Good point, hallucinations only add to the fake news problem and artificial content problem.
I’ll counter with this: how do you know the stuff you look up online is legit? Should we go back to encyclopedias? Who writes those?
Edit: in case anyone isn’t aware, GPT “hallucinates” made up information in specific cases when temperature and top_p settings aren’t optimized, wasn’t saying anyone’s opinion was a hallucination of course
Some generative chatbots will say something then link to where the info is from. That’s good because I can followup
Some will just say something. That’s bad and I’ll have to search myself afterwards.
It’s the equivalent of a book with no cover or a webpage where I can’t see what website it’s on. Maybe it’s reputable, maybe it’s not. Without a source I can’t really decide
Good point, hallucinations only add to the fake news problem and artificial content problem.
I’ll counter with this: how do you know the stuff you look up online is legit? Should we go back to encyclopedias? Who writes those?
Edit: in case anyone isn’t aware, GPT “hallucinates” made up information in specific cases when temperature and top_p settings aren’t optimized, wasn’t saying anyone’s opinion was a hallucination of course
Some generative chatbots will say something then link to where the info is from. That’s good because I can followup
Some will just say something. That’s bad and I’ll have to search myself afterwards.
It’s the equivalent of a book with no cover or a webpage where I can’t see what website it’s on. Maybe it’s reputable, maybe it’s not. Without a source I can’t really decide