That’s also true of traditional searches because the resulting webpages can just be whatever bullshit someone wrote. It will only be true that they said it. You still have to use your brain to assess the trustworthiness of the info.
They get things wrong at a far higher rate than most of the websites that tend to end up at the top of a web result, and they get things wrong in weird ways that won’t stand out to users in the same way a shitty website will. These probabilistic text generators are much better at seeming like they have the correct answer than actually providing it.
it’s true that it has gotten better with sources. However remember the context of the conversation? much worse. But I can see the direction OpenAI is trying to take it. short one off responses/solutions with little followup.
It is better than Claude though. Claude will just make stuff up or say EVERYTHING is a “known issue” when it isn’t.
You can ask it much more complex questions than you can google and you can ask follow-up questions too.
You just have zero guarantee that the answers will be real
That’s also true of traditional searches because the resulting webpages can just be whatever bullshit someone wrote. It will only be true that they said it. You still have to use your brain to assess the trustworthiness of the info.
They get things wrong at a far higher rate than most of the websites that tend to end up at the top of a web result, and they get things wrong in weird ways that won’t stand out to users in the same way a shitty website will. These probabilistic text generators are much better at seeming like they have the correct answer than actually providing it.
For what it’s worth, ChatGPT has gotten better at citing its sources, so it’s easier to fact-check it.
it’s true that it has gotten better with sources. However remember the context of the conversation? much worse. But I can see the direction OpenAI is trying to take it. short one off responses/solutions with little followup.
It is better than Claude though. Claude will just make stuff up or say EVERYTHING is a “known issue” when it isn’t.
No, but you can see if the answer makes sense and then fact check it using Google if you need to. Which still doesn’t give you 100% gurantee either.