Link to original tweet:
https://twitter.com/sayashk/status/1671576723580936193?s=46&t=OEG0fcSTxko2ppiL47BW1Q
Screenshot:
Transcript:
I’d heard that GPT-4’s image analysis feature wasn’t available to the public because it could be used to break Captcha.
Turns out it’s true: The new Bing can break captcha, despite saying it won’t: (image)
Or when it tells you that it can do something it actually can’t, and it hallucinates like crazy. In the early days of ChatGPT I asked it to summarize an article at a link, and it gave me a very believable but completely false summary based on the words in the URL.
This was the first time I saw wild hallucination. It was astounding.
It’s even better when you ask it to write code for you, it generates a decent looking block, but upon closer inspection it imports a nonexistent library that just happens to do exactly what you were looking for.
That’s the best sort of hallucination, because it gets your hopes up.
Yes, for a moment you think “oh, there’s such a convenient API for this” and then you realize…
But we programmers can at least compile/run the code and find out if it’s wrong (most of the time). It is much harder in other fields.