According toforecasts, generative AI, the software engine behind ChatGPT, could sharply boost productivity and add trillions of dollars to the global economy. But whether we look back at steam power or the internet, history teaches us that there is quite a time lag between the arrival of a new technology and its broad adoption.
So to reassure you that AI isn’t going to steal our jobs next week, we decided to share with you the funniest and stupidest interactions with ChatGPT that were uploaded tothe subreddit of the same name.
This post may includeaffiliate links.
As writer Ian Bogostpointed outinThe Atlantic, there are many reasons why ChatGPT sometimes struggles.But first and foremost, it lacks the ability to truly understand the complexity of human language and conversation. Chatbots are trained to generate words based on a given input, but they do not have the ability to truly comprehend the meaning behind those words.This means that the responses they put together can be shallow and lacking in depth and insight.
As writer Ian Bogostpointed outinThe Atlantic, there are many reasons why ChatGPT sometimes struggles.
But first and foremost, it lacks the ability to truly understand the complexity of human language and conversation. Chatbots are trained to generate words based on a given input, but they do not have the ability to truly comprehend the meaning behind those words.
This means that the responses they put together can be shallow and lacking in depth and insight.
John P. Nelson, who is a Postdoctoral Research Fellow in Ethics and Societal Implications of Artificial Intelligence at Georgia Institute of Technology, agrees that so far, large language models, for all their complexity, “are actually really dumb.““ChatGPT can’t learn, improve or even stay up to date without humans giving it new content and telling it how to interpret that content, not to mention programming the model and building, maintaining and powering its hardware,” NelsonwroteinThe Conversation.
John P. Nelson, who is a Postdoctoral Research Fellow in Ethics and Societal Implications of Artificial Intelligence at Georgia Institute of Technology, agrees that so far, large language models, for all their complexity, “are actually really dumb.”
“ChatGPT can’t learn, improve or even stay up to date without humans giving it new content and telling it how to interpret that content, not to mention programming the model and building, maintaining and powering its hardware,” NelsonwroteinThe Conversation.
“In my own testing, ChatGPT summarized the plot of J.R.R. Tolkien’s ‘The Lord of the Rings,’ a very famous novel, with only a few mistakes. But its summaries of Gilbert and Sullivan’s ‘The Pirates of Penzance’ and of Ursula K. Le Guin’s ‘The Left Hand of Darkness’ – both slightly more niche but far from obscure – come close to playing Mad Libs with the character and place names,” Nelson explained.“It doesn’t matter how good these works’ respective Wikipedia pages are. The model needs feedback, not just content.”
“In my own testing, ChatGPT summarized the plot of J.R.R. Tolkien’s ‘The Lord of the Rings,’ a very famous novel, with only a few mistakes. But its summaries of Gilbert and Sullivan’s ‘The Pirates of Penzance’ and of Ursula K. Le Guin’s ‘The Left Hand of Darkness’ – both slightly more niche but far from obscure – come close to playing Mad Libs with the character and place names,” Nelson explained.
“It doesn’t matter how good these works’ respective Wikipedia pages are. The model needs feedback, not just content.”
Large language models don’t actually understand or evaluate information, so they depend on humans to do it for them. “They are parasitic on human knowledge and labor. When new sources are added into their training data sets, they need new training on whether and how to build sentences based on those sources,” Nelson added.“They can’t evaluate whether news reports are accurate or not. They can’t assess arguments or weigh trade-offs. They can’t even read an encyclopedia page and only make statements consistent with it, or accurately summarize the plot of a movie. They rely on human beings to do all these things for them.”
Large language models don’t actually understand or evaluate information, so they depend on humans to do it for them. “They are parasitic on human knowledge and labor. When new sources are added into their training data sets, they need new training on whether and how to build sentences based on those sources,” Nelson added.
“They can’t evaluate whether news reports are accurate or not. They can’t assess arguments or weigh trade-offs. They can’t even read an encyclopedia page and only make statements consistent with it, or accurately summarize the plot of a movie. They rely on human beings to do all these things for them.”
Then, Nelson highlighted, they paraphrase and remix what humans have said, and rely on yet more human beings to tell them whether they’ve paraphrased and remixed well.“If the common wisdom on some topic changes – for example, whether salt is bad for your heart or whether early breast cancer screenings are useful – they will need to be extensively retrained to incorporate the new consensus.”
Then, Nelson highlighted, they paraphrase and remix what humans have said, and rely on yet more human beings to tell them whether they’ve paraphrased and remixed well.
“If the common wisdom on some topic changes – for example, whether salt is bad for your heart or whether early breast cancer screenings are useful – they will need to be extensively retrained to incorporate the new consensus.”
They were paid no more than US$2 an hour, and many understandably reported experiencing psychological distress because of this work.
“Large language models illustrate the total dependence of many AI systems, not only on their designers and maintainers but on their users,” Nelson added. “So if ChatGPT gives you a good or useful answer about something, remember to thank the thousands or millions of hidden people who wrote the words it crunched and who taught it what were good and bad answers.“As the researcher said, far from being an autonomous superintelligence, ChatGPT is, like all technologies, nothing without us.
“Large language models illustrate the total dependence of many AI systems, not only on their designers and maintainers but on their users,” Nelson added. “So if ChatGPT gives you a good or useful answer about something, remember to thank the thousands or millions of hidden people who wrote the words it crunched and who taught it what were good and bad answers.”
As the researcher said, far from being an autonomous superintelligence, ChatGPT is, like all technologies, nothing without us.
Continue reading with Bored Panda PremiumUnlimited contentAd-free browsingDark modeSubscribe nowAlready a subscriber?Sign In
Continue reading with Bored Panda Premium
Unlimited contentAd-free browsingDark mode
Unlimited content
Ad-free browsing
Dark mode
Subscribe nowAlready a subscriber?Sign In
See Also on Bored Panda
Modal closeAdd New ImageModal closeAdd Your Photo To This ListPlease use high-res photos without watermarksOoops! Your image is too large, maximum file size is 8 MB.Not your original work?Add sourcePublish
Modal close
Add New ImageModal closeAdd Your Photo To This ListPlease use high-res photos without watermarksOoops! Your image is too large, maximum file size is 8 MB.Not your original work?Add sourcePublish
Modal closeAdd Your Photo To This ListPlease use high-res photos without watermarksOoops! Your image is too large, maximum file size is 8 MB.Not your original work?Add sourcePublish
Add Your Photo To This ListPlease use high-res photos without watermarksOoops! Your image is too large, maximum file size is 8 MB.
Add Your Photo To This List
Please use high-res photos without watermarks
Ooops! Your image is too large, maximum file size is 8 MB.
Not your original work?Add source
Modal closeModal closeOoops! Your image is too large, maximum file size is 8 MB.UploadUploadError occurred when generating embed. Please check link and try again.TwitterRender conversationUse html versionGenerate not embedded versionAdd watermarkInstagramShow Image OnlyHide CaptionCropAdd watermarkFacebookShow Image OnlyAdd watermarkChangeSourceTitleUpdateAdd Image
Modal closeOoops! Your image is too large, maximum file size is 8 MB.UploadUploadError occurred when generating embed. Please check link and try again.TwitterRender conversationUse html versionGenerate not embedded versionAdd watermarkInstagramShow Image OnlyHide CaptionCropAdd watermarkFacebookShow Image OnlyAdd watermarkChangeSourceTitleUpdateAdd Image
Upload
UploadError occurred when generating embed. Please check link and try again.TwitterRender conversationUse html versionGenerate not embedded versionAdd watermarkInstagramShow Image OnlyHide CaptionCropAdd watermarkFacebookShow Image OnlyAdd watermark
Error occurred when generating embed. Please check link and try again.
TwitterRender conversationUse html versionGenerate not embedded versionAdd watermark
InstagramShow Image OnlyHide CaptionCropAdd watermark
FacebookShow Image OnlyAdd watermark
ChangeSourceTitle
You May Like50 AI Art Fails That Are Both Horrifying And HilariousEglė Bliabaitė50 Times Clueless Older People On Social Media Cracked People UpAurelija Rakauskaitė30 Architecture Fails That Are So Bad They’re FunnyIlona Baliūnaitė
Eglė Bliabaitė
Aurelija Rakauskaitė
Ilona Baliūnaitė
Funny