如果你最近讓ChatGPT做某件事情卻被它拒絕,或者它傲慢地推諉責(zé)任,不只你有這樣的遭遇。
去年12月,ChatGPT在X上承認(rèn),人們擔(dān)心它變得太懶惰。當(dāng)時(shí),奧特曼寫(xiě)道“我們從11月11日以后就沒(méi)有更新模型,這絕非有意為之”,這讓數(shù)以十億計(jì)依賴(lài)這款軟件寫(xiě)郵件和編寫(xiě)代碼的上班族(和學(xué)生)松了一口氣。
OpenAI的億萬(wàn)富翁CEO山姆·奧特曼表示,ChatGPT在實(shí)現(xiàn)新年計(jì)劃的時(shí)候起步緩慢,因?yàn)樗憩F(xiàn)出越來(lái)越多壞習(xí)慣,例如懶惰的傾向,堅(jiān)持不做它能做的事情,甚至變得有些傲慢。正如奧特曼近日在X上宣布軟件更新時(shí)承認(rèn)的那樣,ChatGPT機(jī)器人“現(xiàn)在應(yīng)該變得更懶惰了”。
但如果你看過(guò)奧特曼最近幾個(gè)月的評(píng)論,你會(huì)發(fā)現(xiàn)這款機(jī)器人的個(gè)性(或者個(gè)性的幻覺(jué))很難抑制。他在12月發(fā)帖稱(chēng):“模型的行為無(wú)法預(yù)測(cè)。”他補(bǔ)充道,訓(xùn)練聊天機(jī)器人模型可能像訓(xùn)練寵物狗一樣,它們都不會(huì)始終給出相同的反應(yīng)。
奧特曼寫(xiě)道:“即使使用相同數(shù)據(jù)集進(jìn)行的不同訓(xùn)練,可能使模型出現(xiàn)明顯不同的個(gè)性、書(shū)寫(xiě)風(fēng)格、拒絕行為、評(píng)價(jià)表現(xiàn)甚至政治偏好?!?/p>
OpenAI并未回應(yīng)置評(píng)請(qǐng)求。
ChatGPT的行為變化
在去年年末,越來(lái)越多人對(duì)這款對(duì)話AI平臺(tái)表示不滿(mǎn),用戶(hù)抱怨它給出的回答與他們的要求不匹配。在Reddit上,有數(shù)十名用戶(hù)描述了這款聊天機(jī)器人在執(zhí)行任務(wù)時(shí)的不服從性,例如用戶(hù)要求它寫(xiě)一篇與一個(gè)主題有關(guān)的簡(jiǎn)短總結(jié),卻收到了一篇論文。一位Reddit用戶(hù)寫(xiě)道:“如果我說(shuō)‘再短一些,用最多三四句話’,它會(huì)給我有十句話的一個(gè)段落?!?
甚至在去年早些時(shí)候,就已經(jīng)有一些警告信號(hào)。例如,斯坦福大學(xué)/加州大學(xué)伯克利分校的一項(xiàng)研究,其名稱(chēng)《ChatGPT的行為正在如何發(fā)生變化?》(How is ChatGPT’s behavior changing over time?)可謂恰如其分。研究發(fā)現(xiàn),AI軟件處理某些任務(wù)的準(zhǔn)確度有起伏或巨大波動(dòng),在短短三個(gè)月后,它解決數(shù)學(xué)問(wèn)題和識(shí)別素?cái)?shù)的準(zhǔn)確度,從約97%下降到不足3%。
ChatGPT用戶(hù)還曝出了其他問(wèn)題,例如這款機(jī)器人會(huì)說(shuō)在2023年最近的更新之后它無(wú)法瀏覽互聯(lián)網(wǎng)。在用戶(hù)鼓勵(lì)它說(shuō)“你可以的”時(shí),這款聊天機(jī)器人似乎記起了它能瀏覽互聯(lián)網(wǎng)。Reddit上討論的話題包括人們是否認(rèn)為較早版本的軟件運(yùn)行效果更好,還有人認(rèn)為這款軟件的新穎性引起的狂熱,讓人們誤以為它的舊版本比現(xiàn)在的版本更機(jī)智。
開(kāi)發(fā)者羅伯·林奇表示,這款聊天機(jī)器人在冬季放慢節(jié)奏,可能是因?yàn)樗A(yù)期人類(lèi)也會(huì)在節(jié)日期間放慢速度,也有可能這是另外一個(gè)無(wú)法解釋的故障。林奇在X上發(fā)文稱(chēng),他對(duì)這款聊天機(jī)器人進(jìn)行了測(cè)試,當(dāng)它以為時(shí)間是12月而不是5月時(shí),就會(huì)給出更簡(jiǎn)短的回答。
自2022年11月ChatGPT公開(kāi)發(fā)布以來(lái),它已經(jīng)擁有超過(guò)17億用戶(hù)。人們借助這款軟件寫(xiě)郵件、信函和報(bào)告,而更精通技術(shù)的用戶(hù)則用它來(lái)尋找程序漏洞,編寫(xiě)代碼和進(jìn)行研究分析。
上班族利用這款軟件增強(qiáng)自己的工作能力,并聲稱(chēng)它幫助他們“更聰明地工作,而不是工作得更辛苦”。
令人擔(dān)憂(yōu)的是,研究發(fā)現(xiàn)了聊天機(jī)器人在評(píng)估某些職業(yè)和行為時(shí)表現(xiàn)出的性別偏見(jiàn),它會(huì)將男性與醫(yī)生和上班關(guān)聯(lián)在一起,而與女性關(guān)聯(lián)的則是護(hù)士和烹飪。
在客戶(hù)服務(wù)和醫(yī)療保健等領(lǐng)域,行業(yè)專(zhuān)家警告,過(guò)度依賴(lài)ChatGPT可能意味著失去一些領(lǐng)域依舊需要培養(yǎng)的人與人之間的聯(lián)系和同理心。如果你與客戶(hù)服務(wù)機(jī)器人通了40分鐘電話,你就能深有體會(huì)。但現(xiàn)在我們不得不等著看看ChatGPT變懶惰的問(wèn)題能否解決。(財(cái)富中文網(wǎng))
翻譯:劉進(jìn)龍
審校:汪皓
如果你最近讓ChatGPT做某件事情卻被它拒絕,或者它傲慢地推諉責(zé)任,不只你有這樣的遭遇。
去年12月,ChatGPT在X上承認(rèn),人們擔(dān)心它變得太懶惰。當(dāng)時(shí),奧特曼寫(xiě)道“我們從11月11日以后就沒(méi)有更新模型,這絕非有意為之”,這讓數(shù)以十億計(jì)依賴(lài)這款軟件寫(xiě)郵件和編寫(xiě)代碼的上班族(和學(xué)生)松了一口氣。
OpenAI的億萬(wàn)富翁CEO山姆·奧特曼表示,ChatGPT在實(shí)現(xiàn)新年計(jì)劃的時(shí)候起步緩慢,因?yàn)樗憩F(xiàn)出越來(lái)越多壞習(xí)慣,例如懶惰的傾向,堅(jiān)持不做它能做的事情,甚至變得有些傲慢。正如奧特曼近日在X上宣布軟件更新時(shí)承認(rèn)的那樣,ChatGPT機(jī)器人“現(xiàn)在應(yīng)該變得更懶惰了”。
但如果你看過(guò)奧特曼最近幾個(gè)月的評(píng)論,你會(huì)發(fā)現(xiàn)這款機(jī)器人的個(gè)性(或者個(gè)性的幻覺(jué))很難抑制。他在12月發(fā)帖稱(chēng):“模型的行為無(wú)法預(yù)測(cè)?!彼a(bǔ)充道,訓(xùn)練聊天機(jī)器人模型可能像訓(xùn)練寵物狗一樣,它們都不會(huì)始終給出相同的反應(yīng)。
奧特曼寫(xiě)道:“即使使用相同數(shù)據(jù)集進(jìn)行的不同訓(xùn)練,可能使模型出現(xiàn)明顯不同的個(gè)性、書(shū)寫(xiě)風(fēng)格、拒絕行為、評(píng)價(jià)表現(xiàn)甚至政治偏好?!?/p>
OpenAI并未回應(yīng)置評(píng)請(qǐng)求。
ChatGPT的行為變化
在去年年末,越來(lái)越多人對(duì)這款對(duì)話AI平臺(tái)表示不滿(mǎn),用戶(hù)抱怨它給出的回答與他們的要求不匹配。在Reddit上,有數(shù)十名用戶(hù)描述了這款聊天機(jī)器人在執(zhí)行任務(wù)時(shí)的不服從性,例如用戶(hù)要求它寫(xiě)一篇與一個(gè)主題有關(guān)的簡(jiǎn)短總結(jié),卻收到了一篇論文。一位Reddit用戶(hù)寫(xiě)道:“如果我說(shuō)‘再短一些,用最多三四句話’,它會(huì)給我有十句話的一個(gè)段落?!?
甚至在去年早些時(shí)候,就已經(jīng)有一些警告信號(hào)。例如,斯坦福大學(xué)/加州大學(xué)伯克利分校的一項(xiàng)研究,其名稱(chēng)《ChatGPT的行為正在如何發(fā)生變化?》(How is ChatGPT’s behavior changing over time?)可謂恰如其分。研究發(fā)現(xiàn),AI軟件處理某些任務(wù)的準(zhǔn)確度有起伏或巨大波動(dòng),在短短三個(gè)月后,它解決數(shù)學(xué)問(wèn)題和識(shí)別素?cái)?shù)的準(zhǔn)確度,從約97%下降到不足3%。
ChatGPT用戶(hù)還曝出了其他問(wèn)題,例如這款機(jī)器人會(huì)說(shuō)在2023年最近的更新之后它無(wú)法瀏覽互聯(lián)網(wǎng)。在用戶(hù)鼓勵(lì)它說(shuō)“你可以的”時(shí),這款聊天機(jī)器人似乎記起了它能瀏覽互聯(lián)網(wǎng)。Reddit上討論的話題包括人們是否認(rèn)為較早版本的軟件運(yùn)行效果更好,還有人認(rèn)為這款軟件的新穎性引起的狂熱,讓人們誤以為它的舊版本比現(xiàn)在的版本更機(jī)智。
開(kāi)發(fā)者羅伯·林奇表示,這款聊天機(jī)器人在冬季放慢節(jié)奏,可能是因?yàn)樗A(yù)期人類(lèi)也會(huì)在節(jié)日期間放慢速度,也有可能這是另外一個(gè)無(wú)法解釋的故障。林奇在X上發(fā)文稱(chēng),他對(duì)這款聊天機(jī)器人進(jìn)行了測(cè)試,當(dāng)它以為時(shí)間是12月而不是5月時(shí),就會(huì)給出更簡(jiǎn)短的回答。
自2022年11月ChatGPT公開(kāi)發(fā)布以來(lái),它已經(jīng)擁有超過(guò)17億用戶(hù)。人們借助這款軟件寫(xiě)郵件、信函和報(bào)告,而更精通技術(shù)的用戶(hù)則用它來(lái)尋找程序漏洞,編寫(xiě)代碼和進(jìn)行研究分析。
上班族利用這款軟件增強(qiáng)自己的工作能力,并聲稱(chēng)它幫助他們“更聰明地工作,而不是工作得更辛苦”。
令人擔(dān)憂(yōu)的是,研究發(fā)現(xiàn)了聊天機(jī)器人在評(píng)估某些職業(yè)和行為時(shí)表現(xiàn)出的性別偏見(jiàn),它會(huì)將男性與醫(yī)生和上班關(guān)聯(lián)在一起,而與女性關(guān)聯(lián)的則是護(hù)士和烹飪。
在客戶(hù)服務(wù)和醫(yī)療保健等領(lǐng)域,行業(yè)專(zhuān)家警告,過(guò)度依賴(lài)ChatGPT可能意味著失去一些領(lǐng)域依舊需要培養(yǎng)的人與人之間的聯(lián)系和同理心。如果你與客戶(hù)服務(wù)機(jī)器人通了40分鐘電話,你就能深有體會(huì)。但現(xiàn)在我們不得不等著看看ChatGPT變懶惰的問(wèn)題能否解決。(財(cái)富中文網(wǎng))
翻譯:劉進(jìn)龍
審校:汪皓
If you asked ChatGPT to do something lately and it refused, or sassily gave you the runaround, you’re not alone.
On X last December, ChatGPT acknowledged that people were concerned it’s gotten too lazy. Last December, Altman wrote that “we haven’t updated the model since Nov. 11, and this certainly isn’t intentional,” much to the relief of billions of workers (and students) who have come to rely on the software to write emails and code.
ChatGPT has had a slow start to its New Year’s resolution, according to OpenAI’s billionaire CEO Sam Altman, as it’s displayed an increasing number of bad habits, from a tendency to be lazy to an insistence it can’t do things that it can to even just being kind of sassy. As Altman acknowledged yesterday on X, along with the news of a software update, the ChatGPT bot “should now be much less lazy.”
But if you look at Altman’s comments for months now, the bot’s personalities (or illusions of them) are hard to suppress. “Model behavior can be unpredictable,” he wrote in a December post. Training chat models, he added, might be more similar to training puppies—they don’t always respond the same.
“Different training runs even using the same datasets can produce models that are noticeably different in personality, writing style, refusal behavior, evaluation performance, and even political bias,” Altman wrote.
OpenAI did not respond to a request for comment.
ChatGPT’s changing behavior
As last year wound down, complaints against the conversational AI platform rose as users reported responses that didn’t match their requests. On Reddit, dozens of users described the bot’s disobedience on tasks, like asking it to write a short summary on a topic and instead receiving an essay. One Redditor wrote, “And if i say, ‘much shorter, 3-4 sentences max,’ it’ll give me a paragraph with ten sentences.’”
Warning signs emerged even earlier last year, notably including a Stanford/Berkeley study, aptly named “How is ChatGPT’s behavior changing over time?” It found that the AI software had drifts, or wild fluctuations in accuracy with certain tasks, and its facility in solving math problems and identifying prime numbers had dropped from about 97% to under 3% accuracy after just three months.
ChatGPT users have described other issues, like the bot saying it can’t browse the internet beyond its recent update in 2023. When the user encouraged it, typing in “yes you can,” the bot seemed to remember it could browse the internet. Reddit discussions include whether people think older versions of the software worked better, while others think that the frenzy over the newness of the software makes it seem like it used to be more tactful than now.
One developer, Rob Lynch, posited that the chatbot may have been taking it slow for the winter, possibly due to its expectation that its human counterparts also slow down during the holidays, or just another unexplained glitch. Lynch posted on X that he had run tests on the bot and it would reply in shorter answers when it thought it was December rather than May.
Since ChatGPT’s debut to the public in November 2022, it has amassed an estimate of over 1.7 billion users. People have been using the software to write emails, letters, and reports, and the more tech-savvy depend on it to debug and write codes and research analysis.
Workers have bolstered their manpower with the software and claim it helps them “work smarter, not harder.“
Of concern, studies have revealed the chatbot’s gender bias through its assessment of certain careers and actions, associating men with doctors and going to work, and women to nurses and cooking.
In fields like customer service and health care, industry experts warn that overreliance on ChatGPT could mean a loss of human connection and empathy that some fields still need to foster. If you’ve ever been on a 40-minute call with a customer service robot, you already know. But now we have to wait and see if the puppy’s laziness problem fades.