针眼是什么原因引起的| 绮字五行属什么| 什么叫手足口病| 北京大学什么专业最好| 煲鸡汤放什么材料好| as医学上是什么意思| 婚动是什么意思| 孔子是什么家| 手是什么生肖| 什么情况下需要安装心脏起搏器| 38妇女节送老婆什么礼物| 和田玉对身体有什么好处| 12月27号是什么星座| 什么什么的荷叶| 3月24日是什么星座| 小便分叉是什么症状| 小腿肌肉痛什么原因| 弥可保是什么药| 香肠炒什么菜好吃| 卵巢畸胎瘤是什么病| 什么叫高危行为| 五月26日是什么星座| 艾灰有什么作用和功效| 中国文字博大精深什么意思| 子宫肌瘤是什么意思| cpi是什么| 海马体是什么意思| 海椒是什么辣椒| 天池为什么没有鱼| 氮泵是什么| studio什么牌子| 豪五行属什么| 男人头发硬说明什么| 晒伤了涂什么药| 脑袋疼挂什么科| 女人脖子后面有痣代表什么| 梦见老公出轨了是什么征兆| 基因是什么意思| 炒菜什么油最好| 什么是变异性哮喘| 什么都| sheen是什么牌子的手表| 包皮龟头炎用什么药膏| 殁送是什么意思| 经期吃榴莲有什么好处和坏处| 茯苓长什么样子图片| 和硕是什么意思| 节节草有什么作用| 什么时候出伏| 倒打一耙的前一句是什么| 做梦梦到搬家什么意思| 根是什么生肖| 爱出油的人身体缺什么| 尿糖1十是什么意思| 朱雀是什么| 血常规检查能查出什么| 霉菌阳性是什么意思| 提肛运动有什么好处| 什么是天乙贵人| 腹泻吃什么| 2017属什么生肖| 神态自若是什么意思| 燕窝什么时候吃好| 喉炎吃什么药| xsh是什么意思| 西瓜不可以和什么同食| 什么是阻生智齿| 艾滋病英文缩写是什么| 心胆气虚吃什么中成药| 昆仑山在什么地方| 辅酶q10的作用是什么| 9.3是什么日子| 高数是什么| 为什么喝水血糖也会高| 像蜈蚣一样的虫子叫什么| 32岁属什么| 淀粉酶是什么| 孩子发烧挂什么科| mankind是什么意思| 上环什么时候去最合适| 七月种什么菜| 男人做梦梦到蛇是什么意思| 芡实不能和什么一起吃| 94年属什么今年多大| 什么面玲珑| 梦见枪毙人是什么意思| 经常催吐有什么危害| 中国的国树是什么树| 心悸是什么| 为什么会厌学| 怀孕哭对宝宝有什么影响| dbm是什么意思| 枣红色配什么颜色好看| 吃什么可以补精子| 门前的小树已成年是什么歌| 血液病有什么症状| 一什么木瓜| 什么人招蚊子| 烟酒不沾的人什么性格| 躺下就头晕是什么原因| 626什么星座| 血压有点低是什么原因| 夏至为什么吃馄饨| 牛奶洗脸有什么好处| 梦见盗墓是什么意思| 妇科炎症吃什么食物好| 颈动脉斑块做什么检查| 肾虚吃什么食物能补| 少白头是什么原因| 葬花是什么意思| 火箭是干什么用的| 中耳炎有什么症状| cup什么意思| 过江龙是什么意思| 星月菩提是什么| 摩什么擦什么| 分拣员是做什么的| 热量是什么| 子宫内膜增厚是什么原因引起的| 什么样的脚好看| 空调干燥是什么意思| 小龙女叫什么名字| 莫言是什么学历| 刺激是什么意思| 查幽门螺旋杆菌挂什么科| 回心转意是什么意思| 玄胡又叫什么| 什么蛋营养价值最高| 什么是野鸡大学| 吕布的武器叫什么| 肝风内动是什么原因造成的| 菊花什么时候开放| 左肋骨下方隐隐疼痛是什么原因| 字是什么结构| 幽默什么意思| baumwolle是什么面料| 什么利尿| 肝功能挂什么科| 甘油三酯高吃什么药好| 肛裂是什么感觉| 梦见狗咬别人是什么意思| 丹毒是什么病| 昆仑雪菊有什么作用| 为什么女生喜欢腹肌| 小孩摇头是什么原因| 糖尿病吃什么水果最好| 脑梗是什么病严重吗| 西瓜跟什么不能一起吃| 胃痉挛是什么| 独什么心什么| 什么食物补血效果最好最快| 为什么会突然得荨麻疹| 马甲线是什么意思| 击剑什么意思| 手机充电发烫是什么原因| 24小时动态脑电图能查出什么| 吃什么精力旺盛有精神| 昶字五行属什么| 40岁属什么生肖| 通风什么征兆| 雄五行属什么| 什么事情只能用一只手去做| 猫砂是什么| 前列腺增生有什么症状表现| 两毛二是什么军衔| 6月25日是世界什么日| 吃什么有助于降血压| 咽喉痛吃什么药好得快| trendiano什么牌子| 五月十八什么星座| 水飞蓟是什么| 收尿干什么用的| 鲭鱼是什么鱼| 单核细胞偏高是什么原因| 言音读什么| 异的偏旁是什么| 老年人脚肿是什么原因引起的| 1960年属鼠的是什么命| 结婚40年是什么婚| 九月初四是什么星座| 小病不治下一句是什么| 膈应什么意思| esr是什么意思| 乳腺增生结节吃什么药效果好| 阑尾炎吃什么水果| 2002年属什么生肖| 颈椎病最怕干什么活| 头孢曲松是什么药| 十八岁属什么生肖| c1e驾照能开什么车| 末梢神经炎是什么症状| 81是什么意思| 春梦是什么| 不动产权是什么意思| 回迁房是什么意思| 鸾凤和鸣什么意思| 捷字五行属什么| 喝什么去火| 肠道细菌感染吃什么药| 釜底抽薪是什么意思| 龙跟什么生肖配对最好| 后羿和嫦娥是什么关系| 感冒口苦是什么原因| 农历五月十九是什么日子| 小孩子流鼻血是什么原因引起的| 切诺是什么药| 寸金难买寸光阴什么意思| 喉咙上火吃什么药| 终止是什么意思| 茗字五行属什么| 什么妖魔鬼怪什么美女画皮| 腮边长痘是什么原因| 什么是外心| pass是什么意思| 庹是什么意思| 什么是无性婚姻| 曾毅玲花什么关系| 吃芥末有什么好处| 曲高和寡什么意思| 时髦是什么意思| 长期熬夜吃什么可以补回来| 吃什么能让子宫瘤变小| 腰不好挂什么科| 撸管是什么意思| 献血证有什么用| 什么叫点映| 活性炭和木炭有什么区别| bart是什么意思| 蔡明是什么民族| 鼻窦炎是什么病| 6月7号是什么星座| 什么叫精索静脉曲张啊| 淹死是什么感觉| 什么是化学| 1981年属鸡是什么命| 鸡柳是什么肉| 湿疹和热疹有什么区别| 休闲裤配什么鞋子好看| 人活着的意义是什么| 玉帝叫什么名字| 经常中暑的人体内缺什么| 中国海警是什么编制| 吕布的马叫什么| 手指甲出现竖纹是什么原因| 张国立老婆叫什么名字| 情人节送什么| 1.20是什么星座| 三合生肖是什么意思| mpa是什么意思呀| 戴玉手镯有什么好处| 降噪是什么意思| 鹿茸有什么功效| 替代品是什么意思| 吃避孕药为什么要吃维生素c| 嗯哼是什么意思| 免签国家是什么意思| 阳历八月份是什么星座| 猫咪冠状病毒什么症状| 苡字取名寓意是什么| 11月25日什么星座| 心跳快吃什么药| 黄泉是什么意思| 乳痈是什么病| 频发室性早搏吃什么药| 百度
Skip to content
Unsafe at any seed

传统纸媒探索实用融媒体之路 在精不在多

百度 他与官晶华育有2个女儿,但他与肥肥的女儿郑欣宜年初受访时,透露自己一度穷到户头只剩21元人民币,因未满35岁,还不能动用妈妈的4千万人民币的遗产,对女儿经济窘境,郑少秋全然不知,还对媒体说:“为什么她不找我?”

A look inside the hallucinating artificial minds of the famous text prediction bots.

Benj Edwards | 202
There's something about the way this applicant writes that I can't put my finger on... Credit: Aurich Lawson | Getty Images
There's something about the way this applicant writes that I can't put my finger on... Credit: Aurich Lawson | Getty Images
Story text

Over the past few months, AI chatbots like ChatGPT have captured the world's attention due to their ability to converse in a human-like way on just about any subject. But they come with a serious drawback: They can present convincing false information easily, making them unreliable sources of factual information and potential sources of defamation.

Why do AI chatbots make things up, and will we ever be able to fully trust their output? We asked several experts and dug into how these AI models work to find the answers.

“Hallucinations”—a loaded term in AI

AI chatbots such as OpenAI's ChatGPT rely on a type of AI called a "large language model" (LLM) to generate their responses. An LLM is a computer program trained on millions of text sources that can read and generate "natural language" text—language as humans would naturally write or talk. Unfortunately, they can also make mistakes.

In academic literature, AI researchers often call these mistakes "hallucinations." But that label has grown controversial as the topic becomes mainstream because some people feel it anthropomorphizes AI models (suggesting they have human-like features) or gives them agency (suggesting they can make their own choices) in situations where that should not be implied. The creators of commercial LLMs may also use hallucinations as an excuse to blame the AI model for faulty outputs instead of taking responsibility for the outputs themselves.

Still, generative AI is so new that we need metaphors borrowed from existing ideas to explain these highly technical concepts to the broader public. In this vein, we feel the term "confabulation," although similarly imperfect, is a better metaphor than "hallucination." In human psychology, a "confabulation" occurs when someone's memory has a gap and the brain convincingly fills in the rest without intending to deceive others. ChatGPT does not work like the human brain, but the term "confabulation" arguably serves as a better metaphor because there's a creative gap-filling principle at work, as we'll explore below.

The confabulation problem

It's a big problem when an AI bot generates false information that can potentially mislead, misinform, or defame. Recently, The Washington Post reported on a law professor who discovered that ChatGPT had placed him on a list of legal scholars who had sexually harassed someone. But it never happened—ChatGPT made it up. The same day, Ars reported on an Australian mayor who allegedly found that ChatGPT claimed he had been convicted of bribery and sentenced to prison, a complete fabrication.

Shortly after ChatGPT's launch, people began proclaiming the end of the search engine. At the same time, though, many examples of ChatGPT's confabulations began to circulate on social media. The AI bot has invented books and studies that don't exist, publications that professors didn't write, fake academic papers, false legal citations, non-existent Linux system features, unreal retail mascots, and technical details that don't make sense.

And yet despite ChatGPT's predilection for casually fibbing, counter-intuitively, its resistance to confabulation is why we're even talking about it today. Some experts note that ChatGPT was technically an improvement over vanilla GPT-3 (its predecessor model) because it could refuse to answer some questions or let you know when its answers might not be accurate.

"A major factor in Chat's success is that it manages to suppress confabulation enough to make it unnoticeable for many common questions," said Riley Goodside, an expert in large language models who serves as staff prompt engineer at Scale AI. "Compared to its predecessors, ChatGPT is notably less prone to making things up."

If used as a brainstorming tool, ChatGPT's logical leaps and confabulations might lead to creative breakthroughs. But when used as a factual reference, ChatGPT could cause real harm, and OpenAI knows it.

Not long after the model's launch, OpenAI CEO Sam Altman tweeted, "ChatGPT is incredibly limited, but good enough at some things to create a misleading impression of greatness. It's a mistake to be relying on it for anything important right now. It’s a preview of progress; we have lots of work to do on robustness and truthfulness." In a later tweet, he wrote, "It does know a lot, but the danger is that it is confident and wrong a significant fraction of the time."

What's going on here?

How ChatGPT works

ChatGPT hovering in the library, as one does.
An AI-generated image of a chatbot hovering in the library, as one does.
An AI-generated image of a chatbot hovering in the library, as one does. Credit: Benj Edwards / Stable Diffusion

To understand how a GPT model like ChatGPT or Bing Chat confabulates, we have to know how GPT models work. While OpenAI hasn't released the technical details of ChatGPT, Bing Chat, or even GPT-4, we do have access to the research paper that introduced their precursor, GPT-3, in 2020.

Researchers build (train) large language models like GPT-3 and GPT-4 by using a process called "unsupervised learning," which means the data they use to train the model isn't specially annotated or labeled. During this process, the model is fed a large body of text (millions of books, websites, articles, poems, transcripts, and other sources) and repeatedly tries to predict the next word in every sequence of words. If the model's prediction is close to the actual next word, the neural network updates its parameters to reinforce the patterns that led to that prediction.

Conversely, if the prediction is incorrect, the model adjusts its parameters to improve its performance and tries again. This process of trial and error, though a technique called "backpropagation," allows the model to learn from its mistakes and gradually improve its predictions during the training process.

As a result, GPT learns statistical associations between words and related concepts in the data set. Some people, like OpenAI Chief Scientist Ilya Sutskever, think that GPT models go even further than that, building a sort of internal reality model so they can predict the next best token more accurately, but the idea is controversial. The exact details of how GPT models come up with the next token within their neural nets are still uncertain.

In the current wave of GPT models, this core training (now often called "pre-training") happens only once. After that, people can use the trained neural network in "inference mode," which lets users feed an input into the trained network and get a result. During inference, the input sequence for the GPT model is always provided by a human, and it's called a "prompt." The prompt determines the model's output, and altering the prompt even slightly can dramatically change what the model produces.

For example, if you prompt GPT-3 with "Mary had a," it usually completes the sentence with "little lamb." That's because there are probably thousands of examples of "Mary had a little lamb" in GPT-3's training data set, making it a sensible completion. But if you add more context in the prompt, such as "In the hospital, Mary had a," the result will change and return words like "baby" or "series of tests."

Here's where things get a little funny with ChatGPT, since it's framed as a conversation with an agent rather than just a straight text-completion job. In the case of ChatGPT, the input prompt is the entire conversation you've been having with ChatGPT, starting with your first question or statement and including any specific instructions provided to ChatGPT before the simulated conversation even began. Along the way, ChatGPT keeps a running short-term memory (called the "context window") of everything it and you have written, and when it "talks" to you, it is attempting to complete the transcript of a conversation as a text-completion task.

A diagram showing how GPT conversational language model prompting works
A diagram showing how GPT conversational language model prompting works.
A diagram showing how GPT conversational language model prompting works. Credit: Benj Edwards / Ars Technica

Additionally, ChatGPT is different from vanilla GPT-3 because it has also been trained on transcripts of conversations written by humans. "We trained an initial model using supervised fine-tuning: human AI trainers provided conversations in which they played both sides—the user and an AI assistant," wrote OpenAI in its initial ChatGPT release page. "We gave the trainers access to model-written suggestions to help them compose their responses."

ChatGPT has also been tuned more heavily than GPT-3 using a technique called "reinforcement learning from human feedback," or RLHF, where human raters ranked ChatGPT's responses in order of preference, then fed that information back into the model. Through RLHF, OpenAI was able to instill in the model the goal of refraining from answering many questions it cannot answer reliably. This has allowed the ChatGPT to produce coherent responses with fewer confabulations than the base model. But inaccuracies still slip through.

Why ChatGPT confabulates

Natively, there is nothing in a GPT model's raw data set that separates fact from fiction. That guidance comes from a) the prevalence of accurate content in the data set, b) recognition of factual information in the results by humans, or c) reinforcement learning guidance from humans that emphasizes certain factual responses.

The behavior of LLMs is still an active area of research. Even the researchers who created these GPT models are still discovering surprising properties of the technology that no one predicted when they were first developed. GPT's abilities to do many of the interesting things we are now seeing, such as language translation, programming, and playing chess, were a surprise to researchers at one point (for an early taste of that, check out 2019's GPT-2 research paper and search for the term "surprising").

So when we ask why ChatGPT confabulates, it's difficult to pinpoint an exact technical answer. And because there is a "black box" element of the neural network weights, it's very difficult (if not impossible) to predict their exact output given a complex prompt. Still, we know some basic things about how why confabulation happens.

Key to understanding ChatGPT's confabulation ability is understanding its role as a prediction machine. When ChatGPT confabulates, it is reaching for information or analysis that is not present in its data set and filling in the blanks with plausible-sounding words. ChatGPT is especially good at making things up because of the superhuman amount of data it has to work with, and its ability to glean word context so well helps it place erroneous information seamlessly into the surrounding text.

"I think the best way to think about confabulation is to think about the very nature of large language models: The only thing they know how to do is to pick the next best word based on statistical probability against their training set," said Simon Willison, a software developer who often writes on the topic.

In a 2021 paper, a trio of researchers from the University of Oxford and OpenAI identified two major types of falsehoods that LLMs like ChatGPT might produce. The first comes from inaccurate source material in its training data set, such as common misconceptions (e.g., "eating turkey makes you drowsy"). The second arises from making inferences about specific situations that are absent from its training material (data set); this falls under the aforementioned "hallucination" label.

Whether the GPT model makes a wild guess or not is based on a property that AI researchers call "temperature," which is often characterized as a "creativity" setting. If the creativity is set high, the model will guess wildly; if it's set low, it will spit out data deterministically based on its data set.

Recently, Microsoft employee Mikhail Parakhin, who works on Bing Chat, tweeted about Bing Chat's tendency to hallucinate and what causes it. "This is what I tried to explain previously: hallucinations = creativity," he wrote. "It tries to produce the highest probability continuation of the string using all the data at its disposal. Very often it is correct. Sometimes people have never produced continuations like this."

Parakhin said that those wild creative leaps are what make LLMs interesting. "You can clamp down on hallucinations, and it is super-boring," he wrote. "[It] answers 'I don't know' all the time or only reads what is there in the Search results (also sometimes incorrect). What is missing is the tone of voice: it shouldn't sound so confident in those situations."

Balancing creativity and accuracy is a challenge when it comes to fine-tuning language models like ChatGPT. On the one hand, the ability to come up with creative responses is what makes ChatGPT such a powerful tool for generating new ideas or unblocking writer's block. It also makes the models sound more human. On the other hand, accuracy to the source material is crucial when it comes to producing reliable information and avoiding confabulation. Finding the right balance between the two is an ongoing challenge for the development of language models, but it's one that is essential to produce a tool that is both useful and trustworthy.

There's also the issue of compression. During the training process, GPT-3 considered petabytes of information, but the resulting neural network is only a fraction of that in size. In a widely read New Yorker piece, author Ted Chiang called this a "blurry JPEG of the web." That means a large portion of the factual training data is lost, but GPT-3 makes up for it by learning relationships between concepts that it can later use to reformulate new permutations of these facts. Like a human with a flawed memory working from a hunch of how something works, it sometimes gets things wrong. And, of course, if it doesn't know the answer, it will give its best guess.

We cannot forget the role of the prompt in confabulations. In some ways, ChatGPT is a mirror: It gives you back what you feed it. If you feed it falsehoods, it will tend to agree with you and "think" along those lines. That's why it's important to start fresh with a new prompt when changing subjects or experiencing unwanted responses. And ChatGPT is probabilistic, which means it's partially random in nature. Even with the same prompt, what it outputs can change between sessions.

All this leads to one conclusion, one that OpenAI agrees with: ChatGPT as it is currently designed, is not a reliable source of factual information and cannot be trusted as such. "ChatGPT is great for some things, such as unblocking writer's block or coming up with creative ideas," said Dr. Margaret Mitchell, researcher and chief ethics scientist at AI company Hugging Face. "It was not built to be factual and thus will not be factual. It's as simple as that."

Can the fibbing be fixed?

Trusting an AI chatbot's generations blindly is a mistake, but that may change as the underlying technology improves. Since its release in November, ChatGPT has already been upgraded several times, and some upgrades included improvements in accuracy—and also an improved ability to refuse to answer questions it doesn't know the answers to.

So how does OpenAI plan to make ChatGPT more accurate? We reached out to OpenAI multiple times on this subject over the past few months and received no response. But we can pull out clues from documents OpenAI has released and news reports about the company's attempts to guide ChatGPT's alignment with human workers.

As previously mentioned, one of the reasons why ChatGPT has been so successful is because of extensive training using RLHF. As OpenAI explains, "To make our models safer, more helpful, and more aligned, we use an existing technique called reinforcement learning from human feedback (RLHF). On prompts submitted by our customers to the API, our labelers provide demonstrations of the desired model behavior and rank several outputs from our models. We then use this data to fine-tune GPT-3."

OpenAI's Sutskever believes that additional training through RLHF can fix the hallucination problem. "I'm quite hopeful that by simply improving this subsequent reinforcement learning from human feedback step, we can teach it to not hallucinate," Sutskever said in an interview with Forbes earlier this month.

He continued:

The way we do things today is that we hire people to teach our neural network to behave, to teach ChatGPT to behave. You just interact with it, and it sees from your reaction, it infers, oh, that's not what you wanted. You are not happy with its output. Therefore, the output was not good, and it should do something differently next time. I think there is a quite high chance that this approach will be able to address hallucinations completely.

Others disagree. Yann LeCun, chief AI scientist at Meta, believes hallucination issues will not be solved by the current generation of LLMs that use the GPT architecture. But there is a quickly emerging approach that may bring a great deal more accuracy to LLMs with the current architecture.

"One of the most actively researched approaches for increasing factuality in LLMs is retrieval augmentation—providing external documents to the model to use as sources and supporting context," said Goodside. With that technique, he explained, researchers hope to teach models to use external search engines like Google, "citing reliable sources in their answers as a human researcher might, and rely less on the unreliable factual knowledge learned during model training."

Bing Chat and Google Bard do this already by roping in searches from the web, and soon, a browser-enabled version of ChatGPT will as well. Additionally, ChatGPT plugins aim to supplement GPT-4's training data with information it retrieves from external sources, such as the web and purpose-built databases. This augmentation is similar to how a human with access to an encyclopedia will be more factually accurate than a human without one.

Also, it may be possible to train a model like GPT-4 to be aware of when it is making things up and adjust accordingly. "There are deeper things one can do so that ChatGPT and similar are more factual from the start," said Mitchell, "including more sophisticated data curation and the linking of the training data with 'trust' scores, using a method not unlike PageRank... It would also be possible to fine-tune the model to hedge when it is less confident in the response."

So while ChatGPT is currently in hot water over its confabulations, there may be a way out ahead, and for the sake of a world that is beginning to rely on these tools as essential assistants (for better or worse), an improvement in factual reliability cannot come soon enough.

Listing image: Aurich Lawson | Getty Images

Photo of Benj Edwards
Benj Edwards Senior AI Reporter
Benj Edwards is Ars Technica's Senior AI Reporter and founder of the site's dedicated AI beat in 2022. He's also a tech historian with almost two decades of experience. In his free time, he writes and records music, collects vintage computers, and enjoys nature. He lives in Raleigh, NC.
202 Comments
小孩流鼻涕咳嗽吃什么药 什么洗面奶最好用 检查肺挂什么科 南宁有什么特产 肾与性功能有什么关系
属鸡的守护神是什么菩萨 cg动画是什么意思 孕妇吃火龙果有什么好处 咳嗽挂什么科 见人说人话见鬼说鬼话是什么意思
肺部磨玻璃结节需要注意什么 海棠花长什么样 四书五经是什么 生抽和酱油有什么区别 vivo是什么品牌手机
炒房是什么意思 胸前长痘痘是什么原因 黑玫瑰代表什么意思 巧妙是什么意思 情人节送妈妈什么花
真空是什么意思hcv9jop6ns9r.cn 女性尿路感染什么原因引起的wuhaiwuya.com 做喉镜挂什么科jinxinzhichuang.com 龟头炎用什么药好hcv9jop1ns2r.cn 什么牌子的益生菌调理肠胃比较好hcv8jop2ns6r.cn
外籍是什么意思hcv8jop5ns3r.cn g6pd筛查是检查什么hcv9jop1ns1r.cn 疣长什么样hcv8jop3ns9r.cn 天女散花是什么意思hcv8jop2ns8r.cn 女菩萨是什么意思hcv8jop0ns9r.cn
剑锋金命五行缺什么hcv8jop1ns0r.cn 脾胃不好吃什么水果好hcv9jop3ns2r.cn 胎盘厚度代表什么xscnpatent.com 朋友生日送什么礼物好wuhaiwuya.com 抬头是什么hcv8jop8ns3r.cn
白兰地是什么hcv9jop2ns9r.cn cosplay是什么hcv9jop6ns1r.cn 电瓶车充不进电是什么原因hcv8jop0ns6r.cn pm2.5是什么意思hcv9jop6ns3r.cn 梦到蜈蚣是什么意思hcv9jop0ns4r.cn
百度