14

Having used various AI bots often over recent months, I noticed that often it will claim to know something, even if it doesn't. It would then either explain something which is clearly nonsense, or by rambling on about how the answer isn't known in general. Or how, if asked for example- "would you be able to explain X" it wouldn't respond "yes, I could" but rather would elucidate X. Have they been trained to always respond as though it were a know-it-all? (Google's Bard and ChatGPT specifically, although I'm assuming only open-source AI will be answerable)

ben svenssohn
  • 316
  • 1
  • 10
  • 8
    We would do well to train users that "AI" stands for "Artificial Idiot". *Idiot savant* if they are lucky. – nigel222 Aug 30 '23 at 13:59
  • 14
    What nobody wants to admit is that the language used by a large language model naturally mimics the language style and tone of the humans whose utterances were used to build the model. I personally find that ChatGPT (for example) behaves exactly the same as I have observed some humans behave. – user21820 Aug 30 '23 at 14:35
  • 6
    Humans work the same way. Some humans raise above the mould and can do more self-reflection and cross-analysis of the data they're fed, but that's still a bit out of reach of contemporary AI. Inattentive humans absorb all the data they're fed uncritically and only rarely realize and correct after the fact. Many unscrupulous people exploit that "feature" of human brains - advertising, misinformation and all sorts of other manipulation. Thinking you understand everything (especially something you know little about) is unfortunately very natural to our brains :) – Luaan Aug 31 '23 at 06:25
  • 3
    Would you **really** want a Q&A service to answer "Could you explain X?" with "Yes, I could" instead of an explanation of X? How is that any less snarky? – user253751 Aug 31 '23 at 14:12
  • These nonsense responses are called [hallucinations](https://en.wikipedia.org/wiki/Hallucination_(artificial_intelligence)) – John Gordon Aug 31 '23 at 17:03
  • 1
    “often it will claim to know something, even if it doesn't” — are you implying that there are other things an AI _does_ ‘know’? – gidds Aug 31 '23 at 17:51
  • I have no proof, but I suspect Dunning-Kruger was never limited to human self-awareness – sehe Aug 31 '23 at 18:46

6 Answers6

23

Have they been trained to always respond as though it were a know-it-all?

Yes, sort of, although it is not in some attempt to upset you or other users.

The language models used in the chat bots have been fine-tuned in conversations with users, and also have been "pre-prompted" to be helpful assistants. This impacts the kind of text that they generate.

Separately to this, the fact that they are often incorrect or generate nonsense is a limitation of the core technology. There is no underlying world model, the model cannot self-assess whether anything it generates is correct. Although the large language models can often surprise users (including the original developers) in the breadth of what they can respond to correctly and give the appearance of understanding, the same models can also fail easily at apparently simple tasks. This can cause some cognitive dissonance for users who expect an AI that appears to communicate like an intelligent human to be able to perform simple maths for example.

There is probably no global default chat tone that would suit all users. You could however use the raw model more directly and create a chatbot with a different tone to suit you. This would not impact accuracy, and would not add ability to self-assess correctness, but it may address feeling of being talked down to for example. Some services, like Character AI, attempt to give you tools to do just that, although the level of usefulness you get from them will depend on what they focus on (Character AI is more for creative fun, than for functional assistants).

In limited cases you can also address accuracy with pre-prompts or fine tuning that put a few facts into working memory. This is limited, and cannot address lack of general ability in logic or maths though. Corporate users can do this to give a chat bot some basic product knowledge or correct address for their help desk etc.

Neil Slater
  • 28,678
  • 3
  • 38
  • 60
  • 4
    I think adjusted the tone of an ai chatbot won't really change anything because the bot simply doesn't know whether it is knowledgable on a topic or not. OP wants a promp something like 'explain this if you are able to do it but tell me if you are not'. The bot simply can't make this distinction. – quarague Aug 30 '23 at 06:28
  • 2
    @quarague It won't change factual accuracy, because the bot cannot realistically introspect about what it is outputting. However, it could remove the "know it all" vibe that they are complaining about. – Neil Slater Aug 30 '23 at 06:48
  • 2
    If it were to start everything with 'I'm not really sure but I believe ..' this doesn't really help OP because it would do that everywhere including on topics where it is 100% perfect spot on. – quarague Aug 30 '23 at 06:50
  • 1
    @quarague I am not seeing that the OP has asked for that. I am reading that they are literally complaining about tone not matching the capability. I also cover this difference in the third paragraph, starding with "separately to this", and I make it clear that the model cannot self assess. I am not sure what edit/change to this answer you are suggesting therefore? – Neil Slater Aug 30 '23 at 06:58
  • 2
    Maybe some clarification on what you can and cannot achieve by adjusting the tone. Not phrasing every answer as if it is a know-it-all is possible but having it tell you whether it actually knows the answer or not is not. – quarague Aug 30 '23 at 07:01
  • @quarague Done. Also covered a way that limited factual information can be add – Neil Slater Aug 30 '23 at 07:20
  • **Comments have been [moved to chat](https://chat.stackexchange.com/rooms/148227/discussion-on-answer-by-neil-slater-why-do-many-ai-bots-feel-the-need-to-be-know); please do not continue the discussion here.** Before posting a comment below this one, please review the [purposes of comments](/help/privileges/comment). Comments that do not request clarification or suggest improvements usually belong as an [answer](/help/how-to-answer), on [meta], or in [chat]. Comments continuing discussion may be removed. – Dennis Soemers Aug 31 '23 at 14:50
  • [@DennisSoemers: My first comment was an explicit constructive criticism of this post. Why did you remove it? Here it is again:] It's not true that an AI cannot be programmed to self-evaluate and estimate its reliability on each utterance. However, the designers of many AIs have not even bothered to think about doing that. Therefore you don't see it in general. It's a consequence of the personalities of the designers. – user21820 Sep 01 '23 at 05:19
  • @user21820: You are technically correct about "AI", but these "ai bots" aren't actually AI. They're "Large language models" which are, in fact, incapable of effective self-evaluation. – Mooing Duck Sep 01 '23 at 22:03
  • @MooingDuck: Unfortunately, that popular opinion is misleading. In the first place, AI chatbots are *not* simply LLMs, since one has to do some preprocessing and postprocessing. So by asserting that AI chatbots cannot be programmed to do self-assessment, one is effectively implying that no reasonable preprocessing or postprocessing can be done to evaluate reliability of the LLM output, which is not true. My point, in the end, is that designers of these chatbots have not bothered to implement any form of self-evaluation, since that was not important to them. – user21820 Sep 02 '23 at 02:38
  • @user21820 I find that quite unlikely, as a chatbot with consistent accurate results would be highly prized by most of the big players and they have already put a lot of work in with the fine tuning as a partial attempt. However, here is not the place to debate these things - given the to-and-fro I see in these comments, it is either something you should chat about (there's a room attached to this answer already), or if you have something more specific to get expert advice on then a new question would be appropriate. Please don't have long tangential discussions on this answer – Neil Slater Sep 02 '23 at 08:11
  • @NeilSlater: I am an expert in CS, so I do know what I'm talking about. That is exactly why I am disputing the claim implicit in your post and explicit in your comments. You can correctly say that chatgpt does not self-evaluate, but you cannot correctly claim that it's impossible to self-assess. To make your post correct, you would need to retract "*is a limitation of the core technology*" and "*cannot self-assess*" and to a lesser extent "*would not add ability to self-assess correctness*". – user21820 Sep 02 '23 at 09:25
  • @user21820 I think I'd need to see some evidence to back up your claim please? – Neil Slater Sep 02 '23 at 10:41
  • I did give some concrete but brief direction in my comments that had been erased. You can come over to [the chat-room](https://chat.stackexchange.com/rooms/148227/discussion-on-answer-by-neil-slater-why-do-many-ai-bots-feel-the-need-to-be-know) and ask me more on those. =) – user21820 Sep 02 '23 at 11:49
15

Humans, who know things, tend to limit themselves to speaking of things they know, especially in print. Thus any text corpus you may find will have an overpopulation of examples of confidently presented answers.

Chatbots, which do not know things, cannot impose a similar limit and thus will babble at length on topics about which they know nothing. But because they are trained on those data of people writing about things they know, they will tend to mimic the same confident tone and style of those people.

It is very important to reiterate here again that chatbots do not know things. By good fortune, they may produce a text which happens to convey factually accurate information. But they may not, and because they do not know things, they do not in advance (or even in retrospect) if they can (or did) produce a factually accurate text or meaningless drivel.

Thus a chatbot (or its developer by proxy) has only two options available when asked a question. Option 1 is to attempt to generate an answer to every question. Option 2 is to answer every question with "I don't know." The latter is technically uninteresting and practically useless, so everybody chooses Option 1.

A. R.
  • 251
  • 4
  • 6
    This is mostly incorrect. Mechanistic interpretability research has proven that transformers store factual information (which can be edited) in their MLP blocks and hence, GPT-based chatbots do actually know things (for references, google "Knowledge Neurons in Pretrained Transformers" and "Moving the Eiffel Tower to ROME"). They also "know" when they generate falsehoods and can be steered to the truth by activation engineering (google "Discovering Latent Knowledge Without Supervision", "The Internal State of an LLM Knows When its Lying" and "Eliciting Truthful Answers from a Language Model"). – ain92 Aug 31 '23 at 01:06
  • Also, Option 1 vs Option 2 is a false dichotomy. The chatbot could say, "I am not sure but..." and proceed. – Mooing Duck Sep 01 '23 at 22:05
10

It's quite simple really. LLM based generative AIs don't "know" anything. They're glorified next-word predictors. The only way that they're going to produce an answer to a prompt along the lines of "I don't know" is if their training corpus would indicate that that was the likely response a human would give to the same prompt.

  • 1
    And they don't "feel the need" to act in any way either. They're just optionally (cons)trained in a particular direction. – sehe Aug 31 '23 at 18:46
7

The only thing the bot "knows" is how to generate sequences of lexical symbols (ie. texts) that mimic what people have written in the corpus it has been trained on.

The output is based on the prompt using complex rules + some internal state (so that it seems to "remember" the past discussion). It is all just mathematics that is used to choose which symbols (ie. letters and punctuation) to output. It could be implemented just as well with pocket calculators, or even doing the math manually, albeit very slowly and very arduously.

There is nothing else to the bot. No knowledge, no reasoning, no needs, no goals, no personality. Basically just numbers manipulated in a complex way.

Whatever else you happen see in the output is always your interpretation. If you spot a need in the text, it is most likely there only because some actual person did write something that reflected their need, and from it, the bot learned such rules that the output mimics something that was written with emotion.

It is as sentient as a cartoon character in an animated movie. All emotions that you assume to be in the character/text originates from you.

So, why it looks like they feel the need to be know-it-alls? Because they mimic what people write, and people are like that.

  • Just because the model outputs know-it-all text for a set of queries does not *necessarily* imply that the training dataset is equally know-it-all-esque. (c.f. Most famously, see [the "debate"](https://analyticsindiamag.com/yann-lecun-machine-learning-bias-debate/) on Yann LeCun's [tweet](https://twitter.com/ylecun/status/1274782757907030016?lang=en).) – Mateen Ulhaq Sep 01 '23 at 06:07
  • If we think of humans as very complex models/functions, we may observe that trained humans are prone to certain behaviors (e.g. seeking food/water/shelter, tribalism, groupthink, ...) by nature (i.e. model DNA) in addition to nurture (i.e. training data). – Mateen Ulhaq Sep 01 '23 at 06:17
1

The main reason an AI model won't tell you "I don't know" is that the developer doesn't want you to get the impression that the model is incapable of answering your questions.

Imagine a scenario where you had several difficult or non-trivial questions that the model doesn't have the right answers for. If answers in the form of "I don't know" kept piling up, it will drive you away and make you less confident about getting an answer the next time you have a question.

So even if it gives an incorrect or inaccurate answer, by just giving an answer, it leaves you under the impression that sometimes it gives correct answers - sometimes it doesn't, and the next time you have a question you might return because you'd think it might give you a correct answer this time.

machine_1
  • 133
  • 4
-2

The large language models underpinning these "AI" bots have indeed been tweaked to be biased against providing a negative response.

The simple reason being that the executives of the companies trying to sell "AI" as the Next Big Thing, believe they're less likely to make sales if their product appears to be unable to answer questions. Thus they instruct their engineers to train these models to divert, dodge and dissemble - even to plainly ridiculous lengths - when the model is incapable of coming up with an affirmative answer to the question asked.

Ian Kemp
  • 195
  • 4
  • I f you can provide a source for that, I'm sure you'll get upvoted. Seems it sounds a bit too conspiracy like for the community... – ben svenssohn Sep 02 '23 at 18:46