In recent months, many people will have experimented with a chatbot like ChatGPT. Useful though they can be, there’s also no shortage of examples of them producing, ahem, erroneous information. Now, a group of scientists from the University of Oxford are asking: is there a legal pathway by which we could require these chatbots to tell us the truth?
The rise of large language models
Amid all the buzz around artificial intelligence (AI), which seems to have reached new heights in the past couple of years, one branch of the field has garnered more attention than any other – at least with those of us who aren’t machine learning experts. It’s the large language models (LLMs), which leverage generative AI to produce often eerily human-sounding responses to almost any query you can dream up.
The likes of ChatGPT and Google’s Gemini are based on models trained on huge amounts of data – which itself raises numerous questions around privacy and intellectual property – to allow them to understand natural language queries and generate coherent and relevant responses. Unlike with a search engine, you don’t need to learn any syntax to help you narrow down your results. Theoretically, you simply ask a question as if you were speaking out loud.
Their capabilities are no doubt impressive, and they certainly sound confident in their answers. There’s just one tiny problem – these chatbots tend to sound equally confident when they’re dead wrong. Which might be okay, if we humans could just remember not to trust everything they’re telling us.
“While problems arising from our tendency to anthropomorphize machines are well established, our vulnerability to treating LLMs as human-like truth tellers is uniquely worrying,” write the authors of the new paper, referring to a situation that anyone who’s ever had an argument with Alexa or Siri will know well.
“LLMs are not designed to tell the truth in any overriding sense.”
It’s easy to tap out a question for ChatGPT and assume that it is “thinking” of the answer in the same way a human person would. That’s how it appears, but that’s not actually how these models work.
Don’t believe everything you read
As the authors explain, LLMs “are text-generation engines designed to predict which string of words comes next in a piece of text.” The truthfulness of their responses is only one metric by which the models are judged during development. In an effort to produce the most “helpful” answer, the authors argue, they can all-too-frequently stray towards oversimplification, bias, and just making stuff up.
The authors of this study are by no means the first to raise the alarm about this, with one paper going so far as to call the models "bullshitters". Professor Robin Emsley, the editor of the journal Schizophrenia, published an account of an experience with ChatGPT in 2023, stating, “What I experienced were fabrications and falsifications.” The chatbot had produced citations for scholarly papers that did not exist, as well as several that were irrelevant to the query. Others have reported the same thing.
They do okay with questions that have a clear, factual answer where – and this bit’s important – that answer has appeared a lot within their training data. These models are only as good as the data they’re trained on. And, unless you’re prepared to carefully fact-check any answer you get from an LLM, it can be very difficult to tell how accurate the information is – especially as many don’t provide links to their source material or give any measure of confidence.
“Unlike human speakers, LLMs do not have any internal conceptions of expertise or confidence, instead always ‘doing their best’ to be helpful and persuasively respond to the prompt posed,” writes the team at Oxford.
They were particularly concerned about the impact of what they call “careless speech”, and of the harm that could be done by such responses from LLMs leeching into offline human conversation. This led them to ask the question of whether there could be a legal obligation imposed on LLM providers to ensure that their models tell the truth.
What did the new study conclude?
Focusing on current European Union (EU) legislation, the authors found that there are few explicit scenarios where a duty is placed on an organization or individual to tell the truth. Those that do exist are limited to specific sectors or institutions, and very rarely apply to the private sector. Since LLMs operate on relatively new technology, the majority of existing regulations were not drawn up with these models in mind.
So, the authors propose a new framework, “the creation of a legal duty to minimize careless speech for providers of both narrow- and general-purpose LLMs.”
You might naturally ask, “Who is the arbiter of truth?”, and the authors do address this by saying that the aim is not to force LLMs down one particular path, but rather to require “plurality and representativeness of sources”. Fundamentally, they propose that makers redress the balance between truthfulness and "helpfulness", which the authors argue is too much in favor of the latter. It’s not simple to do, but it might be possible.
There are no easy answers to these questions (disclaimer: we have not tried asking ChatGPT), but as this technology continues to advance they are things that developers will have to grapple with. In the meantime, when you’re working with an LLM, it may be worth remembering this sobering statement from the authors: “They are designed to participate in natural language conversations with people and offer answers that are convincing and feel helpful, regardless of the truth of the matter at hand.”
The study is published in the journal Royal Society Open Science.