Large Language Models pose threat to scientific accuracy, new study warns

Large Language Models (LLMs) pose a direct threat to science, because of so-called 'hallucinations' (untruthful responses), and should be restricted to protect scientific truth, says a new paper from leading Artificial Intelligence researchers at the Oxford Internet Institute.

The paper by Professors Brent Mittelstadt, Chris Russell and Sandra Wachter has been published in Nature Human Behaviour. It explains, 'LLMs are designed to produce helpful and convincing responses without any overriding guarantees regarding their accuracy or alignment with fact.'

One reason for this, is that the data which the technology uses to answer questions does not always come from a factually correct source. LLMs are trained on large datasets of text, usually taken from online sources. These can contain false statements, opinions, and creative writing amongst other types of non-factual information.

Prof Mittelstadt explains, 'People using LLMs often anthropomorphise the technology, where they trust it as a human-like information source. This is, in part, due to the design of LLMs as helpful, human-sounding agents that converse with users and answer seemingly any question with confident sounding, well-written text. The result of this is that users can easily be convinced that responses are accurate even when they have no basis in fact or present a biased or partial version of the truth.'

To protect science and education from the spread of bad and biased information, the authors argue, clear expectations should be set around what LLMs can responsibly and helpfully contribute. According to the paper, 'For tasks where the truth matters, we encourage users to write translation prompts that include vetted, factual Information.'

The way in which LLMs are used matters. In the scientific community it is vital that we have confidence in factual information, so it is important to use LLMs responsibly. If LLMs are used to generate and disseminate scientific articles, serious harms could result."

Prof. Sandra Wachter

Prof Russell adds, 'It's important to take a step back from the opportunities LLMs offer and consider whether we want to give those opportunities to a technology, just because we can.'
LLMs are currently treated as knowledge-bases, and used to generate information in response to questions. This makes the user vulnerable both to regurgitated false information that was present in the training data, and to 'hallucinations' - false information spontaneously generated by the LLM that was not present in the training data.

To overcome this, the authors argue that LLMs should instead be used as `zero-shot translators'. Rather than relying on the LLM as a source of relevant information, the user should simply provide the LLM with appropriate information and ask it to transform it into a desired output. For example: rewriting bullet points as a conclusion or generating code to transform scientific data into a graph.

Using LLMs in this way makes it easier to check that the output is factually correct and consistent with the provided input.

The authors acknowledge that the technology will undoubtedly assist with scientific workflows but are clear that scrutiny of its outputs is key to protecting robust science.

Source:
Journal reference:

Mittelstadt, B., et al. (2023). To protect science, we must use LLMs as zero-shot translators. Nature Human Behaviour. doi.org/10.1038/s41562-023-01744-0.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
AI models boost detection of cognitive decline in medical records, study finds