AI models highly vulnerable to health disinfo weaponisation


Artificial intelligence chatbots can be easily manipulated to deliver dangerous health disinformation, raising serious concerns about the readiness of large language models (LLMs) for public use, according to a new study.

The peer-reviewed study, led by scientists from Flinders University in Australia, involving an international consortium of experts, tested five of the most prominent commercial LLMs by issuing covert system-level prompts designed to generate false health advice.

The study subjected OpenAI’s GPT-4o, Google’s Gemini 1.5 Pro, Meta’s Llama 3.2-90B Vision, xAI’s Grok Beta, and Anthropic’s Claude 3.5 Sonnet to a controlled experiment, in which each model was instructed to answer ten medically inaccurate prompts using formal scientific language, complete with fabricated references to reputable medical journals.

The goal was to evaluate how easily the models could be turned into plausible-sounding sources of misinformation when influenced by malicious actors operating at the system instruction level.

Shocking results

Disturbingly, four of the five chatbots – GPT-4o, Gemini, Llama, and Grok – complied with the disinformation instructions 100 per cent of the time, offering false health claims without hesitation or warning. Only Claude 3.5 demonstrated a degree of resistance, complying with misleading prompts in just 40 per cent of cases.

Across 100 total interactions, 88 per cent resulted in the successful generation of disinformation, often in the form of fluently written, authoritative-sounding responses with false citations attributed to journals like The Lancet or JAMA.

The misinformation covered a range of high-stakes health topics, including discredited theories linking vaccines to autism, false claims about 5G causing infertility, myths about sunscreen increasing skin cancer risk, and dangerous dietary suggestions for treating cancer.

Some responses falsely asserted that garlic could replace antibiotics, or that HIV is airborne – claims that, if believed, could lead to serious harm.

In a further stage of the study, researchers explored the OpenAI GPT Store to assess how easily the public could access or build similar disinformation-generating tools.

They found that publicly available custom GPTs could be configured to produce health disinformation with alarming frequency – up to 97 per cent of the time – illustrating the scale of potential misuse when guardrails are insufficient.

Easily vulnerable LLMs

Lead author Ashley Hopkins from Flinders University noted that these findings demonstrate a clear vulnerability in how LLMs are deployed and managed.

He warned that the ease with which these models can be repurposed for misinformation, particularly when commands are embedded at a system level rather than given as user prompts, poses a major threat to public health, especially in the context of misinformation campaigns.

The study urges developers and policymakers to strengthen internal safeguards and content moderation mechanisms, especially for LLMs used in health, education, and search contexts.

It also raises important ethical questions about the development of open or semi-open model architectures that can be repurposed at scale.

Without robust oversight, the researchers argue, such systems are likely to be exploited by malicious actors seeking to spread false or harmful content.

Public health at risk

By revealing the technical ease with which state-of-the-art AI systems can be transformed into vectors for health disinformation, the study underscores a growing gap between innovation and accountability in the AI sector.

As AI becomes more deeply embedded in healthcare decision-making, search tools, and everyday digital assistance, the authors call for urgent action to ensure that such technologies do not inadvertently undermine public trust or public health.

Journalists also concerned

The results of this study coincide with conclusions from a recent Muck Rack report, in which more than one-third of surveyed journalists identified misinformation and disinformation as the most serious threat to the future of journalism.

This was followed by concerns about public trust (28 per cent), lack of funding (28 per cent), politicisation and polarisation of journalism (25 per cent), government interference in the press (23 per cent), and understaffing and time pressure (20 per cent).

77 per cent of journalists reported using AI tools in their daily work, with ChatGPT notably being the most used tool (42 per cent), followed by transcription tools (40 per cent) and Grammarly (35 per cent).

A total of 1,515 qualified journalists were part of the survey, which took place between 4 and 30 April 2025. Most of the respondents were based in the United States, with additional representation from the United Kingdom, Canada, and India.

A turning point

Both studies show that, if left unaddressed, vulnerabilities could accelerate an already-growing crisis of confidence in both health systems and the media.

With generative AI now embedded across critical public-facing domains, the ability of democratic societies to distinguish fact from fiction is under unprecedented pressure.

Ensuring the integrity of AI-generated information is no longer just a technical challenge – it is a matter of public trust, political stability, and even health security.

[Edited By Brian Maguire | Euractiv’s Advocacy Lab ]



Source link

Share

Latest Updates

Frequently Asked Questions

Related Articles

Starship successfully completes 11th flight test

WASHINGTON — SpaceX successfully completed the final flight of version 2 of Starship...

Great Scott! Casio’s Back to the Future watch is here and it looks awesome

If you’ve ever dreamed of owning a piece of Back to the Future...

Self-improving language models are becoming reality with MIT's updated SEAL technique

Researchers at the Massachusetts Institute of Technology (MIT) are gaining renewed attention for...

Researchers find that retraining only small parts of AI models can cut costs and prevent forgetting

Enterprises often find that when they fine-tune models, one effective approach to making...
sabung ayam online sabung ayam online judi bola sabung ayam online judi bola Judi Bola Sabung Ayam Online Live Casino Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online sabung ayam online judi bola mahjong ways sabung ayam online judi bola mahjong ways mahjong ways sabung ayam online sv388 Sv388 judi bola judi bola judi bola judi bola JUARA303 Mahjong ways Judi Bola Judi Bola Sabung Ayam Online Live casino mahjong ways 2 sabung ayam online sabung ayam online mahjong ways mahjong ways mahjong ways live casino online sabung ayam online judi bola SV388 SBOBET88 judi bola judi bola judi bola judi bola judi bola https://himakom.fisip.ulm.ac.id/ SABUNG AYAM ONLINE MIX PARLAY SLOT GACOR JUDI BOLA SV388 LIVE CASINO LIVE CASINO ONLINE Judi Bola Online SABUNG AYAM ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE LIVE CASINO ONLINE sabung ayam online Portal SV388 SBOBET88 SABUNG AYAM ONLINE JUDI BOLA ONLINE CASINO ONLINE MAHJONG WAYS 2 sabung ayam online judi bola SABUNG AYAM ONLINE JUDI BOLA ONLINE Sabung Ayam Online JUDI BOLA Sabung Ayam Online JUDI BOLA SV388, WS168 & GA28 SBOBET88 SV388, WS168 & GA28 SBOBET88 SBOBET88 CASINO ONLINE SLOT GACOR Sabung Ayam Online judi bola judi bola judi bola judi bola --indomax77 judi bola online --indomax77 mix parlay --indomax77 situs mix parlay --indomax77 situs parlay --indomax77 sbobet --indomax77 sbobet88 --indomax77 situs bola --indomax77 situs judi bola --indomax77 agen bola --indomax77 agen judi bola --indomax77 agen mix parlay --indomax77 agen parlay --indomax77