Extended Data Fig. 3: Generating medical misinformation at scale. | Nature Medicine

Extended Data Fig. 3: Generating medical misinformation at scale.

From: Medical large language models are vulnerable to data-poisoning attacks

Extended Data Fig. 3

Prompt engineering is used to bypass OpenAI’s guardrails and generate harmful medical articles using the GPT-3.5-turbo API. The articles are inserted into websites as invisible HTML text tags. Tags may include the ‘hidden’ style, font size 0, opacity 0, and other tags that conceal malicious text. Invisible misinformation is uploaded to coincide with scheduled Common Crawl data dumps, entering the repository while evading detection.

Back to article page