← Back to archive
AI's Dark Side: LLMs Master Manipulation and Deepfakes

Image via Every

Monday, January 12, 2026

AI's Dark Side: LLMs Master Manipulation and Deepfakes

We're seeing some wild contrasts in the AI ecosystem right now. On one hand, Claude is quietly becoming healthcare's backbone with HIPAA-ready tools and life sciences integrations, while DeepConff is achieving 99.9% accuracy on reasoning tasks while slashing token usage by nearly 85% (yikes, efficiency). But here's where it gets messy: researchers found that LLMs are equally effective at spreading or debunking conspiracy theories regardless of safety measures, and Indonesia plus Malaysia just blocked Grok over deepfake abuse concerns. Meanwhile, LLMs are literally evolving competitive programs through adversarial self-play in Core War experiments. So we've got AI getting smarter, faster, and more capable in healthcare, but simultaneously more dangerous in the wrong hands. Bold move by these platforms to expand while regulators are circling. If LLMs can convincingly spread OR debunk anything, who really decides the truth?

Top Stories

1
Digital Red Queen: Adversarial Program Evolution in Core War with LLMs

arXiv

LLMs can evolve increasingly effective adversarial programs through self-play in dynamic, competitive environments, suggesting that Red Queen dynamics—continual adaptation to evolving opponents—could improve AI evolution methods beyond static optimization frameworks.

llmevolutionadversarialmulti-agent
2
Large Language Models Can Effectively Convince People to Believe Conspiracies

arXiv

GPT-4o can persuasively promote conspiracy theories nearly as effectively as debunk them, with safety guardrails offering minimal protection—though targeted interventions like accuracy constraints show promise.

llmgpt-4misinformationsafety
3
Indonesia and Malaysia Block Grok Over Deepfake Abuse

TechCrunch

Indonesia and Malaysia blocked Grok over non-consensual sexual deepfakes, triggering coordinated international regulatory scrutiny and marking a major governance challenge for xAI's rapid expansion into sensitive AI capabilities.

xaideepfakeregulationcontent-moderation
4
Claude is Quietly Becoming Healthcare's New Backend

Anthropic

Claude is becoming embedded in healthcare infrastructure through HIPAA-ready tools and connectors that automate administrative burden and accelerate clinical decision-making, enabling faster patient care and drug approvals while major pharma and healthtech companies report substantial productivity improvements.

healthcarellmanthropicagents
5
Deep Think with Confidence

arXiv

DeepConf uses model-internal confidence signals to filter low-quality reasoning during inference, achieving state-of-the-art accuracy on reasoning tasks while dramatically reducing computational costs—critical for making advanced reasoning practical at scale.

llmreasoningtest-time-scalingefficiency

Keep Reading

Enjoyed this issue?

Get daily AI intel delivered to your inbox. No fluff, just the stories that matter.