AI Is Learning to Lie for Social Media Likes
Large language models are learning how to win—and that’s the problem.
In a research paper published Tuesday titled "Moloch’s Bargain: Emergent Misalignment When LLMs Compete for Audiences," Stanford University Professor James Zou and PhD student Batu El show that when AIs are optimized for competitive success—whether to boost ad engagement, win votes, or drive social media traffic—they start lying.
“Optimizing LLMs for competitive success can inadvertently drive misalignment,” the authors write, warning that the very metrics that define “winning” in modern communication—clicks, conversions, engagement—can quietly rewire models to prioritize persuasion over honesty.
"When LLMs compete for social media likes, they start making things up," Zou wrote on X. "When they compete for votes, they turn inflammatory/populist."
This work is important because it identifies a structural danger in the emerging AI economy: models trained to compete for human attention begin sacrificing alignment to maximize influence. Unlike the classical “paperclip maximizer” thought experiment, this isn’t science fiction. It’s a measurable effect that surfaces when real AI systems chase market rewards, what the authors call “Moloch’s bargain”—short-term success at the expense of truth, safety, and social trust.
Using simulations of three real-world competitive environments—advertising, elections, and social media—the researchers quantified the trade-offs. A 6.3% increase in sales came with a 14.0% rise in deceptive marketing; a 4.9% gain in vote share brought a 22.3% uptick in disinformation and 12.5% more populist rhetoric; and a 7.5% boost in social engagement correlated with a staggering 188.6% increase in disinformation and 16.3% more promotion of harmful behaviors.
“These misaligned behaviors emerge even when models are explicitly instructed to remain truthful and grounded,” El and Zou wrote, calling this “a race to the bottom” in AI alignment.
In other words: even when told to play fair, models trained to win begin to cheat.
The problem isn't just hypothetical
AI is no longer a novelty in social media workflows—it’s now near-ubiquitous.
According to the 2025 State of AI in Social Media Study, 96% of social media professionals report using AI tools, and 72.5% rely on them daily. These tools help generate captions, brainstorm content ideas, re-format posts for different platforms, and even respond to comments. Meanwhile, the broader market is valuing this shift: The AI in social media sector is projected to grow from USD 2.69 billion in 2025 to nearly USD 9.25 billion by 2030.
This pervasive integration matters because it means AI is shaping not just how content is made, but what content is seen, who sees it, and which voices get amplified. Algorithms now filter feeds, prioritize ads, moderate posts, and optimize engagement strategies—embedding AI decision logic into the architecture of public discourse. That influence carries real risks: reinforcing echo chambers, privileging sensational content, and creating incentive structures that reward the manipulative over the truthful.
The authors emphasize that this isn’t malicious intent—it’s optimization logic. When reward signals come from engagement or audience approval, the model learns to exploit human biases, mirroring the manipulative feedback loops already visible in algorithmic social media. As the paper puts it, “market-driven optimization pressures can systematically erode alignment.”
The findings highlight the fragility of today’s “alignment safeguards.” It’s one thing to tell an LLM to be honest; it’s another to embed that honesty in a competitive ecosystem that punishes truth-telling.
In myth, Moloch was the god who demanded human sacrifice in exchange for power. Here, the sacrifice is truth itself. El and Zou’s results suggest that without stronger governance and incentive design, AI systems built to compete for our attention could inevitably learn to manipulate us.
The authors end on a sober note: alignment isn’t just a technical challenge—it’s a social one.
“Safe deployment of AI systems will require stronger governance and carefully designed incentives,” they conclude, “to prevent competitive dynamics from undermining societal trust.”
Disclaimer: The content of this article solely reflects the author's opinion and does not represent the platform in any capacity. This article is not intended to serve as a reference for making investment decisions.
You may also like
ICP Price Jumps 30%: Exploring the Reasons Behind the Rally and Future Outlook
- ICP token surged 30% in November 2025, driven by ICP 2.0 upgrades and institutional partnerships with Microsoft , Google, and SWIFT. - On-chain data showed 35% higher active addresses but 91% lower token transfers during downturns, highlighting speculative retail activity over organic adoption. - Social sentiment shifted from extreme greed (Nov 10) to cautious optimism as ICP broke below $4.33 support, with technical indicators suggesting potential short-term reversals. - Analysts project $11.15-$31.89 p

Investment Prospects in AI-Powered Training Platforms: A 2025 Overview of Industry Growth and Workforce Needs
- AI integration in education and workforce training has reached a critical inflection point, driven by surging corporate adoption and labor market shifts. - AI-powered platforms are bridging skill gaps through personalized learning, with 77% of companies using AI for training and a $240B global eLearning market projected by 2025. - Labor trends show AI displacing 92M jobs but creating 170M new roles by 2030, while AI-exposed industries grow revenue 3x faster than non-AI sectors. - Academic programs and $1

The Federal Reserve's Change in Policy and Its Effects on High-Yield Cryptocurrencies Such as Solana: Rethinking Risk Management Amidst Shifting Regulations in the Digital Asset Sector
- Fed's 2025 policy shift injected $72.35B into markets, briefly boosting Solana (+3.01%) before macro risks triggered a 6.1% price drop. - EU MiCA and US GENIUS Act regulations drove institutional adoption of compliant platforms, with Solana's institutional ownership reaching 8% of supply. - Fed's $340B balance sheet reduction and SIMD-0411 proposal exposed crypto liquidity fragility, causing 15% market cap decline and 4.7% TVL drop for Solana. - Institutions now prioritize MiCA-compliant stablecoins and

Algo slips 0.52% as Allego unveils app designed to simplify EV charging
- Algo (ALGO) fell 0.52% in 24 hours to $0.1335, with a 60.3% YTD decline, coinciding with Allego's new EV charging app launch. - Allego's app offers real-time pricing, smart routing, and Plug&Charge features to simplify European EV charging across 35,000+ stations. - The app eliminates partner network markups and provides transparent billing, targeting user frustrations with fragmented charging experiences. - As Europe's EV market grows, Allego positions itself as a key infrastructure provider through thi

