In colonial India during the 1800s, the British government offered a bounty for every dead cobra delivered to authorities. The intent was to reduce the snake population plaguing Delhi. The result was predictable to anyone who understands incentives and unpredictable to those who designed the policy: enterprising locals began breeding cobras specifically to collect the reward. When the government discovered the scheme and cancelled the bounty program, breeders released their now-worthless snakes into the wild, leaving Delhi with more cobras than before the intervention began.
This is the cobra effect -- a perverse incentive where a policy designed to solve a problem makes it measurably worse by creating new incentives that work against the original goal. The term was popularized by German economist Horst Siebert in his 2001 book Der Kobra-Effekt and has since become a standard concept in behavioral economics and policy analysis.
Social media platforms have been running a version of this experiment at global scale for over fifteen years. They optimized their products for engagement -- likes, shares, comments, time spent -- as a proxy for user value. The logic seemed sound: if someone is engaging, they must be getting value. But engagement and value diverged, and the consequences are now visible in public discourse, journalism, politics, and mental health data worldwide.
"When a measure becomes a target, it ceases to be a good measure." -- Charles Goodhart, in "Problems of Monetary Management: The U.K. Experience" (1975)
Defining the Problem: Engagement as a Proxy Gone Wrong
Engagement metrics are quantitative measures of user interaction with digital content -- likes, shares, comments, reactions, time spent on page, session duration, and return frequency. They were originally designed as proxy measures for a quality that is much harder to quantify: whether users are getting genuine value from a platform.
The distinction between a proxy and the thing it represents is critical. A thermometer is a proxy for how warm you feel. Your bank balance is a proxy for financial security. School grades are a proxy for learning. Proxies are useful precisely because they make unmeasurable things measurable. But they break when they are treated as the goal itself rather than an indicator of it.
This is Goodhart's Law in action -- one of the most important concepts in measurement theory. When engagement metrics shifted from being an indicator that platforms monitored to being the objective that platforms optimized for, the metrics attracted content specifically engineered to game them. The signal became noise.
A 2023 report by the Center for Humane Technology estimated that recommendation algorithms powered by engagement optimization influence the content consumption of over 4 billion people daily -- roughly half the world's population. The scale of the cobra effect in social media is unprecedented in the history of perverse incentives.
Why Negative Emotions Win the Engagement Game
The relationship between emotion and online sharing behavior has been studied extensively, and the findings are consistent across platforms, cultures, and time periods.
The Berger-Milkman Virality Research
A landmark 2011 study by Jonah Berger and Katherine Milkman, published in the Journal of Marketing Research, analyzed 7,000 articles from The New York Times over a three-month period to determine what made content go viral. Their findings established a principle that has held up across subsequent research:
Content inducing high-arousal emotions -- awe, anger, anxiety, excitement -- was far more likely to be widely shared than content inducing low-arousal emotions like sadness or contentment. The emotional intensity mattered more than the emotional valence (positive vs. negative). But anger had a structural advantage over awe because it was easier to trigger, faster to form, and more reliably produced sharing behavior.
The Mechanics of Outrage
Anger has what researchers call an "almost unfair structural advantage" in the attention economy:
- Speed of formation: A provocative headline triggers anger before the prefrontal cortex fully processes context. Neuroscientist Antonio Damasio's somatic marker hypothesis (1994) explains this: emotional responses precede and guide cognitive evaluation, not the reverse.
- Low sharing threshold: Outrage-sharing requires almost no effort or comprehension. A 2023 study by researchers at NYU's Center for Social Media and Politics found that users sharing outrage content spent an average of only 4.2 seconds on the article before sharing -- far below the time needed to read even the headline and subheadline.
- Social signaling value: Sharing outrage signals group membership, moral positioning, and ideological identity. It communicates "I am the kind of person who cares about this" without requiring the sharer to do anything substantive about the issue.
- Algorithmic amplification: Every share generates engagement signals -- more impressions, more reactions, more comments -- which cause the recommendation algorithm to surface the content to more users, creating a feedback loop.
The MIT False News Study
A groundbreaking 2018 study by Soroush Vosoughi, Deb Roy, and Sinan Aral at MIT, published in Science, analyzed the spread of 126,000 verified true and false news stories shared by approximately 3 million people on Twitter between 2006 and 2017. Their findings were stark:
| Metric | True News | False News |
|---|---|---|
| Average cascade depth | Rarely exceeded 10 retweet chains | Routinely exceeded 19 retweet chains |
| Speed to reach 1,500 users | 6x slower than false news | 6x faster than true news |
| Reach breadth | Limited | 35% broader dissemination |
| Emotional reactions | Moderate | Significantly higher surprise, fear, and disgust |
False news spread faster, farther, and more broadly than true news -- and the researchers controlled for the influence of bots, finding that humans, not automated accounts, were primarily responsible for the differential spread. False news was more novel, and novelty triggers the high-arousal emotions (surprise, curiosity, outrage) that drive sharing behavior.
"Falsehood diffused significantly farther, faster, deeper, and more broadly than the truth in all categories of information." -- Vosoughi, Roy, and Aral, "The Spread of True and False News Online," Science (2018)
How Platform Algorithms Learned to Serve Outrage
Recommendation algorithms are not programmed to promote outrage. They are programmed to maximize a measurable proxy (engagement) for an unmeasurable goal (user value). The outrage is an emergent property of the optimization -- a cobra that the bounty system bred.
The YouTube Radicalization Funnel
YouTube's recommendation algorithm, designed to maximize watch time, developed a well-documented tendency to serve progressively more extreme content. A viewer who watched a mainstream political commentary video would be recommended a more partisan one, then a more extreme one, because each incremental step toward extremity slightly increased the probability of continued watching.
Guillaume Chaslot, a former YouTube algorithm engineer who later founded AlgoTransparency, documented this pattern systematically beginning in 2016. His research, along with subsequent work by Brendan Nyhan at Dartmouth and Kevin Roose at The New York Times, found that the algorithm systematically overrepresented fringe content relative to its actual production volume. A conspiracy theory video with 10,000 views might be recommended as frequently as a mainstream news video with 10 million views, because the conspiracy video had higher per-user engagement metrics.
The algorithm had no concept of accuracy, credibility, or long-term user welfare. It had one objective: maximize the probability that the current viewer would watch another video. And extreme content, by triggering strong emotional responses, was measurably better at achieving that objective.
YouTube began modifying its algorithm in January 2019, reducing recommendations of what it internally classified as "borderline content" -- material that approached but did not cross the platform's policy lines. The company reported that this intervention reduced views on such content by approximately 70 percent. But the intervention required YouTube to accept a reduction in a metric it had spent years optimizing -- something platforms are structurally reluctant to do because engagement directly correlates with advertising revenue.
Facebook's "Meaningful Social Interactions" Miscalculation
In January 2018, Facebook CEO Mark Zuckerberg announced a major algorithmic shift: the News Feed would deprioritize passive content consumption (watching videos, reading articles) and instead emphasize "meaningful social interactions" -- content that generated comments and shares from friends and family.
The logic was superficially reasonable. A post you comment on is one you care about. A post you share is one that matters to you. Prioritizing these interactions should, in theory, make the experience feel more personal and less like a passive content feed.
The results were catastrophic for the quality of public discourse. As The Wall Street Journal reported in its 2021 "Facebook Files" series (based on internal documents leaked by whistleblower Frances Haugen), Facebook's own internal research showed that the "meaningful interactions" update caused a significant increase in the virality of misinformation and politically divisive content. Comments and shares are not neutral proxies for meaning -- they are generated at disproportionately high rates by content that provokes, offends, and frightens.
| Content Type | Average Reactions | Average Comments | Average Shares |
|---|---|---|---|
| Calm, informational posts | Low | Low | Low |
| Inspiring or uplifting posts | Medium | Low | Medium |
| Partisan political content | High | High | High |
| Outrage-inducing content | Very High | Very High | Very High |
| Misinformation with emotional framing | Very High | Very High | Very High |
This pattern was documented both by Facebook's internal data science team and by external researchers including Laura Edelson at NYU's Online Political Ads Transparency Project. The content that maximized "meaningful social interactions" was systematically different from content that left users feeling informed, connected, or satisfied.
Facebook's internal researchers reportedly recommended changes to mitigate these effects. According to Haugen's testimony before the U.S. Senate in October 2021, those recommendations were largely not implemented because they would have reduced engagement metrics.
The News Industry Caught in the Middle
The consequences of engagement optimization have been most acute and most damaging for journalism. Throughout the 2010s, news publishers became increasingly dependent on Facebook and Google for audience distribution. At the peak of this dependency, some publishers received more than 40 percent of their web traffic via Facebook referrals, according to data from Parse.ly (2017).
This dependency created a powerful incentive to produce content that performed well in algorithmic feeds. Outlets began optimizing headlines for emotional arousal rather than informational accuracy. The phenomenon of clickbait -- headlines that exploit curiosity gaps, trigger outrage, or make exaggerated promises -- became so pervasive that it transformed reader expectations across the entire industry.
Upworthy, founded in 2012, pioneered a systematic approach to headline optimization that the industry called "curiosity gap" headlines: "You Won't Believe What Happened When..." and "This Video Changed Everything I Thought I Knew About..." The approach was enormously effective at generating clicks and shares. At its peak in 2013, Upworthy was reaching 90 million unique visitors per month -- traffic driven almost entirely by Facebook's algorithm.
When Facebook's 2018 algorithm change deprioritized publisher content in favor of personal posts, publishers who had invested in platform-dependent distribution strategies were devastated. LittleThings, a digital publisher that had built its entire business on Facebook video, shut down entirely in February 2018, laying off all 100 employees. The company's CEO attributed the closure directly to the algorithm change.
The irony is sharp: publishers had debased their editorial standards to win an engagement game, and then the engagement game changed, leaving them with neither traffic nor credibility. Those that survived best were publishers that had invested in direct audience relationships -- email newsletters, podcasts, membership programs -- that did not depend on algorithmic amplification.
Goodhart's Law in Practice
The engagement metrics problem in journalism is a textbook illustration of Goodhart's Law and its cousin, Campbell's Law (formulated by social scientist Donald Campbell in 1979): "The more any quantitative social indicator is used for social decision-making, the more subject it will be to corruption pressures and the more apt it will be to distort and corrupt the social processes it is intended to monitor."
Engagement was once a reasonable signal of content value. A high engagement rate on a thoughtful article genuinely indicated that readers found it worthwhile. Once engagement became the target that editorial, distribution, and business decisions optimized for, it attracted content specifically engineered to game the metric -- outrage bait, misinformation, and emotional manipulation -- and ceased to be a reliable indicator of the value it had originally proxied.
Time Spent Is Not Time Well Spent
Beyond individual posts and articles, the engagement metrics problem extends to how platforms measure overall product success. The metric of time-on-site or daily active users (DAU) has driven platform product decisions for over a decade. More time spent equals more ad impressions equals more revenue. The incentive is to keep users scrolling as long as possible.
Research on the actual effects of extended social media consumption complicates this equation substantially.
A 2019 study published in the Journal of Social and Clinical Psychology by researchers at the University of Pennsylvania randomly assigned 143 undergraduates to either continue their normal social media use or limit it to 30 minutes per day for three weeks. The group that limited their use showed significant reductions in loneliness and depression compared to the control group, with the strongest effects among participants who had the highest depression scores at baseline.
A 2020 study from Harvard's T.H. Chan School of Public Health found that passive social media consumption -- scrolling through feeds without actively posting, commenting, or connecting with others -- was associated with decreased well-being, while active use (direct messaging, posting, commenting) showed mixed or slightly positive effects. The distinction matters because platform design overwhelmingly encourages passive consumption through infinite scroll, autoplay, and algorithmic feeds.
A 2021 large-scale study by Andrew Przybylski and Amy Orben at the Oxford Internet Institute, using data from 430,000 adolescents across multiple countries, found that the relationship between digital technology use and well-being was real but extremely small -- comparable in magnitude to the effect of wearing glasses or eating potatoes. The researchers argued that the public narrative about social media and mental health was often more alarmist than the evidence supported, while acknowledging that specific features (algorithmic feeds, infinite scroll, social comparison mechanisms) likely had larger effects than overall screen time.
Time-on-site is not equivalent to satisfaction. A person kept awake at 2am by an anxiety-inducing feed is generating engagement metrics while experiencing harm. A person who spends 15 minutes reading a thoughtful article and then closes the app feeling informed generates less engagement than someone who hate-scrolls for two hours.
Tristan Harris, a former Google design ethicist and co-founder of the Center for Humane Technology, popularized the phrase "time well spent" to describe a different design philosophy. The question, Harris argues, should not be whether users stayed longer but whether they left feeling better off than when they arrived. This framing explicitly challenges the assumption that more engagement equals more value.
The Self-Perpetuating Content Machine
The cobra effect extends far beyond platforms. It has reshaped the incentive structures of everyone who produces content for the internet.
Individual creators on YouTube, TikTok, and Instagram learn quickly through direct feedback which content types outperform others. A video expressing outrage about a public figure will typically generate 3-5x the engagement of a calm, nuanced analysis of the same topic. Over time, creators who want to grow their audiences are trained -- by the feedback loop of metrics and algorithms -- to produce increasingly emotionally provocative content.
A 2022 study by researchers at Brown University and the University of Cambridge, published in PNAS, found that social media users learned to express more outrage over time because the platform's reward structure (likes, shares, follower growth) reinforced moral outrage expression. The researchers called this "incentivized moral outrage" and demonstrated that the pattern held even for users who did not initially post outrage-laden content -- the platform trained them to do so.
News organizations face internal pressure from digital editors who track real-time analytics dashboards. A headline that generates thousands of shares within the first hour will be promoted on the homepage; a nuanced, carefully reported investigation that generates thoughtful but fewer shares may get buried. The analytics dashboard becomes the de facto editor, overriding editorial judgment about what stories matter most.
Political actors across the ideological spectrum have learned that controversy, provocation, and conflict generate more attention than policy substance. A 2021 analysis by the Pew Research Center found that U.S. congressional members who posted more negative, more emotional content on social media received significantly more engagement than those who posted substantive policy content. The incentive structure rewards performative conflict over governance.
What Better Metrics Look Like
The good news is that alternatives to raw engagement metrics exist and have been tested. Several platforms and researchers have proposed -- and in some cases implemented -- metrics that attempt to capture value rather than mere attention.
Satisfaction Surveys
Platforms can randomly sample users after content exposure and ask whether they feel better or worse informed, whether the content was worth their time, and whether they would recommend the platform to others. YouTube has used variations of this approach since 2019 to supplement engagement data in training its recommendation algorithm. Internal reports suggest that incorporating satisfaction signals reduced recommendations of "regrettable" content -- content that users clicked on but later wished they had not.
Saves and Bookmarks
When a user saves or bookmarks content, they are signaling intent to return to it -- a signal of perceived value that does not reward purely reactive, emotion-driven content. Instagram and Pinterest have both elevated saves as a ranking signal, and several content creators have reported that optimizing for saves (by producing useful, reference-worthy content) produces more sustainable audience growth than optimizing for likes or comments.
Informed Sharing
Twitter/X introduced a prompt in 2020 asking users if they wanted to read an article before sharing it. The platform reported a 40 percent increase in users opening articles before sharing. This simple friction point -- a single prompt -- significantly reduced impulsive sharing of content that users had not actually read. Facebook implemented a similar feature in 2021 for news articles.
Diverse Information Exposure
Researchers at the MIT Media Lab developed a metric called the "Exposure Diversity Score" that tracks whether a user's feed exposes them to different perspectives, topics, and sources. Platforms designed around this metric would penalize echo chambers rather than reward them -- a direct counter to the algorithmic tendency to serve users more of what they have already engaged with.
| Metric Type | What It Measures | Perverse Incentive | Better Alternative |
|---|---|---|---|
| Likes | Immediate emotional reaction | Approval-seeking, lowest-common-denominator content | Satisfaction survey, "was this valuable?" prompt |
| Shares | Viral amplification | Outrage and controversy | Bookmarks, saves, informed sharing with friction |
| Comments | Volume of response | Conflict, debate, hostility | Proportion of constructive replies, reply quality scoring |
| Time on site | Session length | Infinite scroll, anxiety-inducing design | Session quality rating, "time well spent" assessment |
| Daily active users | Habitual return | FOMO-inducing notifications, withdrawal loops | Weekly active users with expressed intent to return |
| Click-through rate | Interest signal | Clickbait, curiosity-gap manipulation | Click-through with completion rate |
Platform Accountability and the Regulatory Horizon
Regulation of social media algorithms is an active and accelerating area of policy development in the European Union, the United States, and the United Kingdom.
The EU's Digital Services Act (2024)
The Digital Services Act (DSA), which became fully applicable to very large online platforms (those with more than 45 million EU users) in February 2024, requires platforms to:
- Conduct and publish risk assessments of their recommendation systems, including effects on civic discourse, election integrity, and mental health
- Provide users with at least one feed option that is not based on profiling (i.e., a chronological or non-algorithmic feed)
- Give researchers access to data necessary to study the platform's societal effects
- Maintain transparency reports on content moderation decisions and algorithmic amplification
This represents a fundamental shift: platforms can no longer treat engagement optimization as a purely internal business decision. They are now required to demonstrate awareness of the harms their optimization choices may create and to take measurable steps to mitigate them.
U.S. Legislative Developments
In the United States, the Kids Online Safety Act (KOSA) and various state-level laws (notably California's Age-Appropriate Design Code Act) have targeted the specific application of engagement-maximizing design to minors. The legislative argument -- supported by internal research leaked from Facebook/Meta in 2021 -- is that platforms had evidence their engagement-optimized products caused harm to teenagers and failed to act on that knowledge.
Frances Haugen, the Facebook whistleblower, testified before the U.S. Senate Commerce Committee in October 2021 that Facebook's own researchers had concluded that Instagram was making body image issues worse for one in three teenage girls, and that the company had not implemented recommended changes because doing so would reduce engagement metrics.
"The platform knew its algorithm was amplifying content that made teenage girls feel worse about their bodies. It knew this from its own research. And the metric was still clicks." -- Frances Haugen, testimony before the U.S. Senate, October 2021
Lessons for Content Creators and Publishers
If you create content for the web, understanding the cobra effect in engagement metrics is not just intellectually interesting -- it is strategically essential.
Chasing engagement metrics without questioning what they measure is building on sand. Platforms change their algorithms with little warning and no obligation to protect publishers who built businesses around them. LittleThings, Upworthy, and dozens of other publishers learned this lesson at the cost of their businesses. Publishers who invested in sustainable audience relationships -- email lists, podcasts, membership models, direct subscriptions -- weathered algorithm changes far better than those dependent on platform amplification.
Your audience's long-term trust is worth more than any individual post's engagement. A reader who returns to your publication because they consistently find it useful is more valuable than a thousand one-time visitors driven by an outrage-bait headline. The Information, Stratechery, and The Economist have all built sustainable digital businesses by optimizing for reader trust and willingness to pay, not for social media engagement.
Measure what you actually care about. If your goal is to inform, measure comprehension and return rates. If your goal is to inspire action, measure downstream behavior. If your goal is to build community, measure the quality and constructiveness of interactions, not just their volume. The metric you choose to optimize will shape the content you produce -- so choose the metric that aligns with the outcome you actually want.
The Deeper Lesson: Metrics Are Models, Not Reality
The engagement metrics problem is, at its core, a specific instance of a universal challenge: the gap between what we can measure and what we actually value. This gap exists in education (test scores vs. learning), in healthcare (procedure volume vs. patient outcomes), in policing (arrest rates vs. public safety), and in virtually every domain where quantitative metrics are used to manage complex systems.
The sociologist William Bruce Cameron wrote in 1963: "Not everything that can be counted counts, and not everything that counts can be counted." This observation, often misattributed to Einstein, captures the fundamental tension. Engagement can be counted. User value, civic health, informed citizenship, and genuine human connection cannot be counted with the same precision. When organizations optimize for what can be counted at the expense of what cannot, the results are predictably perverse.
The cobras are everywhere. The question is whether we are still paying the bounty -- and whether we have the wisdom to redesign the incentive before the snakes outnumber us.
References and Further Reading
- Berger, J., & Milkman, K. L. "What Makes Online Content Viral?" Journal of Marketing Research, vol. 49, no. 2, 2012, pp. 192-205.
- Vosoughi, S., Roy, D., & Aral, S. "The Spread of True and False News Online." Science, vol. 359, no. 6380, 2018, pp. 1146-1151.
- Siebert, H. Der Kobra-Effekt: Wie man Irrwege der Wirtschaftspolitik vermeidet. Deutsche Verlags-Anstalt, 2001.
- Goodhart, C. A. E. "Problems of Monetary Management: The U.K. Experience." Papers in Monetary Economics, Reserve Bank of Australia, 1975.
- Campbell, D. T. "Assessing the Impact of Planned Social Change." Evaluation and Program Planning, vol. 2, no. 1, 1979, pp. 67-90.
- Hunt Allcott, et al. "The Welfare Effects of Social Media." American Economic Review, vol. 110, no. 3, 2020, pp. 629-676.
- Brady, W. J., et al. "How Social Learning Amplifies Moral Outrage Expression in Online Social Networks." Science Advances, vol. 7, no. 33, 2021.
- Przybylski, A. K., & Orben, A. "Screens, Teens, and Psychological Well-Being: Evidence From Three Time-Use-Diary Studies." Psychological Science, vol. 30, no. 5, 2019.
- Haidt, J. The Anxious Generation: How the Great Rewiring of Childhood Is Causing an Epidemic of Mental Illness. Penguin Press, 2024.
- Harris, T., & Raskin, A. "Ledger of Harms." Center for Humane Technology, 2023. humanetech.com.
- Haugen, F. Testimony before the U.S. Senate Committee on Commerce, Science, and Transportation. October 5, 2021.
- European Commission. "Digital Services Act." Official Journal of the European Union, 2022.
- Chaslot, G. "How Algorithms Can Learn to Discredit the Media." Medium / AlgoTransparency, 2018.
- Pew Research Center. "Social Media and Political Engagement." pewresearch.org, 2021.
- Roose, K. "The Making of a YouTube Radical." The New York Times, June 8, 2019.
- Damasio, A. Descartes' Error: Emotion, Reason, and the Human Brain. Putnam, 1994.
Frequently Asked Questions
What is the cobra effect in social media?
The cobra effect in social media describes how optimizing for a metric (like engagement or likes) creates incentives that produce the opposite of the intended outcome. Platforms designed to connect people instead amplify conflict and outrage because negative, divisive content generates more clicks, shares, and comments than calm or nuanced posts.
Why does outrage drive social media engagement?
Outrage triggers strong emotional responses that lower the threshold for sharing and commenting. Research from the University of Pennsylvania found that each moral-emotional word in a tweet increased its retweet rate by approximately 20%. Platforms reward this behavior algorithmically because high engagement signals relevance, creating a feedback loop that favors inflammatory content.
How did Facebook's algorithm changes affect news publishers?
In 2018, Facebook shifted its News Feed algorithm to prioritize 'meaningful social interactions' — comments and shares over passive likes. News publishers saw organic reach drop by 50-70% for informational content, while emotionally charged, partisan stories performed better. Many publishers were forced to adopt more sensational framing to survive on the platform.
What are better alternatives to engagement as a social media metric?
Healthier alternatives include 'satisfied' or 'inspired' reactions rather than just likes, return visitor rates, saves and bookmarks (which indicate intent to revisit), and surveys measuring whether content made users feel informed or uplifted. Some platforms experiment with hiding public like counts to reduce social comparison pressure.
Can platforms be redesigned to reduce perverse incentives?
Yes. Research by the Center for Humane Technology and academics like Renee DiResta suggests interventions including friction before resharing (prompting users to read before sharing), removing real-time like counts, and ranking feeds by diverse information exposure rather than pure engagement. Twitter's Birdwatch and Community Notes program represents one attempt to add accuracy as a dimension alongside engagement.