Register Domain Transfer Domain Submit a Ticket
Society

Social Media and AI: How Platforms Weaponize Machine Learning to Capture, Shape, and Radicalize

March 2026 18 min read EasWrk Research

The Engagement Engine

The modern social media platform is not, at its core, a communications tool. It is a prediction machine: an enormously sophisticated system designed to forecast, with increasing precision, which piece of content will keep you scrolling for the next thirty seconds. Every tap, pause, like, and share feeds a continuously updated model of your psychological profile. That model has one purpose: to maximize the amount of time you spend inside the platform, because time spent translates directly into advertising revenue.

Engagement maximization is not an incidental feature of social media design. It is the central engineering objective. When Facebook engineers describe optimizing for "meaningful social interaction," or when TikTok's recommendation system is characterized as connecting users with content they love, they are describing engagement maximization in softer language. The metric that matters is time-on-platform, and every design choice, from the infinite scroll to the variable-ratio reinforcement schedule of the notification system, is oriented toward that metric.

The machine learning systems that power this engagement optimization are genuinely impressive technical achievements. TikTok's "For You" algorithm reportedly requires as few as a handful of video completions before it can predict your preferences with unsettling accuracy. Meta's News Feed ranking model processes thousands of signals per impression. YouTube's recommendation system, as documented by researchers and engineers who worked on it, was explicitly optimized for watch time, a decision with profound consequences that its designers did not fully anticipate.

The mechanism that makes this so powerful is rooted in basic behavioral psychology. Variable-ratio reinforcement underpins the social media notification system. It is the same principle that makes slot machines addictive. You do not know when your next post will get engagement, when someone you follow will post something interesting, or when a notification will arrive. This uncertainty creates a compulsive checking behavior that is difficult to resist even when users consciously want to disengage. The platforms did not invent this dynamic; they codified it in software at industrial scale.

The economic logic is straightforward. Social media companies earn their revenue primarily from advertising. Advertising rates are determined by audience size and engagement levels. An engaged user who spends three hours per day on a platform is worth far more to advertisers than a user who checks in briefly and leaves. This creates a direct financial incentive to engineer maximum engagement, independent of any consideration of whether that engagement is positive, healthy, or accurate. A 2018 study by MIT Media Lab researchers Soroush Vosoughi, Deb Roy, and Sinan Aral, published in Science, found that false news spreads significantly faster than true news on Twitter, reaching more people more quickly across virtually every category of information (Vosoughi, Roy, & Aral, 2018). The study analyzed 126,000 rumor cascades spread by approximately 3 million people over more than a decade. The researchers found that false news was 70 percent more likely to be retweeted than true news, and reached its first 1,500 people six times faster. Crucially, this effect was not driven by bots. Human beings were the primary vectors of misinformation spread.

Why does false information spread faster? Because it tends to be more emotionally arousing: more novel, more surprising, more outrage-inducing. Emotional arousal is precisely what the engagement algorithm is selecting for. The platforms are not deliberately amplifying misinformation. They are amplifying emotion, and misinformation tends to be more emotionally potent than accurate reporting. The result is structurally equivalent, but the mechanism matters for how we think about accountability and remedy.

The engagement engine does not only select for misinformation. It selects for conflict, for moral outrage, for identity-affirming content of all kinds. Research by William Brady and colleagues at NYU found that the use of moral-emotional language in tweets increased retweet rates by approximately 20 percent for each moral-emotional word added. The algorithm does not know the difference between accurate and inaccurate, between healthy and harmful. It knows only what you click on, what you watch to completion, and what makes you come back.

Algorithmic Identity and the Filter Bubble

In 2011, internet activist Eli Pariser published "The Filter Bubble," a book that introduced the concept of algorithmic personalization as a form of epistemic segregation (Pariser, 2011). Pariser argued that as platforms learned individual user preferences and tailored content accordingly, they would create self-reinforcing information environments in which users were progressively insulated from perspectives that challenged their worldview. The filter bubble, in Pariser's framing, was not merely a political phenomenon. It was a fundamental restructuring of how individuals encountered reality.

More than a decade after that diagnosis, the empirical picture is both more complex and more troubling than Pariser's original formulation suggested. The filter bubble concept has been challenged on several empirical grounds. Some researchers have found that algorithmic personalization actually exposes users to somewhat more diverse content than they would encounter through purely self-directed browsing, because recommendation systems optimize for engagement, and mild novelty is engaging. A 2019 study by Guess and colleagues found that fake news consumption was concentrated among a small subset of users, primarily older conservatives in the United States, rather than being a broadly distributed phenomenon produced by filter bubbles per se.

But these challenges to the filter bubble hypothesis do not rehabilitate the platforms. They merely complicate the picture in ways that are arguably worse. The problem is not only that people are exposed to homogeneous content. The algorithmic identity formation process is more active and more manipulative than simple exposure filtering suggests. The algorithm does not merely reflect your preferences back at you. It shapes those preferences over time, nudging you toward more extreme, more emotionally engaging versions of the content you already consume.

Identity shaping is deeply intertwined with commercial logic. Platforms profit from the creation of stable, predictable audience segments. A user with a clearly defined identity is a more valuable advertising target than an undifferentiated one. A passionate supporter of a particular political candidate, a devoted follower of a particular dietary philosophy, an enthusiastic consumer of a particular genre of outrage content: all are easier to sell to. The platform has a financial interest in helping you become more certain about who you are, even if what you are becoming is a more polarized, more tribal, more advertisable version of your previous self.

Zeynep Tufekci, a sociologist and technology critic at Columbia University, has written extensively about the ways in which algorithmic amplification differs from human editorial judgment. In a landmark 2018 piece in The New York Times, Tufekci described YouTube's recommendation algorithm as "one of the most powerful radicalizing instruments of the 21st century" (Tufekci, 2018). Her observation was based on a pattern she noticed while researching YouTube: the platform's recommendations reliably pushed viewers toward more extreme content regardless of their starting point. Someone watching a mainstream political speech would be directed toward more partisan content; someone watching fitness videos would be nudged toward more extreme dietary philosophies; someone watching videos about vegetarianism might find themselves three clicks away from radical animal liberation content.

The "rabbit hole" effect is not accidental. It reflects the optimization logic of a recommendation system trained on watch time. Extreme content tends to be more engaging than moderate content, and the algorithm, with no understanding of the social consequences of radicalization, simply follows the engagement signal. The result is a recommendation architecture that systematically rewards escalation.

The filter bubble and identity formation dynamics are particularly powerful in adolescence, when identity formation is developmentally central and self-concept is most fluid and most susceptible to social influence. A teenager whose social life is substantially organized around Instagram or TikTok is not merely consuming content. They are being shaped by a feedback loop in which their emerging sense of self is continuously reflected, amplified, and monetized by systems designed to maximize their engagement. The implications of this for mental health are significant and will be addressed in detail below.

The Radicalization Pipeline

The most consequential documented harm of algorithmic amplification is the radicalization pipeline: the systematic movement of individuals toward more extreme ideological positions through iterative content recommendation. The pipeline metaphor is apt because it describes a process that has consistent stages, moves in a predictable direction, and is powered by an external force, the recommendation algorithm, rather than purely by the user's own choices.

The mechanisms of radicalization via social media are well-documented. Guillaume Chaslot, a former YouTube engineer, spent years after leaving the company documenting how the recommendation algorithm systematically amplified extreme and conspiratorial content. His analysis found that the algorithm was significantly more likely to recommend conspiratorial, extreme, and outrage-inducing content than its share of available content would suggest. This was not because YouTube's engineers wanted to radicalize their users. It was because this content was more engaging, and engagement was what the algorithm was optimizing for.

Frances Haugen, the Facebook data scientist who became one of the most consequential whistleblowers in the history of Silicon Valley, provided detailed documentation of these dynamics in her October 2021 congressional testimony. Haugen had spent nearly two years in Facebook's Civic Integrity team before leaving and providing thousands of pages of internal documents to regulators and journalists. Her testimony established several critical points with documentary evidence: that Facebook's own researchers had identified the engagement-amplification of harmful content as a systemic problem; that proposed fixes had been internally rejected or limited because they reduced engagement metrics; and that the company's leadership was aware that its algorithm was contributing to polarization, radicalization, and harm, and had chosen not to act substantively on that knowledge (Haugen, 2021).

The internal Facebook research that Haugen disclosed included a study titled "Carol's Journey to QAnon," which documented how a fictional middle-aged American woman following mainstream conservative content would, within two weeks, be served QAnon conspiracy content through the recommendation algorithm without ever actively seeking it. The research was internal. It was not published. And the platform's response to its own findings was not to fix the underlying recommendation logic but to implement superficial adjustments that preserved engagement while offering a degree of plausible deniability.

Radicalization through social media is not limited to political extremism. The same mechanisms operate in fitness communities, dietary subcultures, parenting forums, and religious spaces. Research on pro-eating-disorder communities on Instagram, for example, has documented how the recommendation algorithm connects users seeking general wellness content to increasingly extreme communities promoting disordered eating. Instagram's own internal research, leaked in the Haugen documents, found that the platform made body image issues worse for a substantial proportion of teenage girls who reported experiencing them, a finding the company had for years publicly denied.

The radicalization pipeline is also platform-specific in its character. Facebook radicalizes primarily through group dynamics and shareable content; YouTube through autoplay and recommendation chains; TikTok through compressed exposure to algorithmically curated short-form video. Twitter/X operates through a different but related mechanism: the amplification of moral outrage through retweets, quote-tweets, and trending topics. Each platform's architecture shapes the form that radicalization takes, but the underlying dynamic is common to all of them: engagement optimization rewards escalation.

Understanding radicalization as a systemic product of algorithmic design, rather than a phenomenon produced solely by the malicious intent of individual bad actors, has important implications for regulation and remedy. If radicalization were primarily a content problem, the solution would be content moderation. But if it is fundamentally an algorithmic design problem, if the architecture of recommendation itself is the engine of radicalization, then content moderation alone is structurally insufficient. Moderating content while preserving the engagement-optimizing recommendation algorithm is analogous to cleaning up oil spills while continuing to operate the pipeline that produces them.

Teen Mental Health: The Evidence

No aspect of the social media and AI debate has generated more public attention, or more scientific controversy, than the question of the relationship between social media use and adolescent mental health. The evidence base is large, contested, and politically charged, but a careful reading of the literature supports conclusions that are both specific and serious.

Jean Twenge, a psychology professor at San Diego State University, was among the first researchers to document population-level shifts in adolescent mental health that correlated with smartphone adoption. In her 2017 book "iGen," Twenge presented data showing sharp increases in depression, anxiety, loneliness, and suicide ideation among American teenagers beginning around 2012, the year that smartphone ownership became majority among this age group (Twenge, 2017). The correlation was notable not only for its magnitude but for its specificity: the deterioration in mental health indicators was significantly sharper among girls than boys, and it was most pronounced for measures of mental health that plausibly connected to social comparison and appearance, areas directly relevant to Instagram's content ecology.

Jonathan Haidt, a social psychologist at NYU's Stern School of Business, has become the most prominent academic voice arguing that social media causes significant mental health harm in adolescents, particularly girls. In "The Anxious Generation," published in 2024, Haidt synthesizes a large body of research to argue that the shift to phone-based childhood has been catastrophic for teen mental health, driven primarily by social comparison, cyberbullying, sleep disruption, and the displacement of in-person interaction (Haidt, 2024). In earlier work with colleagues, Haidt documented correlations between social media use and depression, anxiety, and loneliness across multiple countries and datasets.

The causal mechanism Haidt and Twenge identify is social comparison: adolescents, especially girls, measure their appearance, social status, and life quality against the curated highlight reels of peers and influencers. Instagram in particular has been identified as especially harmful in this regard, because its visual format makes appearance comparison central and because its influencer culture creates an enormous supply of aspirational content against which ordinary teenagers inevitably measure themselves unfavorably.

Facebook's own internal research, disclosed by Haugen, found that 32 percent of teenage girls said that when they felt bad about their bodies, Instagram made them feel worse. Among teenagers who reported suicidal thoughts, 13 percent of British users and 6 percent of American users traced the impulse directly to Instagram. These findings were produced by Facebook's own researchers, using Facebook's own data, and they were not made public.

The scientific debate around social media and mental health is, however, genuinely contested in important ways. Haidt and Twenge's correlation-based arguments have been challenged by several researchers who argue that the effect sizes found in most studies are small, that correlational evidence cannot establish causation, and that the deterioration in teen mental health has multiple potential causes beyond social media. A 2020 paper by Haidt and Nick Allen in Nature reviewed the meta-analytic literature and found that while negative effects of social media on mental health are real, their magnitude has often been overstated in public discourse (Haidt & Allen, 2020). The relationship is real but more modest than some of its most prominent advocates have suggested.

What the evidence does support, fairly robustly, is that heavy social media use is associated with worse mental health outcomes for adolescents, particularly girls, and that this relationship is mediated by social comparison, cyberbullying, and sleep disruption. What remains contested is the precise causal mechanism, the magnitude of the effect, and how much of the observed deterioration in teen mental health is attributable to social media as opposed to other social and economic factors. The policy implications of these debates are significant, but the uncertainty in the scientific literature does not absolve platforms of responsibility for design choices that their own internal research indicates cause harm.

Surveillance Capitalism and Data Harvesting

In 2019, Harvard Business School professor Shoshana Zuboff published "The Age of Surveillance Capitalism," a landmark work of social theory that reframed the fundamental nature of the digital economy (Zuboff, 2019). Zuboff's argument was that the leading technology platforms had developed a new form of economic logic, distinct from industrial capitalism, in which human experience itself, the granular data of daily life, thought, and behavior, was appropriated as raw material for the production of behavioral prediction products.

In Zuboff's framing, social media platforms are not primarily in the business of providing communication services. They are in the business of harvesting behavioral surplus, the data that your interactions generate beyond what is needed to provide the service, and using machine learning to extract from that surplus predictive intelligence about your future behavior. This predictive intelligence is then sold to advertisers, political campaigns, and other interested parties who wish to influence your behavior in particular directions.

The scale of data collection is staggering. Facebook's advertising system, for example, allows advertisers to target users based on not only their explicitly stated preferences but their inferred psychological characteristics, relationship status changes, recent life events, political leanings, religious affiliations, health conditions, and thousands of other attributes derived from behavioral data analysis. Cambridge Analytica's harvesting of data from up to 87 million Facebook profiles, used to build psychological targeting models for political campaigns in the United States and elsewhere, was presented as an aberration when it was exposed in 2018. In reality, it was an extreme example of a practice that is central to the platform's commercial model.

The data collection extends beyond what happens on the platform. Facebook's tracking pixels are embedded in an enormous proportion of the commercial web, allowing the company to build behavioral profiles of individuals across their entire internet activity, not just within Facebook itself. Instagram, WhatsApp, and Facebook are combined into a unified data infrastructure, allowing Meta to correlate behavior across what users often perceive as separate services. The Off-Facebook Activity tool, introduced in 2019, allows users to see some of this cross-site tracking. The list of companies that have shared data with Facebook about your activity on their sites is typically hundreds of entries long.

For children, the data harvesting dimensions are especially concerning. The Children's Online Privacy Protection Act (COPPA) in the United States prohibits the collection of personal data from children under 13 without parental consent, but platforms have historically done little to verify user age. Internal Facebook documents disclosed by Haugen included research into attracting younger users, including pre-teens, as a growth strategy, a finding that raised serious questions about whether the company's compliance with COPPA was substantive or merely nominal.

The surveillance capitalism model is reinforced by the structure of platform data practices in ways that make regulation difficult. The behavioral data that platforms collect is not merely stored passively. It is processed by machine learning systems that continuously refine behavioral models, which are then used to make real-time decisions about content ranking, advertising targeting, and product recommendations. This creates a feedback loop in which more engagement generates more data, which enables better prediction, which drives more engagement. Breaking this loop requires not just data protection regulation but fundamental changes to the algorithmic recommendation architecture. Platforms have strong financial incentives to resist those changes.

Regulatory Responses and Their Limits

The regulatory response to social media's documented harms has been significant in ambition but limited in effect. The European Union's Digital Services Act (DSA), which entered into force in November 2022 and became fully applicable to very large platforms in August 2023, represents the most comprehensive regulatory framework yet applied to social media in any major jurisdiction (European Union, 2022). The DSA requires large platforms to conduct and publish risk assessments for systemic risks including threats to public health, democratic discourse, and children's safety; mandates external auditing of these assessments; requires algorithmic transparency measures; and enables the European Commission to investigate and impose fines of up to six percent of global annual turnover for violations.

The DSA's transparency requirements are particularly significant in the context of this analysis. For the first time, platforms are legally required to give approved researchers access to data and algorithms sufficient to audit their claims about safety and governance. This is a meaningful departure from the voluntary, opaque approach that characterized the pre-DSA era, in which platforms' representations about their own practices were essentially unverifiable from the outside.

In the United States, federal regulation of social media has been largely blocked by Section 230 of the Communications Decency Act, which provides platforms with broad immunity from liability for user-generated content. The immunity was designed to encourage the development of internet services in the early commercial internet era, and it has been widely credited with enabling the growth of platforms that allow user expression. But critics argue that it has also insulated platforms from accountability for algorithmic design decisions. The amplification of harmful content is a platform choice, not user-generated content, yet courts have frequently treated recommendation algorithms as covered by Section 230.

Regulatory responses have also struggled with the transnational nature of platform operations. A platform regulated in California reaches users in every country on earth. Regulations passed in the European Union apply to platforms based in California. This jurisdictional mismatch creates opportunities for regulatory arbitrage and limits the effectiveness of any single jurisdiction's approach. The DSA's approach of focusing on systemic risk at the platform level, rather than on individual pieces of content, represents a more structurally sophisticated regulatory strategy, but its ultimate effectiveness will depend on enforcement resources and political will that are not yet established.

At the state level in the United States, several states have passed laws restricting minors' use of social media, requiring age verification, or mandating parental consent. These laws have faced First Amendment challenges, with courts reaching inconsistent conclusions about their constitutionality. The Supreme Court's 2024 term included major social media cases that will shape the legal framework for years to come, but the fundamental tension between platform editorial discretion and public interest regulation remains unresolved.

The limits of content moderation as a regulatory strategy are becoming increasingly apparent. Platform content moderation operations are expensive, inconsistent across languages and regions, prone to both over- and under-removal, and increasingly automated in ways that create their own problems of bias and error. More fundamentally, content moderation addresses symptoms while leaving the underlying engagement optimization architecture intact. Removing a piece of harmful content does not change the algorithm that would recommend the next piece of harmful content in its place.

What Comes Next

The trajectory of social media and AI is moving in several directions simultaneously, and not all of them are toward greater harm. The deployment of generative AI in social media content creation introduces new and potentially more serious versions of existing problems. Synthetic media, including AI-generated text, images, audio, and video, can now be produced at industrial scale, at low cost, and with sufficient fidelity to deceive casual viewers. The combination of generative AI content production with engagement-optimizing recommendation systems creates conditions for information environments that are not merely distorted by bias and misinformation but actively manufactured at scale by automated systems.

The specific risk is the commodification of synthetic influence. Political actors, commercial interests, and state-level adversaries can now generate coordinated inauthentic content at a scale and personalization level that was previously impossible. A synthetic influence operation can generate thousands of individually tailored pieces of content, each optimized for a specific audience segment's psychological profile, delivered through algorithmic recommendation systems that amplify based on engagement signals that the synthetic content is specifically engineered to trigger. The detection of such operations is a genuine technical challenge, and platforms' track records on identifying and removing coordinated inauthentic behavior suggest that they are not equipped to address this threat at the scale it will likely operate.

At the same time, the political and regulatory momentum for genuine structural reform is building in ways that were not evident five years ago. The Haugen disclosures, the DSA, the growing scientific literature on adolescent mental health, and a series of high-profile incidents connecting platform algorithms to real-world violence have created conditions for more serious regulatory action than anything that has been implemented to date. The question is whether regulatory frameworks can be designed with sufficient specificity and technical sophistication to address the actual mechanisms of harm, rather than merely their most visible symptoms.

Some researchers and advocates have proposed structural interventions that go beyond content moderation and transparency requirements. These include: algorithmic choice requirements, which would allow users to opt out of engagement-optimized recommendation in favor of chronological or user-curated feeds; interoperability mandates, which would reduce platform lock-in and allow users to migrate to alternatives; mandatory risk assessments for algorithm updates; and data minimization requirements that would limit the volume of behavioral surplus platforms can collect and retain.

None of these interventions is sufficient on its own, and each faces significant implementation challenges. The accumulation of evidence makes a compelling case that the current architecture of social media represents a genuine public health and democratic governance problem that warrants structural response, not merely surface-level content regulation. That evidence includes epidemiological data on adolescent mental health, the MIT misinformation research, the Haugen disclosures, the documented radicalization pipeline, and Zuboff's surveillance capitalism analysis.

The engagement engine will not reform itself. Its financial logic is too clear, its entrenchment too deep, and its political influence too significant. What comes next will depend on whether regulators, researchers, and the public can maintain the clarity of analysis necessary to address the machine for what it actually is.

References

  1. Vosoughi, S., Roy, D., & Aral, S. (2018). The spread of true and false news online. Science, 359(6380), 1146-1151.
  2. Haidt, J. (2024). The Anxious Generation. Penguin Press.
  3. Haidt, J., & Allen, N. (2020). Scrutinizing the effects of digital technology on mental health. Nature, 578(7794), 226-227.
  4. Pariser, E. (2011). The Filter Bubble. Penguin Press.
  5. Twenge, J.M. (2017). iGen. Atria Books.
  6. Zuboff, S. (2019). The Age of Surveillance Capitalism. PublicAffairs.
  7. Tufekci, Z. (2018, March 10). YouTube, the Great Radicalizer. The New York Times.
  8. Haugen, F. (2021, October 5). Congressional testimony before the U.S. Senate Commerce Subcommittee on Consumer Protection.
  9. European Union. (2022). Digital Services Act. Regulation (EU) 2022/2065 of the European Parliament and of the Council.
  10. Keyes, O., et al. (2019). A mulching proposal. CHI '19 Extended Abstracts. ACM.