Mark Zuckerberg sat in a Los Angeles courtroom. Under oath. Under lights. Under the gaze of parents who say his platforms destroyed their children. They didn't ask if they wanted their kids addicted to Instagram and Facebook. No one was consulted if the algorithm should be designed to maximize engagement at the expense of mental health. Nobody asked if children under 13 should be on platforms that explicitly prohibit them. They just built the machine. Fed it children. And counted the engagement metrics. The Landmark Case The trial in Los Angeles is unprecedented. It combines claims from multiple lawsuits into a single proceeding that alleges: Meta deliberately designed Instagram and Facebook to be addictive to minors The company knew its platforms harmed teen mental health Internal research showed the damage, and Meta suppressed it Children under 13 were allowed on platforms despite terms of service prohibiting it Algorithmic manipulation was used to maximize engagement without regard for wellbeing This isn't a class action about data privacy. This is a product liability case about a defective product that was marketed to children with known dangers. "If a toy company knew their product was choking children and kept selling it, we'd call that criminal. When a tech company knows their platform is harming children and keeps growing it, we call it innovation." — Plaintiff's attorney, allegedly. The public was never asked if the comparison was fair. The comparison is fair. What Meta Knew The most damning evidence comes from Meta's own internal research. Documents revealed in litigation show: The Instagram Teen Mental Health Study Meta's own researchers found that Instagram made body image issues worse for one in three teenage girls. The internal presentation stated: "We make body image issues worse for one in three teen girls." The response from Meta leadership? The research was buried. The product continued. The marketing intensified. The Addiction by Design Documents Internal communications show Meta engineers discussing: Variable reward mechanisms — The same psychology that makes slot machines addictive Infinite scroll — Eliminating natural stopping points Push notifications — Triggering compulsive checking behavior Social validation feedback loops — Likes, comments, and shares as dopamine triggers Fear of missing out (FOMO) — Algorithmic amplification of social anxiety These weren't accidental features. They were deliberately designed to maximize time spent on the platform. The Children Problem Meta knew millions of children under 13 were using Instagram despite the terms of service prohibiting it. Internal documents show: Age verification was deliberately weak Marketing strategies targeted younger demographics Features were developed with teenage users in mind The "under 13" prohibition was a legal shield, not an enforcement reality No community input if children should be protected. They calculated that protecting children would reduce growth. The Roblox Dimension The trial has also exposed disturbing allegations about Roblox, the gaming platform popular with children: Predator access — The platform allegedly provided tools that allowed adults to identify and contact children Inadequate moderation — Content moderation was insufficient to prevent grooming Virtual currency manipulation — Children were spending real money without understanding the value Addictive mechanics — Game design employed the same engagement-maximizing techniques as social media Roblox has over 70 million daily active users, the majority of whom are children. The platform presents itself as a safe space for kids. The allegations suggest it was anything but. People were kept in the dark parents if a platform with 70 million daily users — mostly children — should have robust safety protections. It didn't have them. The Algorithmic Manipulation Understanding how social media addicting children requires understanding the algorithm: The Engagement Optimization Loop Social media algorithms are optimized for one metric: engagement. The more time you spend, the more content you consume, the more ads you see, the more money the platform makes. For children, this optimization creates a feedback loop: Child opens app Algorithm serves content designed to trigger emotional response Emotional response drives continued engagement Engagement signals algorithm to serve more triggering content Repeat until child is unable to stop The Emotional Manipulation Research shows that content triggering strong emotions — outrage, anxiety, envy, excitement — generates more engagement than neutral content. The algorithm learns this and serves increasingly emotionally provocative content. For developing minds, this creates: Emotional dysregulation — Difficulty managing intense feelings Anxiety disorders — Constant low-level stress from provocative content Depression — Social comparison and FOMO Attention deficits — Inability to focus on non-stimulating activities The Dopamine Hijacking Every like, comment, and share triggers a dopamine release — the same neurotransmitter involved in drug addiction. For adolescent brains, which are still developing impulse control mechanisms, this creates: Compulsive checking — Can't resist looking at notifications Withdrawal anxiety — Distress when unable to access the platform Tolerance — Need for increasing amounts of stimulation Loss of interest — In activities that don't provide instant gratification No one was given a voice if hijacking children's dopamine systems was ethical. They measured the engagement. The Kids Off Social Media Act In response to growing public outrage, Congress is advancing the Kids Off Social Media Act, which would: Ban social media for children under 13 — With actual enforcement mechanisms Require age verification — Real verification, not self-reported birthdates Prohibit algorithmic targeting of minors — No engagement-optimized feeds for kids Mandate safety features — Default privacy settings, content filtering, time limits Create accountability — Penalties for platforms that fail to comply The bill has bipartisan support, which in 2026 is practically a miracle. It also has fierce opposition from the tech industry, which argues it would: Stifle innovation Be technically impossible to enforce Violate First Amendment rights Drive children to less regulated platforms these arguments are the same ones the tobacco industry made about age restrictions on cigarettes. "We can put age verification on alcohol, tobacco, firearms, and gambling. But somehow age verification on social media is 'technically impossible.' Funny how that works." — Senator, allegedly. The Parental Consent Fiction Social media platforms claim they require parental consent for users under 13. In practice: Consent is a checkbox — A child can check a box saying they're 13 or have parental permission No verification occurs — Platforms don't verify age or parental consent Design targets youth — Features, marketing, and content appeal to younger users Enforcement is performative — Accounts identified as underage are rarely removed The Children's Online Privacy Protection Act (COPPA) requires parental consent for data collection from children under 13. Platforms comply technically while violating the spirit: They prohibit under-13 users in their terms They make no real effort to enforce the prohibition They design products that appeal to under-13 users They collect data from users they know are underage The public was sidelined if the legal fiction of "parental consent" was adequate. It wasn't. The Mental Health Crisis The consequences of social media addiction in children are measurable and devastating: Teen depression has increased 60% since 2012, correlating with smartphone adoption Self-harm hospitalizations for girls aged 10-14 have tripled Suicide rates for teen girls have increased 70% in the past decade Anxiety disorders affect 1 in 3 adolescents, up from 1 in 10 twenty years ago Correlation isn't causation, but Meta's own research established the causal link. They knew. They kept going. Communities were ignored if the mental health of a generation was worth the advertising revenue. They calculated that it was. Raising the Standard For Parents Delay social media access — The longer you wait, the better Monitor usage — Know what platforms your children use and how much time they spend Have conversations — Talk to your children about how algorithms work and why they're designed to be addictive Model behavior — Children learn from watching your phone habits Demand school policies — Push for phone-free school environments For Everyone Support the Kids Off Social Media Act — Contact your representatives Hold platforms accountable — Support litigation and regulatory action Fund research — Support independent research on social media's effects on children Change the narrative — Challenge the idea that children "need" social media For Educators Teach media literacy — Help students understand algorithmic manipulation Create phone-free spaces — Implement and enforce phone-free classrooms Support affected students — Recognize signs of social media addiction and mental health issues What This Costs You Mark Zuckerberg sat in that courtroom because parents said his platforms addicted their children. The internal documents prove Meta knew. The algorithm was designed to hook kids. Children under 13 were on the platform despite the rules. And the mental health of a generation was traded for engagement metrics. It was forced on everyone if children should be addicted to their platforms. They didn't ask if the mental health damage was acceptable. They didn't ask if parents consented to their children being algorithmically manipulated. They just built the machine. Fed it children. And counted the money. Now the parents are counting the cost. --- Related: AI Chatbots Drove Children to Suicide The Engagement Bait Economy Snapchat My AI Spy