Could the tide be turning on our tacit acceptance of the role big tech plays in molding our minds?

Last month President Biden declared, in his State of the Union address, that “we must hold social media platforms accountable for the national experiment they’re conducting on our children for profit.”   He credited the courage of Facebook whistleblower Frances Haugen, who he hosted as his guest at the joint session of Congress and whose revelations last autumn made social media’s impact on the mental health of young people around the world undeniable.  It is a relief that Haugen’s inside story has finally struck a chord in the United States.  But the issues she highlights were not a surprise to many.

As campaigners like Privacy International have flagged, in the current data economy even our mental health is for sale.  Many mental health websites around the world share information about their visitors, including, in some cases, answers to self-assessment questionnaires about mental health.  That information, whether about children or the adults around them, is highly valuable in an online ecosystem where it is assumed to be legal to prey on and play with people’s emotional states.  And it is the kind of information that could be fed into the algorithms that decide the price, or availability, of your medical insurance.


Reports from Australia in 2017 claimed that Facebook had offered advertisers real time access to the emotional states of teenagers and young adults, allowing them to be targeted when they were at their lowest ebb.  It was a claim that Facebook denied.  But last year, Reset Australia found that it could buy advertising targeting thousands of children with dangerous interests like extreme weight-loss, alcohol and gambling, for a few dollars.  Surveillance advertising is the exploitation of all of our mental states for someone else’s benefit.

But that weighted blanket Facebook wants to sell you to calm your anxious dreams is just the tip of the iceberg.  The surveillance advertising business model is the oil that drives disinformation about Covid 19, turning it into “a partisan dividing line” instead of an infectious disease.  That business model is also the pusher of conspiracy theories that leads people to take up arms on the steps of the Capitol.  It is the fuel for Russian information warfare in the current crisis in Ukraine, and it has been targeting democracies around the world for years.  The algorithms that support surveillance advertising thrive on division, whatever the topic.

Campaigners and legislators have been grappling with these issues for more than a decade.  Earlier this year campaigners in Europe had a ground-breaking win with a ruling from the Belgian Data Protection Authority that consent pop ups that are used to legitimize massive online tracking by advertisers are in fact a breach of EU law.  Meta’s response to increased EU regulation had been to threaten that it may withdraw its business from Europe.  Perhaps that would be no bad thing.  If its business model cannot respect our rights, maybe it’s time for a new tech paradigm.


In the EU, the Digital Services Act and the AI Act attempt to limit the human rights impacts of technology.  And the UK’s Online Safety Bill, touted as a flagship piece of legislation to make the internet safer, was published last week. It will no doubt provoke more intense debates that pit safety against freedom of speech. But the bill’s focus on content is simultaneously too broad and too narrow and fails to touch the real problem.  It is the systems, not the content, that cause the real harm. And the issues caused by business models built on surveillance, profiling and targeting go far beyond what we say. They affect how we feel, how we behave, how we spend, and how we vote.

Even China has introduced regulation this month to tackle the influence of recommender algorithms that manipulate the way we see the world.  The US may be late to the party, but as the home of many of the tech giants that affect all our lives, its newfound interest in regulation may be a game changer.

But the reality is that any genuine move to address the harms must go beyond legislating to protect children online or to police content.  It is the business model that uses vast troves of data on each of us to understand what we think. It determines how to press our individual emotional buttons in order to change our opinions and our actions.  That is the biggest threat to our collective human future, and our children will only be safe when we are all free of it.

Susie Alegre is a senior fellow at the Centre for International Governance Innovation (CIGI) and an international human rights lawyer based in the UK, with specific expertise in in the intersection of human rights and technology. She is the author of the upcoming book Freedom to Think: The Long Struggle to Liberate our Minds.