Social media companies can’t be forced to block teens from seeing ‘harmful’ content, judge rules

Share

A federal judge has ruled that social media companies can’t be required to block certain types of content from teens. The ruling will prevent some aspects of a controversial social media law in Texas from going into effect.

The ruling came as the result of tech industry groups’ challenge to the Securing Children Online Through Parental Empowerment (SCOPE) Act, a Texas law that imposes age verification requirements and other policies for how social media companies treat teenage users. But, as The Verge points out, the measure also requires companies to “prevent the known minor’s exposure to harmful material,” including content that “glorifies” self-harm and substance abuse.

It’s that latter requirement that was struck down, with the judge saying that “a state cannot pick and choose which categories of protected speech it wishes to block teenagers from discussing online.” The judge also criticized the language used in the law, writing in his decision that terms like “glorifying” and “promoting” are “politically charged” and “undefined.”

At the same time, the judge left other aspects of the law, including age verification requirements and bans on targeted advertising to minors, in place. NetChoice, the tech industry group that challenged the law, has argued that measures like the Scope Act require major tech companies to increase the amount of data collected from minors.

The Texas law, originally passed last year, is one of many across the country attempting to change how social media platforms deal with underage users. New York recently passed two laws restricting social media companies’ ability to collect data on teenage users, and requiring parental consent for younger users to access “addictive” features like algorithmic feeds. California lawmakers also recently passed a measure, which has yet to be signed into law by the governor, that requires social media companies to limit notifications to minors and restrict them from “addictive” algorithms.