老夫子传媒

漏 2024 | 老夫子传媒
Southern Oregon University
1250 Siskiyou Blvd.
Ashland, OR 97520
541.552.6301 | 800.782.6191
Listen | Discover | Engage a service of Southern Oregon University
Play Live Radio
Next Up:
0:00
0:00
0:00 0:00
Available On Air Stations

California lawmakers want more social media regulation. Here are the legal hurdles that could face

Facebook employees take a photo in front of the new Meta sign at the company headquarters in Menlo Park on Oct. 28, 2021.
Tony Avelar
/
AP Photo
Facebook employees take a photo in front of the new Meta sign at the company headquarters in Menlo Park on Oct. 28, 2021.

A California bill would hold social media companies legally responsible for addicting kids to their platforms. Tech lobbyists, digital rights advocates, and others say the proposal would run afoul of federal law and the U.S. Constitution.

When Sophie Szew first downloaded Instagram at her 10th birthday party, she was exposed to a flurry of information that 鈥減romoted eating disorders,鈥 she told California lawmakers. By 15, she said, she was following 鈥渆very starvation regimen recommended鈥 by Instagram鈥檚 鈥渆xplore鈥 page.

Szew, now 20, spoke in Sacramento at a Senate hearing in April in support of an making its way through the Legislature. It would hold companies legally responsible for using algorithms and design features that addict young people.

鈥淪tanding with me today is a generation that knows all too well what it is like to be harmed by flawed systems,鈥 she said.

The past few years have seen a steady drip of research and reporting about the effects of social media on teens. That has translated into a stream of legislative activity across the country, with several states passing or considering laws that regulate how social media companies do business.

In 2022, when posing as 13-year old girls in the U.S., U.K., Australia and Canada, they scrolled through videos served to them, stopping to watch and like videos about eating disorders, body image, and mental health. They found that TikTok recommended suicide content within 2.6 minutes and eating disorder content within 8 minutes.

In 2021, after a whistleblower leaked documents, the Wall Street Journal  that researchers at Meta, formerly known as Facebook, found that 32% of teen girls said that when they felt bad about their bodies, Instagram made them feel worse. Meta owns Instagram.

CalMatters reached out to Twitter, Reddit, Mastodon, TikTok and Meta, which also owns WhatsApp, for comment. Twitter鈥檚 press inbox responded automatically with a poop emoji. Reddit, Mastodon, and TikTok did not respond.

In a statement, Meta鈥檚 Global Head of Safety, Antigone Davis, wrote that Meta wants teens to be : 鈥淲e don鈥檛 allow content that promotes suicide, self-harm or eating disorders, and of the content we remove or take action on, we identify over 99% of it before it鈥檚 reported to us.鈥

Rachel Holland, a Meta spokesperson, offered information on a variety of features teens and parents can use to shape what young people see, like a setting that makes new teen users less likely to encounter sensitive content, and a feature that nudges teens when they鈥檝e been scrolling through the same kind of content for a while.

What the California bill does

The bill takes aim at how social media companies serve up content to young people under 18. Specifically, it prohibits social media companies from using a design, algorithm, feature, or practice, that companies know (or should know) causes young people to:

  • Develop an eating disorder, inflict harm on themselves or others, or become addicted to the social media platform;
  • Receive content that facilitates the purchase of controlled substances, like opioids; that facilitates suicide by offering information on how to die by suicide; or facilitates the sale of guns illegally.

It covers social media companies that earn more than $100 million in revenue per year and have users in California. Companies that violate the law could get sued by public attorneys, and would face penalties of as much as $250,000 per violation.

But the bill also builds in a way for companies to protect themselves from lawsuits: By auditing their designs, algorithms, features, and more at least quarterly for their potential to cause the harms listed in the bill and correcting issues within 30 days.

鈥淚f you look at a bill like this, the crux of the problem is content.鈥
SOPHIE COPE, SENIOR STAFF ATTORNEY, ELECTRONIC FRONTIER FOUNDATION

California isn鈥檛 the only state where legislators are aiming to regulate social media platforms. In March, Utah鈥檚 governor signed a into law, plus that institutes a social media curfew for Utah teens under 18 and requires teens to get their parents鈥 permission to set up social media accounts. Arkansas also passed requiring parental permission for teens to set up social media accounts. State legislators in New Jersey are considering that prohibits social media companies from using features that addict kids, and lawmakers in Minnesota are also mulling . Then, there are the bills that would regulate social media with a different aim: Texas that made it illegal to ban users based on their 鈥渧iewpoints,鈥 and Florida also restricting platforms鈥 ability to ban users.

On top of all the new proposals, there are a bevy of lawsuits ping-ponging their way through the courts that could reshape social media regulation. The Texas and Florida laws were challenged for violating the First Amendment; the Supreme Court will decide their fate. More than 80 cases have been grouped together in one , in which plaintiffs say that the 鈥 like an exploding toaster 鈥 because they addict children.

California bill could face legal hurdles

If the California bill became law, would it be able to dodge and weave its way through legal challenges?

Tech industry lobbyists, as well as some digital rights advocates and internet law experts, say that the bill would immediately bump up against the First Amendment of the U.S. Constitution and federal law. Both outrank or 鈥減re-empt鈥 in lawyer-speak, California laws.

The bill runs afoul of the Constitution because the Supreme Court has generally interpreted the First Amendment as , said Sophie Cope, a senior staff attorney at the nonprofit Electronic Frontier Foundation, which opposes the bill. Newspapers decide their layout, what goes on the front page and which images to use, Cope said. Social media platforms are doing similar things as online publishers when they make decisions about how content is displayed, shared, or promoted, she said.

This bill 鈥済oes directly at the heart of the editorial discretion of these platforms,鈥 Cope said.

鈥淚t鈥檚 the First Amendment that restricts the ability of legislators to control the flow of information at children鈥 to the degree legislators might want, agreed Eric Goldman, a professor at Santa Clara University School of Law and an expert on internet law. His example: a from 1997 about whether a law prohibiting the transmission of porn or 鈥渙bscene or indecent鈥 messages to kids violated the first amendment. The Supreme Court decided unanimously that it did.

Ed Howard, senior counsel for the Children鈥檚 Advocacy Institute at the University of San Diego School of Law and a supporter of the bill, disagrees, and argues that the bill doesn鈥檛 violate the First Amendment, in part because the algorithms that prioritize what content gets displayed don鈥檛 have their own First Amendment rights. And, he said, 鈥渢here is no blanket immunity to harm people through your speech.鈥

Then, there鈥檚 a federal law that could pose a problem for the California bill. Section 230 of the Communications Decency Act generally protects tech companies from legal responsibility for content posted by users on their platforms. If, for example, someone posts a defamatory review of a restaurant on Yelp, the restaurant owner could sue the poster, but Yelp itself would be protected, said Cope.

So, a key question is whether the California bill is taking aim at features and algorithms that a social media company has designed 鈥 for which it may not be protected by Section 230 鈥 or if it鈥檚 really requiring platforms to be accountable for content that users upload.

State Sen. , an Oakland Democrat who authored the bill, argues that the bill holds companies liable for their own products and practices, so it wouldn鈥檛 be preempted by Section 230. Part of the government鈥檚 job, she said, 鈥渋s to protect consumers from harmful business practices. That鈥檚 a very classic role of government.鈥

But, said Cope with the Electronic Frontier Foundation, 鈥渋f you look at a bill like this, the crux of the problem is content.鈥 The bill is structured to make platforms 鈥渞esponsible for bad content reaching young people,鈥 so the federal law should pre-empt it, she said. The features, in other words, aren鈥檛 a problem when they鈥檙e showing #vanlife and knitting videos to young users, but when they鈥檙e serving up starvation diet tips, they are.

Public may be left in suspense

Last week the bill was sent to the Senate Appropriations Committee and added to a special pile of bills that the committee will analyze for costs and benefits. That pile of bills 鈥 known as the 鈥渟uspense file鈥 鈥 is officially about budget analysis. Unofficially, it鈥檚 also a without taking any heat. That鈥檚 because, unlike other votes that state legislators take, votes on suspense file bills are kept secret. If a bill never makes it out of the special pile, the public doesn鈥檛 know whose votes led to its demise.

A version of the bill was introduced last year. No legislator voted against it publicly, but the bill was put on suspense file and died there.

This year鈥檚 bill 鈥渃ould suffer that fate,鈥 said Skinner, 鈥渂ut I鈥檓 hopeful that the amount of increased awareness, increased evidence and action by other states will hopefully put it in much better shape,鈥 she said.

 is a nonprofit, nonpartisan media venture explaining California policies and politics.