Luján Joins Colleagues To Reintroduce Bipartisan Legislation To Strengthen Online Content Moderation & Hold Internet Companies Accountable
Washington, D.C. – U.S. Senators Ben Ray Luján (D-N.M.) joined Brian Schatz (D-Hawai‘i), and John Thune (R-S.D.) to reintroduce the Internet Platform Accountability and Consumer Transparency (Internet PACT) Act, bipartisan legislation which updates the Communications Act of 1934 by requiring social media companies to establish clear content moderation policies and holding them accountable for content that violates their own policies or is illegal. The bill is cosponsored by U.S. Senators Tammy Baldwin (D-Wis.), John Barrasso (R-Wyo.), Bill Cassidy (R-La.), John Hickenlooper (D-Colo.), and Shelley Moore Capito (R-W.Va.).
“Social media companies must develop clear and consistent content moderation policies that actually protect consumers online,” said Senator Luján. “Our world relies on the internet daily and more guardrails are urgently needed to protect users and hold social media companies accountable for dangerous content that exists on their platforms. I’m proud to join my colleagues to reintroduce this bipartisan legislation that will do exactly that.”
“Online companies need to establish clear content moderation policies, actually follow those policies, and respond to consumers when they raise concerns about their implementation,” said Senator Schatz. “By requiring these simple things, our bipartisan bill will better protect consumers and hold companies more accountable.”
“This bipartisan legislation is a common-sense approach to preserving user-generated content and free speech on the internet and holding Big Tech accountable by providing much-needed transparency to online consumers,” said Senator Thune. “In order to keep up with America’s ever-expanding digital landscape, and all of the consumers who depend on it, it’s important to pursue policies – like the Internet PACT Act – that protect online consumers by giving them more control of their online experience.”
There is widespread bipartisan agreement that social media platforms have inconsistent and opaque content moderation practices due to a lack of accountability. To address this, the Internet PACT Act creates more transparency by:
- Requiring online platforms to explain their content moderation practices in an acceptable use policy that is easily accessible to consumers;
- Implementing biannual reporting requirements for online platforms that includes disaggregated statistics on content that has been removed, demonetized, or deprioritized; and
- Promoting open collaboration and sharing of industry best practices and guidelines through a National Institute of Standards and Technology-led voluntary framework.
The Internet PACT Act holds platforms accountable by:
- Requiring large online platforms to provide due process protections to consumers through a defined complaint system that processes reports and notifies users of moderation decisions within twenty-one days, and allows consumers to appeal online platforms’ content moderation decisions;
- Amending Section 230 to require that large online platforms remove court-determined illegal content and activity within four days; and
- Allowing smaller online platforms to have more flexibility in responding to user complaints, removing illegal content, and acting on illegal activity, based on their size and capacity.
The Internet PACT Act protects consumers by:
- Exempting the enforcement of federal civil laws from Section 230 so that online platforms cannot use it as a defense when federal regulators, like the Department of Justice or Federal Trade Commission (FTC), pursue civil actions online;
- Allowing state attorneys general to enforce federal civil laws against online platforms; and
- Requiring the Government Accountability Office to study and report on the viability of an FTC-administered whistleblower program for employees or contractors of online platforms.