The First Amendment does not protect messages posted on social media platforms.
The companies that own the platforms can, and do, remove, promote or limit the distribution of any posts according to corporate policies. But all that might soon change.
The Supreme Court has agreed to hear five cases during this current term, ending June 2024, that collectively give the court the opportunity to reexamine the nature of content moderation — the rules governing discussions on social media platforms such as Facebook and X, formerly known as Twitter — and the constitutional limitations on the government to affect speech on the platforms.
Content moderation, whether done manually by company employees or automatically by a platform’s software and algorithms, affects what viewers can see on a digital media page. Messages that are promoted garner greater viewership and interaction; those that are deprioritized or removed will obviously receive less attention. Content moderation policies reflect decisions by digital platforms about the relative value of posted messages.
As an attorney, professor and author of a book about the boundaries of the First Amendment, I believe that the constitutional challenges presented by these cases will give the court the occasion to advise government, corporations and users of interactive technologies of their rights and responsibilities as communications technologies continue to evolve.
In late October, the Supreme Court heard oral arguments on two related cases in which both sets of plaintiffs argued that elected officials who use their social media accounts either exclusively or partially to promote their politics and policies cannot constitutionally block constituents from posting comments on the officials’ pages.
In one of those cases, O’Connor-Radcliff v. Garnier, two school board members from the Poway Unified School District in California blocked a set of parents — who frequently posted repetitive and critical comments on the board members’ Facebook and Twitter accounts — from viewing the board members’ accounts.
In the other case, Lindke v. Freed, the city manager of Port Huron, Michigan, apparently angered by critical comments about a posted picture, blocked a constituent from viewing or posting on the manager’s Facebook page.
Courts have long held that public spaces, like parks and sidewalks, are public forums, which must remain open to free and robust conversation and debate, subject only to neutral rules unrelated to the content of the speech. The silenced constituents in the two cases insisted that in a world where a lot of public discussion is conducted on social media, digital spaces used by government representatives to communicate with constituents are also public forums and should be subject to the same First Amendment rules as their physical counterparts.
If the Supreme Court rules that public forums can be both physical and virtual, government officials will not be able to arbitrarily block users from viewing and responding to their content or remove constituent comments with which they disagree. But if the Supreme Court rejects the plaintiffs’ argument, the only recourse for frustrated constituents will be to create competing social media spaces where they can criticize and argue at will.
Moderation as editorial choice
Two other cases — NetChoice LLC v. Paxton and Moody v. NetChoice LLC — also relate to the question of how the government should regulate online discussions. Florida and Texas have both passed laws that modify the internal policies and algorithms of large social media platforms by regulating how the platforms can promote, demote or remove posts.
NetChoice, a tech industry trade group representing a wide range of social media platforms and online businesses, including Meta, Amazon, Airbnb and TikTok, contends that the platforms are not public forums, and that the Florida and Texas legislation unconstitutionally restricts the social media companies’ First Amendment right to make their own editorial choices about what appears on their sites.
In addition, NetChoice alleges that by limiting Facebook’s or X’s ability to rank, repress or even remove speech — whether manually or with algorithms — the Texas and Florida laws amount to government requirements that the platforms host speech they didn’t want to, which is also unconstitutional.
NetChoice is asking the Supreme Court to rule the laws unconstitutional so that the platforms remain free to make their own independent choices regarding when, how and whether posts will remain available for view and comment.
To reduce harmful speech that proliferates across the internet — speech that supports criminal and terrorist activity as well as misinformation and disinformation — the federal government has engaged in wide-ranging discussions with internet companies about their content moderation policies.
The Biden administration has regularly advised — some say strong-armed — social media platforms to deprioritize or remove posts the government flagged as misleading, false or harmful. Some posts related to misinformation about COVID-19 vaccines or promoted human trafficking. On several occasions, the officials would suggest that platform companies ban a user who posted the material from making further posts. Sometimes, the company corporate representatives themselves would ask the government what to do with a particular post.
While the public might be generally aware that content moderation policies exist, people are not always aware of how those policies affect the information to which they are exposed. Specifically, audiences have no way to measure how content moderation policies affect the marketplace of ideas or influence debate and discussion about public issues.
In Missouri v. Biden, the plaintiffs argue that government efforts to persuade social media platforms to publish or remove posts were so relentless and invasive that the moderation policies no longer reflected the companies’ own editorial choices. They argue that the policies were, in reality, government directives that effectively silenced — and unconstitutionally censored — speakers with whom the government disagreed.
The court’s decision in this case could have wide-ranging effects on the manner and methods of government efforts to influence the information that guides public debates and decisions.
Lynn Greenky is professor emeritus of Communication and Rhetorical Studies, Syracuse University.
The Sun-Times welcomes letters to the editor and op-eds. See our guidelines.
The views and opinions expressed by contributors are their own and do not necessarily reflect those of the Chicago Sun-Times or any of its affiliates.
This article was originally published on theconversation.com