The Meta Oversight Board’s Human Rights Future

Responses to the new Oversight Board created by Facebook (now Meta) have run the gamut from enthusiastic to overtly suspicious. Many observers are highly skeptical of the Board’s ability to hold Meta accountable or to protect the company’s users. Much of this skepticism is based on a misplaced analogy between the Board and domestic courts. We argue, in contrast, that the core challenges and opportunities that the Oversight Board faces are similar to those confronted by international human rights tribunals. Analyzing the Oversight Board as a de facto human rights tribunal sheds light on the strengths and weaknesses of its structure, decision-making, and potential future trajectory. Seen from this perspective, key design features that commentators have criticized are, in fact, strengths the Board is already using to expand its authority, develop human rights norms, and influence efforts to regulate social media platforms.

This Article is the first to examine the human rights origins of the Oversight Board, the similarities between the new body and international human rights tribunals, and how the Board is using human rights standards to pressure Meta to improve its content moderation policies and to inform ongoing efforts to regulate social media companies. We complement this analysis with a range of recommendations for the Board to become even more effective, as well as explore potential risks and challenges, including backlash, whitewashing, and negative spillover. Although a sanguine vision of the Board’s future is by no means assured, when viewed in light of the experience of international human rights tribunals, we conclude that the Board has the potentical to serve as a meaningful check on Meta and to significantly advance the promotion and protection of human rights online.

Introduction

In late 2020, Facebook (now Meta)1 created an Oversight Board (“OB” or “Board”) to review challenges to the social media platform’s decisions to remove user posts and to advise the company on improving its content moderation policies. At the time, the Board drew widespread notice and scrutiny from civil society groups and commentators, whose responses have run the gamut from enthusiastic to overtly suspicious. The Oversight Board has recently received fresh attention as governments adopt new initiatives to regulate internet platforms.2 Several of these initiatives impose liability on social media companies for retaining or removing user content and establish procedures for users to contest whether such content is illegal or violates the companies’ terms of service.3

Does the Meta Oversight Board represent the first wave of social media companies becoming more accountable to their users and other stakeholders? Does the Board provide a model for government regulators and other internet platforms to move toward greater transparency in content moderation? Or is it merely a fig leaf that seeks to ameliorate Meta’s many public relations challenges while obscuring its human rights harms?

Attempts to answer these questions are hampered by a lack of agreement about how to conceive of this new body. Not quite a corporate entity and not really a court, the Oversight Board is a body of nearly two dozen experts drawn from geographically, culturally, and professionally diverse backgrounds, funded by an independent trust created under Delaware law, empowered by Meta to make decisions on content that affect billions of individuals worldwide. First proposed as a “quasi-judicial unit with autonomous powers” inspired by constitutional law,4 the idea of an independent review mechanism captured the imagination of CEO Mark Zuckerberg, who floated the idea of creating a “Supreme Court of Facebook”5 to help the company make decisions about speech on its platforms.6 Since its launch, commentators have compared the Oversight Board to U.S. courts,7 administrative tribunals,8 constitutional courts,9 private arbitral bodies,10 grievance mechanisms,11 and self-regulatory initiatives.12

Each of these analogies, however, fails to account for the Board’s distinctive design features, the circumstances of its creation, and its potential for developing international norms. Analogies to domestic courts, for example, inaccurately assume that the OB operates within a preexisting and stable legal system with established checks and balances and enforcement powers. Analogies to contractual or self-regulatory mechanisms overlook the public law aspects of the Board’s work or the range of stakeholders that are affected by its decisions.

Commentators also disagree over what this new body should do. Some argue that the Board’s primary goal should be to increase Meta’s accountability to users13 or to provide a mechanism for individual redress.14 Others have emphasized the OB’s role in curbing the company’s adverse human rights impacts or as a forum for negotiating the rules governing online speech.15 The Charter establishing the Oversight Board describes its purpose as protecting free expression “by making principled, independent decisions about important pieces of content and by issuing policy advisory opinions on Facebook’s content policies.”16 Yet the OB also calls itself an “independent grievance mechanism”17 that “promote[s] the rights and interests of users.”18

The Board’s design raises significant doubts about its ability to accomplish any of these objectives. The OB’s authority is limited: It can review Meta’s decisions to remove or allow individual pieces of content on Facebook and Instagram that are contested by users or submitted for review by the company itself. Additionally, it can evaluate Meta’s content moderation policies, but its recommendations are only advisory. The OB is also hampered by jurisdictional and functional constraints. Meta’s news feed ranking and political advertising are outside the Board’s remit, as are the company’s decisions to comply with local law.19 In addition, the Board has been issuing between twelve and sixteen decisions each year. Given the volume of content shared on the platform, such a limited number of individual cases seems unlikely to have much impact.20 Lastly, the Board is perpetual in theory but time limited in practice, since its existence likely depends on continued funding from Meta.

These limitations are daunting. They are not, however, all that new. The central challenges that the Oversight Board faces—as a body of limited, consent-based jurisdiction and constrained authority, facing overwhelming demand, whose existence, legitimacy, and efficacy depend in large part on the continued support of its creator—are similar to the challenges confronted by international human rights courts and quasi-judicial review bodies such as U.N. human rights treaty bodies, the European Court of Human Rights, and the Inter-American Commission on Human Rights (“international human rights tribunals” or “tribunals”).21 We argue that, like international human rights tribunals, the Board can contribute most effectively to the international human rights system not by correcting errors or providing remedies in individual cases but by pressuring Meta to improve its processes and by developing the law.

This Article sketches a comprehensive vision for the Oversight Board as a de facto international human rights tribunal. It does so in three steps. We first explore the influence of human rights law and tribunals on the Board’s origins, its design features, and operations. Previous accounts of the OB have overlooked how these sources affected the process leading to the Board’s creation, the choices regarding the OB’s review powers, and the legal norms it applies.

Second, we consider the structural and functional analogies between the OB and international human rights tribunals. We draw on the extensive experience of these tribunals over the last several decades to highlight the many similarities between the Oversight Board and the tribunals, while also noting salient differences.22 Comparing the OB to these bodies sheds light on the strengths and weaknesses of the Board’s structure, its decision-making processes, and its potential future trajectory.23 The similarities between the two sets of institutions reveal that key OB design features that some observers have criticized—such as its ability to issue only narrow binding rulings while designating its policy guidance as nonbinding—are actually strengths that the Board can use to incrementally expand its authority over time. We also review OB decisions that suggest that the Board is beginning to emulate the course previously charted by international human rights tribunals, for example by issuing detailed policy recommendations and pressing Meta to provide information about its algorithms and communications with governments.

Third, we sketch a vision for the Oversight Board’s human rights future. We argue that the OB should develop methodologies similar to those that the tribunals have used to promote accountability and to contribute to human rights norm development. We contend that the Board is uniquely placed to explain, through an accretion of case-specific decisions on content moderation and broader policy recommendations, how international rules that bind states apply to private social media companies. We show that the Board is already making significant contributions in this regard—some of which are being cited favorably by other human rights bodies—and we highlight opportunities for the OB to influence the regulation of internet platforms by national governments and the European Union.

This sanguine vision of the Board’s human rights future is by no means assured, however. Efforts by the OB to enforce its existing mandate or expand its authority may engender resistance by Meta. The company may also engage in whitewashing, touting its commitment to the Board and to human rights generally in an attempt to avoid more stringent regulation. In addition, skeptics have also questioned the legitimacy of a private monitoring body developing public norms for content moderation that are precise enough to impose meaningful constraints on social media companies while also taking account of the different ways that the value and harm of speech are balanced around the world. There is also a risk of negative spillover if states and companies misuse the Board’s normative output in ways that weaken existing international human rights standards. We take these concerns seriously and suggest a number of ways to mitigate them.

This Article contributes to robust ongoing debates about social media content regulation in several ways. First, we analogize the Oversight Board to international human rights tribunals to shed light on the opportunities, risks, and challenges that the OB is facing and is likely to face. Second, we provide the first comprehensive analysis of the Board’s contributions to human rights doctrine, including its clarification of how freedom of expression and other human rights apply to private internet platforms in a wide array of contexts. Third, we offer recommendations for the Board to shape its future trajectory as a de facto international human rights tribunal, guidance that we hope will also influence other initiatives to address social media harms.

The Article proceeds in three Parts. Part I describes the role of international human rights law and institutions in the Oversight Board’s origins and design and identifies key similarities between the Board and international human rights tribunals. Drawing on the experiences of human rights tribunals, Part II describes two ways in which the Board will have an impact on both Meta and human rights law and explores how these approaches have already emerged in the Board’s decisions. Part III discusses the risks and challenges the OB may face in pursuing this trajectory. A brief conclusion summarizes the Article’s contributions.

I. The Oversight Board as a Human Rights Tribunal

This Part reviews the origins and design features of the Oversight Board and compares and contrasts these features with those of international human rights tribunals.24 This comparison reveals that what many commentators view as weaknesses—in particular, the OB’s circumscribed jurisdiction and authority—may actually help the Board promote human rights.

A. The Influence of Human Rights on the OB’s Origins and Design Features

In this Section, we provide basic information about the Board’s origins and functions, explore how international human rights law and tribunals were discussed during the lead up to the Board’s creation, and analyze the references to human rights in the Board’s founding documents.

1. Origins

The Oversight Board’s trajectory from initial conception to functioning institution was a remarkably short two and a half years. The first public statements about the Board occurred in early 2018, including Mark Zuckerberg’s reference to a “Supreme Court” that would “ultimately make the final judgment call on what should be acceptable speech.”25 The company announced plans for what eventually would be called the Oversight Board in late 2018; the Board’s Charter was revised and finalized during 2019; Meta released the OB’s Bylaws in January 2020; and the inaugural members were appointed by May 2020.26 The Board began operating in the fall of 2020.27

The Board itself does not have a formal legal personality. The Trust Agreement established by Meta creates an irrevocable noncharitable trust under Delaware law.28 The trust is empowered to create a limited liability company that, in turn, has the authority “to establish, administer, and attend to the ongoing operation of the group of individuals who make up the Board Members.”29

The current composition of the Board is diverse along multiple dimensions. Twenty-two members hail from nineteen countries and bring expertise and professional backgrounds in law, human rights, technology regulation, international politics, content moderation, journalism, health communication, intellectual property, management, policy design, and digital rights, among others.30 The Charter provides that “a committee of the board will select candidates to serve as board members based on a review of the candidates’ qualifications and a screen for disqualifications,” and that both Meta and the public may propose candidates.31 The staff that supports the Board’s work also brings a broad range of expertise. The Director of Oversight Board Administration, for example, is the former Executive Director of ARTICLE 19—one of the preeminent civil society organizations on the freedom of expression.32

The Board’s structure was developed in conjunction with a consultative process that took place between the release of the initial draft Charter in January 2019 and the publication of the final Charter in September of that year. The January draft posed questions as “a starting point for discussion on how the board should be designed and formed.”33 From late 2018 through mid-2019, Meta held a series of in-person consultative meetings around the world. The company then published a document summarizing the input from workshops, roundtables, and an online survey.34

During this time, civil society groups and scholars increasingly called for social media platforms to look to international human rights law to guide their content moderation policies.35 These demands featured prominently during the OB consultative process. Advocates argued that international human rights law would “provide the oversight board with a set of tools and a ‘common vocabulary’ for addressing and resolving hard questions around the moderation of online content.”36 They also noted that Meta had already committed to evaluating the human rights impact of some of its policies as a member of the Global Network Initiative.37 Other contributors, however, worried human rights “would not be enough to adjudicate hard cases” and questioned whether they would provide enough clarity to make difficult tradeoffs.38

Two reports commissioned by Meta shed light on the influence of international human rights law and institutions during the consultative process. The first was produced by BSR, a leading business and human rights advocacy organization. Meta asked BSR to review the consistency of “the final charter, bylaws, and operations” of the Board with “human rights-based approaches.”39 The report strongly recommended that Meta “incorporate a firm commitment to international human rights standards into the Oversight Board’s governance charter and bylaws” and “explain the link between human rights and Facebook’s values.”40

The second report, co-authored by a constitutional law scholar and Meta’s Head of Product Policy, discussed potential models for the Board’s design.41 The templates included the Court of Justice of the European Union, the Inter-American Commission and Court of Human Rights, and the European Court of Human Rights (ECtHR).42 Especially noteworthy was a reference to the ECtHR’s use of “precedent as a legitimizing tool to substitute for its lack of coercive power in inducing domestic courts to apply its rulings.”43 In addition to these two reports, advocacy groups submitted briefing papers that discussed the relevance of international human rights tribunals.44

The final version of the OB that emerged from the consultative process explicitly embraces human rights norms and objectives in several ways. Before describing those references, we first outline the OB’s jurisdiction and standing rules to provide context for our analysis.

2. Jurisdiction and Standing

Under the Charter, a case is initiated when an individual Meta user (either the person who posted the content or another user who flagged it for review) or the company asks the Board to review a decision to retain or remove content.45 Once in receipt of such a request, the Board has jurisdiction to receive submissions from the user, Meta, and third parties; apply Meta’s Community Standards and Values;46 seek additional information from the company; and provide a written explanation of its decision. In terms of remedies, the OB can direct Meta to retain or remove the content,47 or (following an October 2022 amendment to the Bylaws) to retain it with a warning screen.48 The Board can also provide non-binding policy guidance “specific to a case decision or upon [the company’s] request, on [Meta’s] content policies.”49

The Bylaws narrow the Board’s jurisdiction described in the Charter. Article 2 of the Bylaws excludes five broad categories of content from OB review (“content posted through marketplace, fundraisers, Facebook dating, messages, and spam”) as well as Meta’s decisions involving intellectual property violations and services on several of Meta’s platforms (“WhatsApp, Messenger, Instagram Direct, and Oculus”).50 The Bylaws also exclude review of content removed by Meta “following the receipt of a valid report of illegality” or content that is unlawful or criminally unlawful in a connected jurisdiction—issues that we discuss in greater detail below.51

The Bylaws also create standing requirements. To request review by the Board, a user “must have an active Facebook or Instagram account.”52 Initially, the OB could only review “content that has been removed for violations of content policies.”53 In April 2021, Meta authorized the Board to review user challenges to content that the company decided to retain on its platform.54

3. Applicable Law and References to Human Rights

The Oversight Board’s founding documents contain multiple references to human rights. Both the Charter and the Delaware Trust that established the limited liability company that created the OB refer to the Board’s purpose as “protect[ing] free expression by making principled, independent decisions about important pieces of content and by issuing policy advisory opinions on Facebook’s content policies.”55

Human rights norms are also central to the law that the Board applies. Although the initial draft of the Charter referred only to Meta’s Community Standards and Values,56 the final version adds “human rights norms protecting free expression” as a third source of law.57 In addition, the Values explain that Meta will “look to international human rights standards” in deciding whether content that otherwise contradicts the company’s Community Standards should nonetheless remain on the platform.58

As we describe below, the Board has fully embraced international human rights law in its decisions, recommendations, and other documents. In its first Annual Report, published in June 2021, the Board emphasized its “conviction that Meta will make content moderation decisions in a fairer, more principled way if it bases those decisions on the international human rights standards to which it has committed itself.”59

B. Shortcomings in Design

Since the Oversight Board’s inception, commentators have identified a range of shortcomings in the design of this new private monitoring body. This Section catalogues these critiques to set the stage for our more favorable assessment that analogizes the OB to an international human rights tribunal.

Perhaps the most trenchant criticism relates to the Board’s highly circumscribed jurisdiction and access rules in individual content moderation cases.60 The Board can only review appeals submitted by individual users and offer recommendations specific to those cases or in response to Meta’s requests for policy guidance.61 There may thus be decisions or policies that the Board cannot consider because no single user has a sufficient interest to challenge them, even though they have significant consequences when viewed cumulatively. In addition, non-Facebook and non-Instagram users cannot submit cases to the Board62 despite the fact that content on the platforms may significantly affect them.63 These limitations may inhibit the Board from considering the systemic or cumulative effects of Meta’s content moderation decisions, which can “result in severe human rights impacts.”64 Individualized ex post decision-making is a poor vehicle for addressing systemic harms.65

Commentators have been especially critical of the OB’s inability to review cases in which Meta removes content to comply with local law. Construed broadly, this provision would preclude the Board from considering situations in which governments censor online content or engage in disinformation campaigns—both of which have significant adverse human rights consequences.66 The Board also lacks authority over important areas of the social media company’s business, including Meta’s content ranking algorithms.67

Another set of critiques relates to the Board’s constrained authority in compelling Meta to disclose relevant information. The Charter commits Meta to “provid[ing] information, in compliance with applicable legal and privacy restrictions, that is reasonably required for the board to make a decision.”68 However, the company appears to reserve to itself the determination of what information is “reasonably required,” and it has rejected requests that would have helped the Board make informed decisions.69

The Charter and Bylaws also adopt a narrow conception of compliance.70 The Bylaws explain that while “[t]he board’s resolution on each case will be binding on Facebook, . . . the policy advisory statement from the board will be considered as a recommendation.”71 Thus, only the outcome of a specific case is binding. The company has agreed to consider the feasibility of applying the Board’s decisions to “identical content with parallel context.”72 However, Meta need not give individual decisions wider application if it concludes that this is not feasible.73 As for OB advisory statements, the Charter commits the company only to take these policy recommendations “into consideration,”74 including by “analyzing the operational procedures required to implement the guidance, considering it in the formal policy development process of Facebook, and transparently communicating about actions taken as a result.”75

Concerns have also been raised about Meta’s ability to circumvent or limit the Board’s powers.76 Although noncompliance could result in reputational consequences, Meta may have an incentive to disregard at least some Board decisions.77 If the company changes its Community Standards or Values, the OB would be obligated to apply these revised rules—at least prospectively.78 Meta also retains significant power to amend the Bylaws.79

Finally, the Board’s existence depends on continued funding from Meta. Meta has funded the OB robustly in the form of an initial contribution to the irrevocable Trust of $130 million in 2019 when the Board was created, and a second contribution of $150 million in 2022.80 After that funding is exhausted, however, Meta will decide whether to provide additional support, although presumably the Board could also seek funding from other sources.

C. A Better Analogy: International Human Rights Tribunals

Many critiques of the Oversight Board and Meta’s ability to circumvent its decisions compare the OB to a domestic court. Seen from this perspective, the Board is indeed inadequate. We contend, however, that a better analogy for the OB is found in international human rights tribunals. The following discussion identifies OB design elements that are similar to central features of those tribunals, and then reviews several salient differences.

1. Similarities

The Oversight Board is similar to international human rights tribunals in several important respects. First, both institutions have limited jurisdiction conferred upon them by their creators, as opposed to national courts that derive their authority from and are part of a broader legal system. The Board’s Charter, much like a treaty establishing an international tribunal, defines the cases the OB can hear, who can bring them, and the legal issues it can address.81 In this respect, the Board mirrors the structure of international monitoring bodies, whose jurisdiction, access rules, and review powers are circumscribed by the consent of the states that created them, and which are vulnerable to their creators imposing new limits on that consent or withdrawing it altogether.82

Critics who lament the Board’s limited ability to act as a meaningful check on Meta underestimate the constraints of this consent-based design. Limited jurisdiction and authority are often necessary compromises because states (like Meta) are unwilling to delegate more expansive review powers to these bodies—at least at the outset. For example, the U.N. Human Rights Committee, created during an era when U.N. members were wary of international monitoring, was given only modest authority to review compliance with the International Covenant on Civil and Political Rights (ICCPR).83 And in regional human rights systems, states could join a treaty without accepting the jurisdiction of the associated court.84

When international review mechanisms are first established, therefore, the choice is often not between more or less authority, but between less authority and no authority at all. This perspective finds support from the fact that no other social media company has yet established a body similar to the OB. A few platforms have created bodies of external experts to advise them on content moderation policies.85 However, these bodies have no binding authority vis-à-vis their respective companies, and they can be disbanded at any time—as recently occurred with Twitter’s Trust and Safety Council.86

Second, international human rights tribunals, like the Board (but unlike most domestic courts), have both binding and nonbinding functions. Regional human rights courts, for example, possess both contentious and advisory jurisdiction.87 Only the former results in legally binding judgments, and the latter is often more frequently used during a tribunal’s early years.88 All of the normative output of the ten U.N. treaty bodies—general comments, country-specific concluding observations, and individual decisions—are authoritative but nonbinding. Much like the OB’s policy guidance to Meta, states must give only “due consideration” to treaty body recommendations.89

Third, the Oversight Board exercises fact-finding and promotional functions comparable to many tribunals. To inform its review of Meta’s decision to remove or retain content, the OB meets with experts and advocacy organizations,90 similar to U.N. treaty bodies that confer with civil society organizations when reviewing the human rights record of state parties.91 The Board also publishes annual reports that analyze how its “decisions have considered or tracked the international human rights implicated by a case,” as well as “Facebook’s implementation and response to board decisions and policy advisory statements.”92 This is similar to the treaty bodies, which report each year to the U.N. General Assembly on their human rights activities.93 The Board has also solicited information outside of the content review process, similar to human rights commissions that proactively investigate violations in particular situations or countries.94

Fourth, both international human rights tribunals and the Board face overwhelming demand. The backlog of tens of thousands of cases pending before the ECtHR is the most extreme example, but the problem is systemic. Other regional courts and commissions, as well as some U.N. treaty bodies, receive far more petitions than they can review within a reasonable timeframe. Simply put, the tribunals cannot be expected to achieve individualized justice given the pervasiveness and scale of human rights harms.95 The Board, too, is unable to provide individual justice in light of the volume of Meta’s content moderation actions worldwide.96

Fifth, international human rights tribunals, like the OB, have only limited ability to enforce their decisions. This is true not only for nonbinding policy guidance and recommendations, but also for legally binding judgments. Even for the ECtHR—widely recognized as the most active and influential human rights court—partial and delayed compliance is the norm, and outright resistance is not uncommon.97 The compliance records of other tribunals are generally weaker. This has not, however, prevented these bodies from requesting information about implementation and following up with states in subsequent interactions98—a practice that the Oversight Board has also adopted.99 These interactions sometimes induce countries to take incremental steps toward compliance, a trend that can be accelerated when civil society groups rely on tribunal decisions as focal points for domestic and transnational advocacy.100

Sixth, and finally, states can deploy a variety of tools to weaken or undermine international human rights tribunals, just as Meta can vis-à-vis the OB. Scholars have catalogued a range of “pushback” and “backlash” techniques, including critiquing specific judgments or doctrines, narrowing a tribunal’s delegated authority, changing the composition of its members, starving it of resources, or withdrawing from its jurisdiction.101 Although resistance to national courts is also common in some countries, opposition to international tribunals can, in the most extreme cases, lead to their de facto suspension.102

Notwithstanding these constraints and limitations, most human rights tribunals have incrementally expanded their competencies over time and developed international norms without eliciting extensive state opposition.103 In later Sections of this Article, we discuss and suggest ways to mitigate the risk that Meta will oppose the OB for making similar moves.

2. Differences

There are also a variety of differences between the Oversight Board and human rights tribunals. We focus here on four distinctions that are likely to have significant influence on the Board’s future trajectory.

The first, and most obvious, difference is that Meta is a private corporation and not a state.104 Some limits on the Board’s jurisdiction reflect the fact that, unlike human rights tribunals, the Board oversees a private company that can incur liability as a result of its activities around the world. States are responsible for violations of international law attributable to them, and their officials may face civil or criminal liability for such violations. However, a commercial entity like Meta also has obligations to shareholders that require it to avoid foreseeable financial or legal risks, which include retaining on its platforms content that violates domestic law. In addition, while human rights treaties are negotiated and ratified by multiple states, the Board has only a single creator. This enhances Meta’s ability to ignore Board decisions and policy recommendations that, in the company’s view, are too aggressive in promoting accountability or developing international norms.105

A second distinction is the vast financial support that Meta has provided to the Board ($280 million as of 2022, as noted above) in comparison to the much lesser funding that states provide to human rights tribunals. Only the ECtHR has anything close to the OB’s budget (€74.5 million for 2022),106 but it reviews a vastly greater number of applications each year (over 36,000 in 2021)107 than does the OB (20 in 2021).108 In comparison, other international human rights tribunals are chronically starved for resources.109

A third difference relates to the ability to screen and select cases for review. Human rights tribunals apply admissibility criteria to determine whether they can adjudicate a petition on the merits. Among the most common is the exhaustion of domestic remedies, a rule that is analogous to requiring a Facebook user to file an internal challenge to the company’s content moderation decision before bringing a case to the Board.110 Once the admissibility criteria have been met, however, the tribunals have no discretion to not hear a case (although they can dispose of petitions without published decisions).111

In contrast, the OB selects which cases it reviews. The Board has prioritized “cases that had a potential to affect lots of users around the world, were critically important to public discourse or raised important questions about Meta’s policies.”112 These included cases involving automated moderation, Meta’s “dangerous individuals and organizations” standard, and press freedom.113 In addition, the Board has selected cases that “raised major implications for applying international human rights standards to moderating content at global scale,” including cases “from outside the Global North.”114

A fourth difference concerns remedies. The Oversight Board’s remedial powers in individual cases are narrow: it can direct Meta to uphold or reverse the company’s internal determination to retain or take down content.115 The remedies awarded by the tribunals are far more capacious. They include declaring that rights have been violated, awarding compensation, and recommending that a state take concrete actions, such as releasing an individual from detention, reopening judicial proceedings, and adopting legislative or other systemic measures to remedy the structural causes of treaty violations.116

D. The Impact of International Human Rights Tribunals: Accountability Promotion and Norm Development

For some commentators, comparing the Oversight Board to international human rights tribunals may reinforce their view that the OB is unlikely to have much of an impact. An extensive literature has critiqued the effectiveness of the tribunals on multiple grounds, including the proliferation of monitoring bodies, the fragmented jurisprudence they produce, their lack of enforcement powers, and the low levels of compliance with their decisions.117 Other studies, in contrast, note that the existence of multiple bodies enables advocates to strategically choose the forum that is most advantageous for seeking accountability.118 They also identify the tribunals’ positive impact on applying international law to new contexts and enabling civil society organizations to push governments to implement human rights protections domestically.119

For those familiar with domestic legal systems, it may seem obvious that the effectiveness of any adjudicatory body depends on whether litigants obey its rulings. In the international arena, however, whether the parties to a dispute follow the orders and provide the remedies that a tribunal awards is only one measure of impact, and arguably not the most important one.120 Other measures of effectiveness include whether tribunal decisions deter future violations;121 influence behavior in a rights-enhancing direction;122 become embedded in domestic law, thus enabling violations to be remedied at home;123 promote the systemic integration of international law;124 and adapt human rights agreements to changing circumstances.125

The wide variety of these metrics reflects the fact that international human rights tribunals engage in rights-promoting activities beyond adjudicating individual complaints. They also investigate systemic violations, review state reports, issue recommendations to improve compliance, and provide interpretive guidance in general comments and advisory opinions.126 The precise functions vary from one institution to another. In general, however, all international human rights tribunals share two overarching objectives: (1) providing a venue to hold states accountable for legal violations, and (2) developing international norms in response to longstanding and emerging human rights issues.

The Sections that follow discuss how human rights tribunals promote accountability and develop international norms and explain how those functions have evolved over time. This sets the stage for our proposal for the Oversight Board to prioritize similar goals.

1. Promoting Accountability

Despite the limits that states initially placed on the powers of international human rights tribunals, these bodies have, over time, incrementally expanded their authority to adjudicate complaints and monitor implementation and compliance. These expansions have enhanced the tribunals’ ability to hold states accountable for violations of human rights law. Although states have not explicitly authorized these expansions, they have generally acquiesced in the tribunals’ augmented review powers, in part due to pressure from civil society organizations.127 A few examples illustrate these developments.

The ECtHR, the oldest and most active human rights tribunal, conceives of itself not merely as the judicial arm of the European Convention on Human Rights but as the guardian of a “constitutional instrument of European public order.”128 The court has adopted numerous doctrinal innovations that reflect this elevated status. It has severed reservations that it deems incompatible with the Convention’s object and purpose, treated interim measures as legally binding notwithstanding states’ positions that such measures are only recommendations, and created an aggregate litigation mechanism to address systemic human rights problems.129

U.N. treaty bodies have also capaciously construed their mandates to review state party reports, formulate general comments, and issue decisions in response to individual complaints.130 For example, the “views” that the Human Rights Committee adopts in response to complaints from individuals are nonbinding, but the Committee expects states to follow them nonetheless. It has also taken concrete steps to monitor compliance, appointing a special rapporteur to evaluate state responses and collecting and publicizing compliance information. In these ways, the Human Rights Committee and other treaty bodies have pushed “the limits of [their] authority and sometimes arguably [gone] beyond it.”131

In the Americas, the Inter-American Court of Human Rights (IACtHR) has developed a range of responses to enhance its effectiveness, including issuing periodic reports on whether states have complied with its judgments and the remedies they award.132 The court also created the doctrine of conventionality control, which requires all domestic governmental actors to give effect to the American Convention on Human Rights as interpreted by the IACtHR. The doctrine reflects a particularly strong view of international judicial supremacy, one in which national judges are compliance partners for the court’s binding judgments and advisory opinions.133

2. Developing International Norms

In addition to enhancing their ability to hold states accountable for violations, the tribunals also contribute to the promotion and protection of human rights via interpretation. The development of international norms stems from a shared conception of human rights treaties as “living instrument[s]”134 that “tak[e] into account the circumstances of contemporary society,”135 as well as progressive trends at the global, regional, and national levels.136 Such interpretations have encompassed a wide array of topics, including the right to life, inhuman and degrading treatment, discrimination against LGBT persons, conscientious objection to military service, and collective bargaining by trade unions.137 In these and other areas, the tribunals have viewed with increasing skepticism rights-restrictive laws and policies once accepted as within a state’s discretion.

Two examples illustrate these trends. The first concerns how consensus emerged around an expanded definition of rape as a form of torture, a practice prohibited in many human rights treaties. As one scholar recently explained:

[W]hen the Special Rapporteur first asserted, in 1986, that rape can constitute torture, this idea was quickly taken up by other bodies. Twelve years later, the ICTY, the ICTR, the IACHR and the ECtHR had all found that the prohibition of torture and ill-treatment had been violated by acts of rape.138

A second example focuses on the general comments of the Committee on Economic, Social and Cultural Rights. Over several decades, the Committee has developed a range of doctrines and principles—including the nonderogable minimum core; obligations to respect, protect, and fulfill; and the extension of duties to private actors—that have augmented and concretized provisions of the International Covenant on Economic, Social and Cultural Rights.139 Individuals and civil society groups, in turn, have relied on these normative developments in litigation, especially in countries in the Global South, to progressively interpret economic and social rights in national constitutions.140 Studies have documented similar examples of the international and domestic diffusion of norms developed by other U.N. treaty bodies and regional human rights courts.141

As these examples reveal, normative development often begins within a single human rights treaty. Over time, however, evolutive interpretations can spread to other international human rights conventions, as well as within and across domestic legal systems. The result is a force multiplier effect that can dramatically enhance the influence and impact of human rights norms. As we later explain, a similar trend is emerging for the Oversight Board, whose normative output is already being cited within the U.N. human rights system.

II. The Oversight Board’s Human Rights Future

As we have explained, commentators have criticized the Oversight Board on numerous grounds, including its narrow jurisdiction, limited capacity to review user challenges, and inability to compel Meta to change its policies. Similar criticisms have been leveled at international human rights tribunals. Yet as the previous Section has shown, the tribunals have incrementally enhanced their authority and progressively expanded international norms to increase their influence over governments.

We now propose a range of strategies for the Oversight Board to chart a similar trajectory. We begin by explaining how the OB has already begun to hold Meta accountable for its content moderation decisions and policies. We then discuss how the Board is developing international human rights norms, in particular with respect to how these norms apply to private social media companies. We then identify strategies for the Board to continue along this trajectory, drawing where appropriate on principles and doctrines that international human rights tribunals have developed.

A. Accountability Promotion

1. Jurisdiction

The Oversight Board’s decisions thus far reflect a fairly broad interpretation of its own jurisdiction. From the beginning, it rejected Meta’s argument that an appeal is rendered moot by the company’s removal or restoration of the content at issue.142 It has also expansively exercised its authority to provide Meta with policy guidance, issuing robust recommendations in nearly every case even though this function appears to have been intended as subsidiary to its review of individual decisions.143 Despite fears that the Board would not be able to address the ways in which Meta’s recommendation systems affect human rights, the OB has capaciously interpreted its policy advisory function to recommend that the company “commission a human rights impact assessment on how its newsfeed, recommendation algorithms, and other features amplify harmful health misinformation and its impacts.”144

In the future, we expect that the Board will continue to face borderline cases in which it will need to determine its own jurisdiction. All international tribunals have such authority. Although the exercise of this power is mostly uncontroversial, a tribunal may encounter “Kompetenz-Kompetenz” disputes over whether to expand its jurisdictional reach.145

For example, the OB may need to clarify its jurisdiction with respect to content removed by Meta to comply with domestic law. Some commentators claim that this exception precludes the Board from “consider[ing] cases that human rights activists care most deeply about: state censorship of content, particularly political speech, in flagrant violation of international human rights standards on freedom of expression.”146 In our view, however, the OB’s founding documents provide leeway for the Board to hear cases raising these issues. At a minimum, the Board has the authority to determine whether the criteria for this exception have been met.

The Charter provides that nothing in the OB’s founding documents “shall be interpreted in a manner that would result in a violation of law by Facebook” or any of its associated entities.147 The Bylaws, in turn, identify three categories of cases that “will not be eligible for the board to review”:

  • Where the underlying content has already been blocked, following the receipt of a valid report of illegality, and not removed for a Community Standards violation;

  • Where the underlying content is criminally unlawful in a jurisdiction with a connection to the content (such as the jurisdiction of the posting party and/or the reporting party) and where a board decision to allow the content on the platform could lead to criminal liability for Facebook, Facebook employees, the administration, or the board’s members; or

  • Where the underlying content is unlawful in a jurisdiction with a connection to the content (such as the jurisdiction of the posting party and/or the reporting party) and where a board decision to allow the content on the platform could lead to adverse governmental action against Facebook, Facebook employees, the administration, or the board’s members.148

As an initial matter, the three exceptions to the OB’s jurisdiction in the Bylaws are broader than those in the Charter, which only requires the Board to refrain from decisions that would result in liability for Meta. However, the exceptions in the Bylaws are more constrained than they may initially appear. The first exception excludes review only when content has been blocked pursuant to a “valid report of illegality.” The practices of both Meta and the Board indicate that this exception only prevents review of content that is removed directly pursuant to a government request. When the OB has asked the company about the existence of such a request, Meta has provided a substantive response.149 In addition, Meta and the Board agree that this exception does not include situations in which the content was flagged by a government entity but was later removed for a Community Standards violation.150

Perhaps the strongest indication that the “valid report of illegality” exception will be interpreted narrowly appears in a 2022 decision of the Board, referred by Meta, in which the company removed a drill music video that had been brought to its attention by an internet referral unit in the United Kingdom.151 Such law enforcement offices, which use internet providers’ terms of service to seek removal of content that the government deems objectionable,152 raise serious human rights concerns.153 The Board found there was insufficient evidence of a credible veiled threat and issued several recommendations for how Meta should respond to such requests in the future.154

If the Board hears cases implicating other exceptions to its jurisdiction, we believe it should determine for itself whether the exceptions are applicable. In addition to demonstrating its independence and rejecting Meta’s ability to exclude cases from Board review simply by asserting the existence of an exception, this approach would have normative benefits. For example, it would provide the Board with an opportunity to address an unresolved tension in the U.N. Guiding Principles on Business and Human Rights (“U.N. Guiding Principles” or “UNGPs”): the obligation of businesses to follow local law while, at the same time, respecting human rights.155 In determining what constitutes a “valid report of illegality,” for example, the OB might draw on nonbinding standards such as the Manila Principles on Intermediary Liability156 or the Global Network Initiative (GNI) Implementation Guidelines.157 More audaciously, the Board might conclude that Meta may disregard domestic criminal or civil legislation that an international tribunal has found to violate human rights, as is the case with several national media laws in Africa.158

We acknowledge, however, that compliance with local law is a sensitive issue that many transnational corporations face. The Board should thus proceed carefully, recognizing that Meta has compliance expertise and a range of business interests affected by domestic legal requirements. However, nothing in the Charter or Bylaws precludes the OB from asking the company to justify its invocation of an exception or provide information about how it responds to government requests and claims of illegality. Nor do the Board’s founding documents prevent it from providing the policy guidance on the overarching issue that the exceptions raise: how Meta reconciles its human rights responsibilities and compliance with local law.

2. Standards of Review

Neither the Charter nor the Bylaws address what standard of review the Oversight Board should apply. This Section identifies a range of circumstances in which the OB will confront whether to give any deference to the decisions, actions, policies, or interpretations of Meta or of governments. We draw upon doctrines and approaches developed by international human rights tribunals to identify factors that the Board should consider when determining appropriate standards of review.

Like the OB’s founding documents, human rights treaties that create international tribunals are silent on standards of review.159 However, the tribunals’ accountability promotion and norm development functions weigh against giving any deference to states regarding the meaning of international law. Consistent with this perspective, the tribunals have generally applied de novo review when interpreting human rights agreements and customary law.160

At the same time, the tribunals have identified circumstances in which deference to governments may be appropriate. This is most famously—and controversially—reflected in the ECtHR’s margin of appreciation doctrine.161 Other human rights tribunals at least implicitly recognize that a degree of deference is appropriate in some contexts.162 This includes situations where “national authorities enjoy comparative institutional advantages”163—for example, in interpreting and applying domestic law, evaluating local conditions, or making certain factual findings—or where international rules permit those authorities “to strike a balance between competing rights and interests” or make “a choice between competing social values.”164

The accountability promotion and norm development functions that the Oversight Board shares with international tribunals favor giving no deference to Meta’s interpretation of international human rights law when the company sets content moderation policies or reviews user posts. The social media company has no special expertise in this regard, nor are there institutional or normative reasons for deference. Accordingly, the Board should exercise de novo review.

Whether the Board should give any deference to Meta’s assessment of whether a post violates its Values or Community Standards is a closer question. The caselaw to date suggests that the OB continues to wrestle with this question. In Video After Nigeria Church Attack, for example, the panel was divided over the question of whether Meta made the right judgment call in removing a video showing the aftermath of a terrorist attack that included shooting-related hashtags. Meta (and the minority of the OB) argued that the hashtags could be viewed as sadistic, glorifying violence, and mocking, and therefore should be removed under three different Community Standards. The majority disagreed. Apparently giving no deference to Meta’s determination, the majority concluded that the hashtags were only designed to raise awareness among airsoft and firearm enthusiasts and thus did not violate the Community Standards.165

Is de novo review of Community Standards appropriate? Meta retains the power under the Charter to modify its internal norms.166 This arguably includes the lesser power to interpret the Values and Community Standards. If the company adopts a general position regarding the meaning of those internal norms, there may be structural or functional reasons for deferring to its position. In assessing how much deference, if any, is appropriate in that context, relevant factors for the Board may include whether the interpretation is based on experience with moderating specific types of online speech, whether Meta follows that interpretation consistently, and whether the company has considered the potential disadvantages of its position.167 In contrast, the arguments for discretion are weaker when reviewing the decisions of front-line moderators applying Community Standards or Values to specific user posts.

The political and social context of user posts is another setting in which standard of review issues may arise. The Board appears to apply something akin to de novo review of Meta’s assessment of how contextual factors affect the likelihood of real-world harm. In Protest in India Against France, for example, a majority of the OB “did not find Facebook’s contextual rationale in relation to possible violence in India in this particular case [to be] compelling.”168 In contrast, a minority of the panel would have “defer[red] to Facebook’s determination that the post presented an unacceptable risk of promoting violence,”169 in part, because the company “relied upon a third party partner assessment” and “consulted regional and linguistic experts.”170

The majority’s position seems compelling. The Board has the resources to engage deeply with stakeholders and develop a nuanced understanding of context in the small number of cases it selects for review. In contrast, while Meta has broad experience working with local governments and evaluating demands for compliance with local law, its understanding of the real-world human rights impact of the speech on its platforms is more limited.171

Another standard of review issue relates to the Board’s evaluation of government decisions and actions. The OB obviously lacks expertise in all of the local laws that may be implicated by its oversight functions. However, it has substantial resources to commission experts to analyze the domestic laws and the decisions of public actors that are relevant to an appeal, a capacity that is important given the risk that governments may not be fully forthcoming. The Board can also initiate interactions with public agencies and officials. In UK Drill Music, for example, the Board received a comment from the London Metropolitan Police and it filed a Freedom of Information Act request to compel the law enforcement body “to provide information about its policies and practice on making requests to social media and streaming companies to review and/or remove content.”172 Taken together, these considerations suggest that the OB should give weight to a government’s understanding of its own laws while preserving the possibility of verifying how the law applies in particular cases.

3. Applicable Law

A third set of strategic issues relates to the legal rules and norms that the OB applies. As previously explained, the Charter directs the OB to “pay particular attention to the impact of removing content in light of human rights norms protecting free expression” when evaluating cases.173 Many Board decisions analyze freedom of expression in depth, as we discuss in detail below. However, the OB has not shied away from addressing other human rights. For example, it has applied the prohibition on gender and racial discrimination as well as the rights to privacy and to an effective remedy.174 The OB has also considered the obligations of the International Covenant on Economic, Social and Cultural Rights, such as the right to health and to participate in culture, as well as the Convention on the Rights of the Child, among other treaties.175

Board decisions have cited to and applied a wide range of international human rights sources. These include international conventions (most often the ICCPR and the International Convention on the Elimination of All Forms of Racial Discrimination), general comments and decisions of treaty bodies (in particular those of the Human Rights Committee), reports submitted to the U.N. Human Rights Council (especially those of the Special Rapporteur on the Right to Freedom of Opinion and Expression), and nonbinding human rights standards (notably, the UNGPs and the hate speech guidelines of the Rabat Plan of Action).176 By engaging with these sources and citing its own prior decisions applying them, the OB is developing a well-reasoned, consistent, and coherent body of human rights jurisprudence.177

There are several accountability-promoting advantages to the Board’s commitment to “plac[e] human rights at the heart of [its] work.”178 The first is stability. Meta can change its Values and Community Standards by applying its ordinary corporate decision-making processes. In contrast, the Charter cannot be amended without majority approval of the trustees and agreement of both the Board and Facebook.179 The Charter’s reference to international human rights law thus serves as an external check on Meta. A second benefit is contemporary relevance. As previously explained, the interpretation of human rights law evolves over time. The Board’s consistent practice of applying not only treaty texts but also the normative output of international tribunals and independent experts thus helps to ensure that it will continue to apply standards that respond to real-world problems and controversies.180

A more challenging issue concerns potential conflicts among the sources that the Board applies. The OB’s founding documents do not indicate a hierarchy among Community Standards, Values, and international human rights law. Most Board decisions to date apply the three sources independently of one another, giving each equal weight and finding no conflict among them.181 However, the Board has indicated that it will “prioritize human rights” when that law conflicts with Meta’s internal rules.182 In Pro-Navalny Protests in Russia, the Board overturned the company’s decision to remove a post by a supporter of the Russian opposition politician, finding that the removal was consistent with Community Standards but inconsistent with international human rights law and with Meta’s Values (which incorporate human rights).183 The OB has also made recommendations on how the company can “bring [its] Community Standard[s] in line with Meta’s values and international human rights standards.”184

If the Board faces other clashes between the company’s internal norms and human rights law, it can apply interpretive tools that international human rights tribunals have developed to mitigate these conflicts. For example, the tribunals presume that international organizations and treaty drafters do “not intend to impose any obligation on member States to breach fundamental principles of human rights.”185 They also apply a pro homine principle to construe textual ambiguities in favor of individual rights and against state sovereignty.186 And under the principle of lex specialis, the tribunals interpret general legal obligations in light of more specific commitments.187 We expect that there will be few inconsistencies among sources that the OB cannot resolve by applying these or other interpretive doctrines.

4. Evidentiary Issues

Once the Oversight Board has exercised jurisdiction and has identified the applicable legal norms, it must determine whether it has an adequate factual record to render a decision. The Board regularly receives information from a variety of sources. It often commissions expert reports on the socio-political and cultural context of the content at issue and receives public comments, including from U.N. Special Rapporteurs and advocacy organizations.188

Meta has thus far actively participated in cases reviewed by the Board. It has provided information about its decisions to remove or retain user posts as well as its content moderation and internal appeals processes.189 In several cases, however, the OB has expressed frustration with the limited information that Meta has provided in response to its requests.190 The company’s disclosures and omissions have also revealed inconsistencies between Meta’s publicly available policies and the rules applied by its content moderators, resulting in Board recommendations to close these gaps and provide greater transparency.191

The OB has yet to address when information in Meta’s possession is “reasonably required for the board to make a decision,” nor has it considered situations in which the company has withheld information “in compliance with applicable legal and privacy restrictions.”192 The Charter does not define these key phrases. To fill in the gaps, the Board may seek guidance from international human rights tribunals that have faced similar issues.

States are expected to cooperate fully in international proceedings.193 Significantly, the tribunal, not the government, determines whether the state has produced all relevant information or conducted an adequate investigation.194 This is so even where the government raises confidentiality or security objections. In such cases, the tribunal undertakes an independent verification as to whether or not there had actually “existed reasonable and solid grounds for treating the documents in question as secret or confidential.”195 When states fail to live up to these obligations, the tribunals have responded by creating evidentiary presumptions to incentivize disclosure. Once an applicant makes out a prima facie case, the failure to produce information plausibly within the state’s control or to conduct an adequate investigation will result in a determination that the applicant’s plausible allegations have been proven.196 The Oversight Board should follow this approach if it confronts similar evidentiary disputes with Meta, including whether confidentiality or security concerns justify withholding relevant information.

5. Compliance

As previously explained, the Charter adopts a narrow definition of compliance: OB decisions are legally binding on Meta only in each specific case. The company has committed to implement the Board’s directives to restore or remove content within seven days, unless doing so “could violate the law,” and to publicize its responses.197 Meta must also analyze “whether it is technically and operationally feasible to apply the board’s decision” to “identical content with parallel context.”198 With regard to policy guidance, the Board’s Charter and Bylaws commit Meta to “considering it in the formal policy development process” and providing a public response within sixty days.199

The OB has applied these provisions to expand the influence of its binding decisions and nonbinding policy recommendations. Most notably, it has created institutional structures to monitor implementation, including a Case Implementation and Monitoring Team, an Implementation Working Group, and an Implementation Committee.200 These bodies enable the Board to press Meta on unresolved compliance issues.201 In addition, the OB is structuring its decisions in ways that facilitate tracking of Meta’s implementation.202 For example, the Board numbers its policy recommendations and asks the company to address each of them separately,203 specifies the actions that would constitute implementation of its recommendations,204 and highlights issues of ongoing noncompliance.205 The Board also asks Meta to provide information to facilitate disclosures that aid its accountability efforts.206

Like many human rights tribunals, the Board is using its reporting function—its quarterly and annual reports—to publicize Meta’s compliance record. It praises the company for its progress, identifies areas where Meta has “misrepresented or misunderstood” the Board’s decisions, and tracks the company’s implementation.207 It has also developed criteria for measuring Meta’s compliance and adopted a grading system to indicate whether Meta is fully responsive, partially responsive, or nonresponsive in relation to each of its decisions and policy recommendations.208

The Board is also pushing the company to define its own rules in the first instance. In Former President Trump’s Suspension, for example, the OB refused to pass judgment on Meta’s indefinite suspension of the former U.S. President because the company had not followed its policies regarding account suspension.209 Requiring Meta to “apply and justify a defined penalty”210 prior to review incentivizes the company to articulate and comply with its own policies and safeguards the Board’s independent supervisory role. The Board also appears to be using its policy advisory authority to push the company to generate information that researchers and activists can use to evaluate Meta’s human rights impacts.211

The transparency and dialogue engendered by the Board’s approach serves several functions. Identifying areas of compliance increases the reputational benefits to the company of following OB decisions and recommendations. Conversely, calling out inconsistencies and highlighting areas for improvement raise the costs of noncompliance and provide information that civil society groups can use to advocate for reforms. These consequences are likely to increase as the Board issues decisions clarifying how international human rights law applies to private social media companies, an issue to which we now turn.

B. Norm Development

The Oversight Board has the potential to make important contributions to developing international human rights norms, particularly as they apply to business entities in the digital sphere. This Section explains why the OB is well positioned to develop these norms, reviews indications that it is already doing so in its caselaw, and suggests how the Board can influence the activities of other human rights bodies as well as government efforts to regulate corporations and social media platforms.

1. The Oversight Board as Norm Generator

There is an urgent need to develop human rights norms, especially freedom of expression, as applied to private actors in the online environment. Commentators have made persuasive arguments that such norms should apply to social media platforms. However, there are several areas in which further development is needed.212

Although private companies have human rights responsibilities under the U.N. Guiding Principles, these are different from government obligations regarding freedom of expression.213 Speech restrictions imposed by platforms do not carry the coercive power that accompanies state action, and the Board has cited this as justification for upholding Meta’s removal of content that a state could not restrict in similar circumstances.214 Platforms also have different interests than governments: they pursue commercial objectives rather than societal welfare, and their own speech interests also play a role. It remains unclear how these interests should be accommodated within the human rights framework governing free expression. The online environment also affects the proportionality analysis in crucial ways.215 For example, the large volume of information collected by social media companies may provide an evidentiary basis for crafting speech norms that more precisely target communications that cause real-world harms.216 These and other differences have important practical consequences for the development of human rights norms relating to online expression.

A few national courts have evaluated social media content moderation under constitutional or fundamental rights standards.217 The Oversight Board is well-placed to expand upon these developments by examining how human rights apply to non-state actors in the digital realm across a range of specific contexts. The OB’s structure facilitates case-specific adjudication: it has sufficient resources to gather relevant information, and its members are professionally and geographically diverse.218 The Board’s ability to issue nonbinding recommendations also enables it to provide more far-reaching guidance than would be possible in a single binding decision.

2. Norm Development in OB Caselaw

The Board’s thirty-nine decisions (as of May 2023)219 have tackled a range of important and unsettled human rights issues. This Section provides an overview of OB caselaw addressing four such issues: corporate respect for freedom of expression, Meta’s due diligence obligations, the cumulative harms of hate speech, and the importance of context in content moderation.220

The Board has identified several ways in which freedom of expression principles apply differently to corporations than to states. In upholding the removal of slurs or other hateful content, for example, the OB has emphasized that a private company has broader discretion to restrict speech than a state would have in a similar context, as long as the company articulates the basis for its restriction and ensures that its determinations are necessary and proportionate.221 In particular, the Board has indicated that removing speech from the platform is a less severe sanction than the civil or criminal penalties applied by states and that Meta has different obligations, in part, due to the complexities of content moderation at scale.222 The Board has also reasoned that social media companies have heightened responsibilities when they hold a dominant position in a market or when they are the primary source of news in a particular location.223 Finally, it has recommended that Meta develop tools to allow users to customize their online experience as a way to respond more narrowly to harmful content.224

Board decisions also elaborate on Meta’s due diligence responsibilities under the U.N. Guiding Principles. The UNGPs direct business enterprises to assess and mitigate the human rights harms that they cause, contribute to, or that are “directly linked” to their operations.225 The OB has held that Meta’s due diligence obligations are heightened in areas of conflict,226 as well as in situations of increased risk of harm,227 and when national protections of human rights, particularly freedom of expression, are weak.228 In such situations, the consequences of error—either of failing to remove harmful content or of removing benign content—are high, and the company must evaluate close cases carefully rather than relying on default rules.229 The Board has also provided insights on the kinds of risks that Meta should consider,230 as well as the company’s obligations with respect to automated decision-making and machine learning, technical design, transparency reporting, and content moderation processes and procedures.231 The OB has also emphasized that automatic closure of appeals to Meta may deprive users of remedies.232

Although commentators have raised concerns about the Board’s ability to consider cumulative harms while reviewing individual cases, the OB has discussed such harms in several hate speech decisions. In Depiction of Zwarte Piet, for example, the Board upheld Meta’s removal of a post containing a blackface character on the grounds that “moderating content to address the cumulative harms of hate speech, even where the expression does not directly incite violence or discrimination, can be consistent with Facebook’s human rights responsibilities in certain circumstances.”233 The majority emphasized not only the psychological impact of repeated negative stereotypes but also the way in which the accumulation of such stereotypes “creates an environment where acts of violence are more likely to be tolerated and reproduce discrimination in a society.”234

Perhaps the most significant development in the Board’s jurisprudence, however, is its insistence that Meta do more to consider the context of speech, despite the complexities of doing so at scale.235 Although a number of recent innovations have improved Meta’s ability to tailor the internal guidance it provides to front-line moderators,236 the volume of content that these moderators review precludes them from considering posts in context or exercising discretion in applying this guidance.237 The company assesses the broader socio-political context only when a challenged post is elevated to the policy staff level,238 although its cross-check program increases the availability of human review for certain types of content.239

In case after case, the Board has been pushing back, emphasizing that “[c]ontext is key” and criticizing content moderation practices and policies that ignore contextual factors.240 In Pro-Navalny Protests in Russia, for example, a supporter of the Russian opposition politician who had been attacked online called out the attacker as a “cowardly bot.”241 Meta removed this response under its anti-bullying policy. The Board did not mince words in overturning the removal: “Facebook’s blunt and decontextualized approach can disproportionately restrict freedom of expression.”242

The Board has also highlighted a range of contextual cues that Meta should consider in assessing the harms of speech. These factors include whether a post relates to a situation of armed conflict,243 how widely it is disseminated,244 the response of other users to the post,245 whether threats or calls for exclusion of a particular group are implicit or explicit,246 the location of the user,247 the timing of the post with respect to offline events,248 the language of the post,249 whether the post incorporated a meme,250 whether the content is “used in protest contexts or other crisis situations where the role of government is a topic of political debate,”251 the existence of “government campaigns of disinformation against opponents and their supporters” in a particular country,252 the history of the region,253 the identity or reputation of the poster,254 the relative power dynamics between the involved parties,255 and inferences about the user’s intent from the language of the post.256

The disagreements between the Board and Meta about the extent to which the company considers context are significant. Meta’s failure to account for context has been at the heart of key debates over its human rights record, from its insufficient Burmese-speaking reviewers during the genocide in Myanmar to its initial across-the-board ban on content depicting female nipples.257 Although the company has invested in additional language capacity, built greater complexity into its rules, created a process to provide additional human review of potential false positives,258 and developed regionally specific lists of slurs,259 its front-line content moderation remains focused on the post itself. The Board—while acknowledging the challenges of context for such a vast platform—has not budged in its insistence that Meta take greater account of external considerations.260 Whether Meta will continue to revise its processes to give meaningful attention to context could be one of the most important future tests of the OB’s legitimacy.

3. The Impact of Norm Generation on Human Rights

The previous Section described how the Oversight Board is addressing a range of unsettled legal issues that arise at the intersection of human rights and content moderation. This Section considers the mechanisms by which these normative developments can affect human rights more broadly. First, OB decisions can provide focal points for civil society groups to mobilize for rights-enhancing legal and policy reforms. Second, the Board’s caselaw enables it to engage in a dialogue with other actors who promote the human rights responsibilities of internet platforms. Third, the norms developed by the OB may influence efforts to regulate social media companies in national and regional legislation.

a. Norms as Resources for Mobilization

The relationship between international human rights tribunals and civil society is “interactive and mutually constitutive.”261 The tribunals (particularly in light of their limited funding) rely on civil society groups to identify, file, and support the cases, reports, and other interventions that provide opportunities to promote accountability and develop international norms. These organizations, in turn, mobilize around favorable tribunal decisions, incorporating them into broader rights-claiming strategies in domestic and transnational venues and pressuring governments to reform restrictive laws and policies in line with the tribunals’ progressive interpretations of human rights law.262

The Oversight Board receives input from civil society organizations, and the norms it develops are likely to be important resources for those organizations. A wide range of individuals and groups have weighed in on every OB case, from the U.N. Special Rapporteur on Minority Issues to academics and advocacy groups.263 However, third parties—both non-Meta users as well as civil society and digital rights advocates—cannot appeal to the Board and do not provide input on case selection.264 We also found no evidence that civil society groups are coordinating with Meta users to “tee up” cases for OB review.

The Board’s greater resources mean that it will need to rely less on civil society for information and support than human rights tribunals—ostensibly an advantage, but one that will limit the input the Board may receive about where its attention is most needed. Less reliance on civil society may mean these groups will have fewer opportunities to direct the Board to issues of concern and cases that would be valuable to their advocacy efforts. We thus recommend that the Board’s Engagement Team establish additional channels of communication to civil society groups to help inform the OB’s case selection strategy.

b. Dialogue Promotion

Norm generation can also influence human rights by promoting a dialogue between rights-holders and duty-bearers.265 Observers of the Oversight Board have expressed similar hopes for dialogue among Meta, its users, and other actors affected by its content moderation policies. evelyn douek, for example, argues that “[b]ecause the substantive rules for platform standards will almost always be a matter of reasonable disagreement and contestation, legitimacy for content moderation systems needs to be earned not by unilaterally issuing ‘correct’ rules but by channeling debate about what those rules should be.”266

Thus far, the Oversight Board’s external engagement has been robust—even more so than the public outreach by human rights tribunals. The OB publicizes the appeals it accepts and solicits broad input on these cases using a simple and accessible template.267 The Board also has an Engagement Team that communicates with external audiences.268 It holds “office hours” to discuss how to submit comments regarding the cases it has accepted,269 organizes stakeholder meetings focused on important issues,270 and shares information about the sources it relies on in its decisions.271

The Board also actively engages with experts in deciding cases. In the Swedish Journalist Reporting Sexual Violence Against Minors, for example, the OB relied heavily on input from a virtual roundtable with seven advocacy organizations working on child sexual exploitation.272 Similarly, in Reclaiming Arabic Words, “members of the Board held informative and enriching discussions with organizations that work on freedom of expression and the rights of LGBTQIA+ people, including Arabic speakers.”273 The OB has not made explicit the extent to which it relies on these interventions, but several of its conclusions track information submitted by third parties.274

However, the Board has yet to evaluate the extent to which these efforts may privilege certain stakeholders or voices over others. Just as Meta’s outreach is directed by and filtered through its staff, the Oversight Board’s staff act as gatekeepers who influence from whom the Board hears.275 The addition of a “user advocate,” as recommended by BSR, would help to improve outreach, especially to users who are members of vulnerable communities.276 The inability of civil society organizations to file appeals also hampers the extent to which these groups can proactively initiate interactions with the Board.

In addition, there is one important constituency not currently included in this dialogue: individuals who do not have Facebook or Instagram accounts.277 Non-users can be deeply affected by the content moderation policies of these platforms, and denying them the opportunity for input undermines the Board’s legitimacy. Until the Charter is revised to permit appeals by non-users or otherwise formalize their involvement, the Board should consider affirmatively seeking input on case selection from advocacy organizations on issues that may not generate appeals from users or Meta.

The Board’s contribution to dialogue would also benefit from more discussion of disagreements among its members. Regional human rights courts and U.N. treaty bodies allow their members to file, under their own names, concurring or dissenting opinions.278 Scholars have identified numerous benefits of such separate opinions, such as improving deliberation and the quality of a tribunal’s reasoning; enhancing judicial independence and transparency; and providing information to other tribunals, future litigants, and the public about alternative future trajectories of the law.279

The Board’s Bylaws provide that OB decisions should “include any concurring or dissenting viewpoints, if the panel cannot reach consensus.”280 Currently, such viewpoints are summarized as part of the written decisions issued by the five-member panel of the Board that hears each case. Although the summaries of minority opinions have been getting longer in recent cases, they are still generally less developed than the majority opinions, and they do not indicate the number of OB members that support the minority position.281 Allowing individual Board members to write separate dissenting or concurring opinions will help identify areas where norms are in flux and would promote dialogue about those junctures.

c. Norm Diffusion

The norms that the Board develops in its caselaw may also have a significant impact on the wider international human rights system.282 Prior to the Board’s creation, a few commentators highlighted the normative potential of social media platforms applying human rights standards. Evelyn Mary Aswad (who later became a member and then co-chair of the Board) observed that if companies applied international human rights law to their content moderation decisions, “there could be an active fountain of new ‘jurisprudence’ involving the ICCPR’s speech protections, which could influence the direction of international freedom of expression rights.”283 Sejal Parmar also queried how OB decisions would affect the work of activists, U.N. bodies, national courts, and international freedom of expression standards.284

Our review of the over three dozen OB cases through May 2023 reveals that the Board is already influencing the development of human rights norms at the global, regional, and national levels.285 In the United Nations, for example, a report of the High Commissioner for Human Rights highlighted the Board’s application of the Rabat Plan of Action, a nonbinding standard for hate speech that the High Commissioner had adopted a decade earlier.286 The Special Rapporteur on Minority Issues reviewed several early OB decisions and praised the Board’s “overarching commitment to protecting vulnerable and marginalised groups.”287 And the Special Rapporteur on the Promotion and Protection of the Right to Freedom of Opinion and Expression highlighted the Board’s potential to offer “valuable lessons” for efforts to remedy “wrongful actions taken on the basis of disinformation or misinformation.”288

The Board’s normative output is also relevant to U.N. treaty bodies, especially the Human Rights Committee, whose General Comment No. 34 on freedoms of opinion and expression is frequently cited to and applied by the OB.289 In addition, the Committee on the Elimination of Discrimination Against Women and the Committee on the Elimination of Racial Discrimination may be interested in the Board’s interpretations of the rights of women and racial minorities. Several OB decisions have opined on how certain manifestations of online violence can reflect discrimination and structural inequality.290 In Gender Identity and Nudity, the Board relied on decisions of the Human Rights Committee as well as the Convention on the Elimination of All Forms of Discrimination Against Women and the UNGPs to conclude that Meta’s Adult Nudity and Sexual Activity Community Standard resulted in “disparate opportunities for expression being made available for women, trans, and gender non-binary people on its platforms.”291

Board decisions applying the UNGPs to social media companies may be especially useful to the Working Group on Business and Human Rights, created in 2011 to focus on the implementation of these principles.292 Another U.N. venue is the Open-Ended Intergovernmental Working Group on Transnational Corporations and Other Business Enterprises with Respect to Human Rights, which is elaborating a binding legal instrument on the human rights responsibilities of private companies.293 The OB caselaw could, for example, inform the drafting of treaty provisions on the due diligence requirements of digital platforms.

Finally, OB decisions may also influence the efforts of supranational and national lawmakers to regulate internet platforms. In late 2022, the European Union (EU) approved the Digital Services Act (DSA), which imposes a comprehensive set of new legal obligations on social media companies and authorizes the EU, its member states, and platform users to enforce them.294 With regard to content moderation, the DSA mandates that digital service providers expeditiously delete posts that are illegal under national or EU law, while also empowering users to challenge removal decisions within the companies, in out-of-court dispute settlement proceedings, and by filing suits in court.295 Although precisely how these review processes will function will not be resolved until the DSA is transposed into national law, commentators have already noted the relationship between the out-of-court dispute settlement provision and the Oversight Board.296

National jurisdictions are also increasingly active in this regulatory space. In 2019, for example, the United Kingdom released the Online Harms White Paper, which envisions legislation regulating online platforms, including through the creation of a new duty of care for social media platforms.297 The French Corporate Duty of Vigilance Law, adopted two years earlier, requires large French companies, or companies with large French subsidiaries, to publish a “vigilance plan” that “establish[es] effective measures to identify risks and prevent severe impacts on human rights.”298

For both supranational and national legislation, Oversight Board jurisprudence may provide guidance with respect to due diligence, how platforms distinguish between lawful and unlawful content, and the potential systematic effects of human rights harms created or facilitated by social media companies. In addition, national courts may look to OB decisions when interpreting such legislation or when reviewing the human rights impact of social media platforms in other areas of law.

III.Risks and Challenges

We have thus far portrayed the Oversight Board in mostly positive terms. In particular, we have discussed the Board’s potential to enhance Meta’s accountability and to develop a case-specific jurisprudence that explains how international human rights law developed for states applies to private social media companies. In fact, the OB’s human rights future is rapidly becoming its human rights present. When we began the research for this Article, many of the proposals that we discussed seemed far off prospects. Yet the Board’s first annual report reveals that the OB is already implementing a number of these strategies as well as embracing human rights as its normative polestar.299

Although we are sanguine about the Oversight Board’s trajectory, it is also important to acknowledge the risks and challenges that the institution is facing or may soon confront. This Part considers four of these issues: (1) pushback and backlash; (2) whitewashing; (3) indeterminacy and legitimacy; and (4) negative spillover. We describe these risks and challenges below, drawing on the experiences of international human rights tribunals and offering a few preliminary suggestions for mitigating them.

A. Pushback and Backlash

The early fears of some commentators that Meta would simply ignore costly or inconvenient Oversight Board decisions and recommendations have not come to pass. The company has complied with all OB rulings ordering it to restore or take down content and has implemented many of the Board’s policy recommendations.300 As the Board’s 2021 Annual Report notes, Meta has made significant changes in response to these recommendations, including informing users about automated review, disclosing information about government requests, and translating its rules into additional languages.301

Meta and the OB are also having constructive interactions about the Board’s future and its sometimes far-reaching policy guidance. The Annual Report asserts that in 2021, “[t]he Board got better at making recommendations and Meta improved at responding to them.”302 This was so even as the company complained about time and resources needed for implementation and argued that the “size and scope of the Board’s recommendations go beyond the policy guidance that we first anticipated when we set up the board.”303 Equally as important, Meta has agreed to provide ongoing financial support for the Board, recently pledging $150 million to the Oversight Board Trust’s initial $130 million endowment to ensure that the OB continues beyond its six-year startup period.304

Yet pushback and backlash remain real concerns for the Oversight Board, as they do for international human rights tribunals.305 As we have explained, the company can modify the Values and Community Standards that the OB applies,306 although it has far less power to change the OB’s jurisdiction or access rules and no control over the content of international human rights law or who is appointed to the Board. Meta can also signal (and, some would argue, is already signaling) its displeasure in less formal ways, such as by “slow walking” compliance with decisions, declining to implement policy recommendations, misreading the Board’s policy decisions,307 and resisting its requests for information.308

The most notable example of pushback that the Board has encountered thus far involved Meta’s decision to withdraw its request for guidance on content moderation related to the Ukraine-Russia war.309 The OB expressed its “disappoint[ment]”310 in the decision—which purportedly stemmed from concerns for the safety of Meta’s employees in Russia311—but did not otherwise challenge it publicly.312 Commentators were quick to criticize the Board’s response. According to one, to ask the Board “for its advice—advice that would not even be binding on the company—and then decide belatedly that such advice might be dangerous calls into question the point of the entire enterprise.”313 Another noted that the incident “sets a precedent for Meta to continue withdrawing requests in the future.”314

Two months after the withdrawal, however, the Board accepted a user appeal of Meta’s removal of a post comparing the Russian army in Ukraine to Nazis and quoting a poem that calls for the killing of fascists.315 It overturned the removal and recommended that the company revise its policies to consider unlawful military intervention as relevant context.316

Given the other areas in which the OB is incrementally “push[ing] Meta in the right direction” on human rights issues,317 acquiescing in the company’s withdrawal of the Ukraine request while later ruling on a user appeal raising similar issues was strategically astute. The Board is likely to be more effective in enhancing Meta’s accountability to users and generating new norms if it proceeds incrementally, choosing its battles carefully and promoting its achievements to a wider audience. Such an approach may also, as we now explain, help to convince skeptics who believe that “the Oversight Board’s only role is to handle the easy questions.”318

B. Whitewashing

Meta’s creation of an independent review mechanism has generated a wide range of responses. Although some human rights experts, civil society groups, and scholars have supported the initiative, participated in consultations leading to the creation of the Board, or submitted third-party interventions, others have been skeptical of or even hostile to the company’s and the Board’s embrace of human rights. A significant concern is that Meta is pledging fealty to international human rights law as a public relations fig leaf to obscure the harms of its content moderation policies and undermine prospects for meaningful regulation or reform.

According to proponents of this whitewashing claim, platforms receive “undeserved legitimacy dividends . . . [by] wrap[ping] themselves in the language of [human rights] even as what is required by that body of norms remains indeterminate and contested.”319 They may also “try to co-opt the vocabulary of human rights to legitimize minor reforms at the expense of undertaking more structural or systemic changes to their moderation processes.”320 For example, the OB Charter refers not to international human rights law, but to human rights norms, perhaps indicating a reluctance by Meta to fully commit itself to legal constraints.321 Lastly, skeptics charge that corporations can manipulate the rules and doctrines of human rights law—such as the legality, legitimate aim, and proportionality requirements of ICCPR Article 19—to serve their own financial and businesses interests.322

These are serious concerns, and Meta’s first Human Rights Report, published in July 2022, does nothing to alleviate them.323 According to numerous civil society critics, this self-congratulatory document “whitewash[es] a history of fomenting hatred, violence, and extremism across the globe” and “present[s] the rosiest image of the company possible.”324 The most glaring omission is Meta’s failure to consider the ways in which its “surveillance advertising business model” and algorithms amplify human rights harms such as ethnic violence, hate speech, and discrimination, especially in Global South countries where its platforms are prevalent.325

The eighty-three-page report discusses the Oversight Board on only two pages (on Providing Access to Remedy). However, even in that short space, Meta both overclaims the Board’s achievements and shortchanges the contributions of other actors, extolling the OB as “defining what a human rights-based approach to content means in practice, to a level of detail, practical application and intellectual rigor, that did not previously exist.”326 Notably, Meta’s appraisal contrasts with the Board’s more measured self-assessment in its Annual Report released a month earlier.327

We have several responses to the whitewashing critique. As an initial matter, it is important to reiterate that the Oversight Board is structurally independent from Meta. It is one thing to argue that the OB should, for example, be more stringent in reviewing the company’s content moderation decisions or more forceful in its policy recommendations. It is quite another to dismiss wholesale an autonomous body that, at the very least, is nudging the platform toward greater accountability and transparency. In addition, the Board’s design is remarkably free of the kinds of structural vulnerabilities that can facilitate corporate capture.328

More substantively, the Board’s actions to date have narrowed Meta’s ability to invoke generic references to human rights to shield itself from external scrutiny. As we have shown, OB decisions and recommendations have clarified, often with granularity, how multipart standards such as ICCPR Article 19 apply to different areas of content moderation. The Board has repeatedly insisted that Meta refrain from actions that are arbitrary, opaque, or discriminatory. In addition, as previously noted, the company has complied with all OB decisions and many policy recommendations. These are significant accomplishments for an independent monitoring body that is just a few years old. They also comprise a more substantial record of achievement than most international human rights tribunals could have claimed at similarly early points in their evolution.

Concerns about business enterprises strategically using the language of corporate social responsibility or human rights to boost their brand, improve their public image, or divert attention from the harms they cause are as old as corporate philanthropy itself.329 Yet if the Board continues along its current trajectory, and if Meta continues to comply with most of its decisions and policy recommendations—both admittedly uncertain prospects—credible charges of whitewashing are likely to diminish over time.

C. Indeterminacy and Legitimacy

Another concern relates to the indeterminacy of human rights law and the legitimacy of the Oversight Board’s attempts to provide greater clarity. On the one hand, skeptics claim that indeterminacy makes it impossible for the OB to impose meaningful constraints on Meta. Yet they also fear that if the Board clears up these ambiguities, it will become “the ultimate arbiter of free speech norms around the world.”330 Such a role would, in their view, be illegitimate: the OB is neither a national court nor an international tribunal but a private oversight body comprised of unaccountable, self-appointed experts whose decisions cannot be challenged or corrected by states, civil society, or Meta users.331 This set of challenges thus asserts, somewhat paradoxically, that the Oversight Board will either have too little influence—or too much. We believe both dimensions of this critique are overstated.

Consider first indeterminacy. Skeptics argue that human rights law is too ambiguous and internally inconsistent to guide Meta’s content moderation decisions or the Board’s review of those determinations. Proponents of this view claim that there is “no accepted comprehensive set of global standards that can operate as the default rules for content moderation.”332

The texts of human rights treaties do lack precision and granularity, at least with regard to provisions—such as freedom of expression, privacy, and nondiscrimination—that are most relevant to content moderation.333 Yet this is a feature of individual rights protections in general. A modest amount of indeterminacy allows international tribunals and domestic courts to clarify and update rights in light of contemporary challenges and normative developments. It also allows decisionmakers and advocates to adopt interpretations of rights that are suitable for local contexts.

This is especially true with respect to freedom of expression. A set of global rules that purported to draw hard and fixed lines between permitted and prohibited speech and apply them across all national contexts would be both normatively problematic and practically unworkable given the diverse views regarding the benefits and harms of speech that often exist both within a single society and across countries.334 Frameworks that channel restrictions through multipart standards, such as those in ICCPR Article 19, strike a different balance: they facilitate challenges to abusive and arbitrary speech restrictions while accommodating a modicum of political and cultural variation and preferences of different polities.335

Critics of indeterminacy also overlook the fact that treaty texts are only the starting point of what international human rights law requires. Just as the scope of domestic speech protections cannot be determined by reviewing constitutions without also examining the caselaw of the courts applying them, so too treaties protecting free expression cannot be understood without consulting the interpretations of human rights tribunals and experts. A difference in the international context is that a multiplicity of institutions engages in this interpretive exercise. The relationship among these bodies is pluralistic rather than hierarchical;336 this, too, is a feature, not a bug. An extensive literature on fragmentation and harmonization explains how interactions among these institutions over time lead to greater particularity and clarity to shared legal texts, even as some divergences remain at the margins.337 We expect that OB decisions and recommendations will contribute to this interpretive enterprise, providing much-needed specificity to how human rights law applies to content moderation decisions across a range of factual contexts.

Yet skeptics also raise the opposite critique, challenging the Oversight Board’s efforts to ameliorate the indeterminacy of human rights standards. According to this view, the Board’s top-down, expert-driven attempts to balance the benefits and harms of speech in different contexts are suspect.338 At a minimum, they are not entitled to greater respect than any other reasonable approach to balancing.

In our view, the legitimacy of the Board’s normative output depends on the audience. Meta users agree to be governed by the company’s Community Standards and Values and by the OB’s review of those standards—including under international human rights law—when they create Facebook or Instagram accounts.339 They also agree that OB decisions are binding on the company. The large and growing number of appeals to the OB suggests that many users see value in challenging Meta’s content moderation decisions. To be sure, the decision to join the platform does not preclude users from objecting to the Board’s legitimacy, any more than the formal consent of states resolves all legitimacy concerns relating to international institutions.340 But it does mean that such challenges should be evaluated differently than legitimacy objections raised by other actors affected by Meta’s system of content moderation—in particular, non-users of Facebook and Instagram.

In assessing the Board’s legitimacy for external audiences, it is also important to underscore that the Board’s interpretations of international human rights law are only persuasive authority.341 Its decisions do not bind any other institution or actor. Rather, the OB’s normative influence will depend on a range of factors that scholars of the legitimacy of international tribunals have identified, such as whether it follows fair and transparent processes, whether it considers the views of a diverse range of affected actors, whether its decisions are well reasoned and persuasive, and whether it continues to maintain its independence from Meta.342

D. Negative Spillover

Although we are ultimately unpersuaded by the indeterminacy and legitimacy critiques of the Oversight Board’s human rights future, we acknowledge a related concern. The OB is a private monitoring body tasked with generating a robust body of public law norms on a relatively narrow and underdeveloped topic. It has extensive resources to generate these norms quickly and disseminate them widely, and it is actively working to increase its output.343 Although we previously championed a vision of the Board’s salutary contributions to human rights, there are also risks of negative spillover from a body with these characteristics. In particular, the OB’s decisions could generate norms that are appropriate when applied to Meta but are mischaracterized or abused by governments or other social media platforms in ways that will be difficult for international human rights tribunals to rectify.344

A useful example is the line of OB decisions holding that Meta can remove offensive expressions such as a racial slur, an image of blackface, and a rat cartoon depicting an ethnic group. In each instance, the Board recognized that Meta’s Community Standard for hate speech “prohibits specific forms of discriminatory expression” that “would raise concerns if imposed by a government at a broader level.”345 Nevertheless, the OB concluded that the company “can regulate such expression[ by] demonstrating the necessity and proportionality of the action.”346 The Board recently articulated this principle in more general terms: “Meta can apply less strict standards for removing content from its platform than those which apply to states imposing criminal or civil penalties.”347

At one level, these statements simply reflect the widely shared view that public and private actors are not similarly situated when it comes to restricting freedom of expression. For one, private actors have their own speech interests and thus “can limit, curate, and privilege content and viewpoints in a way that a state should not.”348 For another, a social media platform that removes an individual user’s post after evaluating its impact in a particular context causes substantially less harm (at least in most instances) than a statute or executive decree that punishes broad categories of disfavored speech or speakers.

Yet states and businesses could invoke the Board pronouncements quoted above in troubling ways. Consider first responses by states. At the most general level, governments may chafe at a legal standard that gives greater leeway to one of the world’s largest and richest companies. Officials might respond to this imbalance by invoking Board decisions to support a weakening of public speech restrictions.349 Such a risk is especially acute given that government regulation of internet platforms often blurs the line between public and private. Even more troublingly, a state might misconstrue the Board’s focus on criminal or civil penalties as tacitly condoning political pressure to remove disfavored speakers or content from online forums, so long as laws on the books do not threaten jail time, fines, or other legal consequences for violators.

Turning to private actors, social media companies that have little interest in fact-specific determinations about whether to remove content may cite the OB’s more deferential review of removal decisions by Meta—which is at least nominally concerned with making such fine-grained distinctions—to justify removing expression from their platforms without applying the legality, legitimacy, and proportionality standard. Firms in other industries may also invoke Board decisions as support for speech restrictions, such as limits on employee tweets or letters to the editor, on matters of public interest that relate to a company’s business. A different type of negative spillover could occur if corporations cite the Board’s pronouncements to seek a watering down of the human rights standards in the UNGPs.

None of this parade of horribles is inevitable. To reduce the risk of these outcomes, the Oversight Board needs, at a minimum, to be aware that its decisions are likely to be scrutinized and invoked by public and private actors other than Meta.350 Armed with that knowledge, the OB can provide concrete guidance as to the scope of the company’s discretion to remove content in ways that may limit the ability of other actors to mischaracterize or abuse the legal principles it develops.

Conclusion

The Meta Oversight Board will not solve the many challenges of content moderation at scale, let alone the public law challenges of private companies governing the speech of billions of people around the world. Addressing these issues requires regulation not only of downstream outputs but also of upstream choices regarding the design of institutions and technology.351 It also requires a good deal of patience. If the experience of international tribunals teaches us anything, it is that human rights accountability and norm development are incremental processes that require a plurality of approaches.

We argue that the Board has the potential to serve as an important check on Meta and to significantly advance the promotion and protection of rights online. We make this case by analogizing the OB to international human rights tribunals and exploring what the tribunals’ trajectories suggest about how the Board can hold Meta accountable and contribute to ongoing efforts to regulate social media platforms. In doing so, this Article provides the first examination of the human rights origins of the OB, the strategies it is using to pressure Meta to improve content moderation, and how the Board is developing human rights norms applicable to private social media companies. We complement this analysis with a range of recommendations about how the Board can be even more effective and have a more lasting impact.

This experiment in governing the “governors”352 has already yielded significant benefits. In just a few years, the Oversight Board has explored how human rights apply differently to corporations than governments, forced Meta to be more transparent with users in concrete ways, pushed the company to consider different types of contexts when moderating content, investigated how Meta responds to government requests to remove user posts, prompted revision of Meta’s policy on penalties, and weighed in on Meta’s cross-check program. To some observers, these may seem small steps when compared to the urgent need for greater user protections in the digital realm. We share this concern. It is for this reason that we envision the Board as only one of many public and private efforts—at the national, regional, and global levels—to push Meta and other social media platforms toward greater respect for human rights. However, when compared to the incremental and hard-fought accomplishments of international tribunals, the Board’s achievements are already impressive.

While we urge the Board to continue along this trajectory, the OB’s early successes also give us pause. A de facto human rights tribunal created by an exceedingly powerful technology company and equipped with impressive resources raises two related concerns. First, the Board is the creation of a social media giant, whose motto was at one point “move fast and break things.”353 One might argue that the international human rights system moves too slowly, but there are also risks in moving too quickly. Although speed helps the Board to keep up with technological developments, it can also lead to blind spots, missed opportunities, and inadequately reasoned decisions. In addition, given the Board’s extensive resources and the pressure it faces to demonstrate results, its efforts may quickly outstrip the normative output and monitoring efforts of established international human rights bodies.

Second, tech companies have a poor record of trying to “solve” human rights or humanitarian problems. These efforts tend to focus on quick-fix technological solutions—shiny new things that suck all the air (and funding) out of the room—rather than long-term investments in infrastructure, the rule of law, or broader social and political change.354 There is a risk that the Board will become a similar “technology of law” that crowds out other potential solutions. Some of the OB’s rhetoric suggests this possibility. The Board describes itself as “a source of best practice for self-regulation,” and it promises to “share [its] expertise about self-regulatory solutions across tech policy conversations, as well as contribute to discussions about content moderation innovation with potential tech industry partners.”355 These statements could be read as a vision for a dominant institution that overshadows and diverts resources from other efforts to hold states and private companies accountable for rights violations, both online and off, or preempts efforts to address more directly the underlying structures of our attention economy and their consequences. Avoiding these pitfalls requires the Board to remain aware of its own limitations.

Notwithstanding these cautionary words, we believe the Board is already making noteworthy contributions to advancing human rights. As it continues to provide Meta with interpretive guidance and recommendations, the OB will gradually shape the company’s content moderation decisions and policies as well as international human rights law writ large. The Board may also assist national and regional lawmakers and courts in regulating social media companies and particularizing their due diligence obligations. And Meta’s engagement with the OB may affect the company’s culture, helping to inculcate the values of transparency, information sharing, and fair process.

The Oversight Board’s most lasting impact, however, will be its contribution to the accountability efforts of other actors. The experiences of international human rights tribunals reveal that long-term change requires deep and genuine engagement with civil society organizations, advocacy groups, and other supporters. The Board’s decisions will thus have the greatest impact if they are deployed by these actors to press for meaningful change by states and companies alike.


* Harry R. Chadwick, Sr. Professor of Law, Duke University. This Article is written in the author’s personal capacity. It does not express opinions on behalf of the U.N. Human Rights Committee, on which the author currently serves as a member. This Article analyzes Oversight Board decisions and related developments through May 1, 2023. **Catherine Roraback Professor of Law and Human Rights, University of Connecticut School of Law. We are grateful for feedback and comments from Agustina del Campo, Stefania Di Stefano, evelyn douek, David Kaye, and Richard Wingfield, as well as participants in the Boston College Law School and University of Minnesota Law School faculty workshops, the European Society of International Law’s Interest Groups on International Courts and Tribunals and on Business and Human Rights, and the City University of London conference, “The Facebook/Meta Oversight Board: International and Regional Law Applications.” We are also grateful for the opportunity to present our findings to Meta’s Human Rights and Governance teams. Excellent research assistance was provided by Noah Buyon, David Del Prete, Ben Greene, Sophia Nelson, and Tatiana Varanko.