Information Integrity on Digital Platforms | Our Common Agenda Policy Brief 8
Introduction
CHAPEAU
The challenges that we are facing can be addressed only through stronger international co-operation. The Summit of the Future, to be held in 2024, is an opportunity to agree on multilateral solutions for a better tomorrow, strengthening global governance for both present and future generations (General Assembly resolution 76/307). In my capacity as Secretary-General, I have been invited to provide inputs to the preparations for the Summit in the form of action-oriented recommendations, building on the proposals contained in my report entitled “Our Common Agenda” (A/75/982), which was itself a response to the declaration on the commemoration of the seventy-fifth anniversary of the United Nations (General Assembly resolution 75/1). The present policy brief is one such input. It serves to elaborate on the ideas first proposed in Our Common Agenda, taking into account subsequent guidance from Member States and more than one year of intergovernmental and multi-stakeholder consultations, and rooted in the purposes and the principles of the Charter of the United Nations, the Universal Declaration of Human Rights and other international instruments.
PURPOSE OF THIS POLICY BRIEF
The present policy brief is focused on how threats to information integrity are having an impact on progress on global, national and local issues. In Our Common Agenda, I called for empirically backed consensus around facts, science and knowledge. To that end, the present brief outlines potential principles for a code of conduct that will help to guide Member States, the digital platforms and other stakeholders in their efforts to make the digital space more inclusive and safe for all, while vigorously defending the right to freedom of opinion and expression, and the right to access information. The Code of Conduct for Information Integrity on Digital Platforms is being developed in the context of preparations for the Summit of the Future. My hope is that it will provide a gold standard for guiding action to strengthen information integrity.
Digital platforms are crucial tools that have transformed social, cultural and political interactions everywhere. Across the world, they connect concerned global citizens on issues that matter. Platforms help the United Nations to inform and engage people directly as we strive for peace, dignity and equality on a healthy planet. They have given people hope in times of crisis and struggle, amplified voices that were previously unheard and breathed life into global movements.
Yet these same platforms have also exposed a darker side of the digital ecosystem. They have enabled the rapid spread of lies and hate, causing real harm on a global scale. Optimism over the potential of social media to connect and engage people has been dampened as mis- and disinformation and hate speech have surged from the margins of digital space into the mainstream.
The danger cannot be overstated. Social media-enabled hate speech and disinformation can lead to violence and death.1A/HRC/42/50; A/77/287; A/HRC/51/53; United Nations, “Statement by Alice Wairimu Nderitu, Special Adviser on the Prevention of Genocide, condemning the recent escalation of fighting in Ethiopia”, press release, 19 October 2022; Office of the United Nations High Commissioner for Human Rights (OHCHR), “Myanmar: Social media companies must stand up to junta’s online terror campaign say UN experts”, press release, 13 March 2023; OHCHR, “Freedom of speech is not freedom to spread racial hatred on social media: UN experts”, statement, 6 January 2023; Special Rapporteur on the promotion and protection of the right to freedom of opinion and expression, “#JournalistsToo: women journalists speak out”, 24 November 2021; and OHCHR, “Sri Lanka: Experts dismayed by regressive steps, call for renewed UN scrutiny and efforts to ensure accountability”, press release, 5 February 2021.
The ability to disseminate large-scale disinformation to undermine scientifically established facts poses an existential risk to humanity (A/75/982, para. 26) and endangers democratic institutions and fundamental human rights. These risks have further intensified because of rapid advancements in technology, such as generative artificial intelligence. Across the world, the United Nations is monitoring how mis- and disinformation and hate speech can threaten progress towards the Sustainable Development Goals. It has become clear that business as usual is not an option.
What is information integrity?
Information integrity refers to the accuracy, consistency and reliability of information. It is threatened by disinformation, misinformation and hate speech. While there are no universally accepted definitions of these terms, United Nations entities have developed working definitions.
The Special Rapporteur on the promotion and protection of the right to freedom of opinion and expression refers to disinformation as “false information that is disseminated intentionally to cause serious social harm”.2A/HRC/47/25, para. 15. Disinformation is described by the United Nations Educational, Scientific and Cultural Organization (UNESCO) as false or misleading content that can cause specific harm, irrespective of motivations, awareness or behaviours.3Kalina Bontcheva and Julie Posetti, eds., Balancing Act: Countering Digital Disinformation While Respecting Freedom of Expression – Broadband Commission Research Report on “Freedom of Expression and Addressing Disinformation on the Internet” (Geneva, International Telecommunication Union (ITU); Paris, UNESCO, 2020).
Information integrity and digital platforms
Digital platforms should be integral players in the drive to uphold information integrity. While certain traditional media can also be sources of mis- and disinformation, the velocity, volume and virality of their spread via digital channels warrants an urgent and tailored response. For the purposes of the present brief, the term “digital platform” refers to a digital service that facilitates interactions between two or more users, covering a wide range of activities, from social media and search engines to messaging apps. Typically, they collect data about their users and their interactions.7The European Commission defines online platforms at “Shaping Europe’s digital future: online platforms”, 7 June 2022. Available at https://digital-strategy.ec.europa.eu/en/policies/online-platforms.
Mis- and disinformation are created by a wide range of actors, with various motivations, who by and large are able to remain anonymous. Coordinated disinformation campaigns by State and non-State actors have exploited flawed digital systems to promote harmful narratives, with serious repercussions.
Many States have launched initiatives to regulate digital platforms, with at least 70 such laws adopted or considered in the last four years.8See OHCHR, “Moderating online content: fighting harm or silencing dissent?”, 23 July 2021. At their core, legislative approaches typically involve a narrow scope of remedies to define and remove harmful content. By focusing on the removal of harmful content, some States have introduced flawed and overbroad legislation that has in effect silenced “protected speech”, which is permitted under international law. Other responses, such as blanket Internet shutdowns and bans on platforms, may lack legal basis and infringe human rights.
Many States and political figures have used alleged concerns over information integrity as a pretext to restrict access to information, to discredit and restrict reporting, and to target journalists and opponents.9See United Nations, “Countering disinformation”, and A/77/287. State actors have also pressured platforms to do their bidding under the guise of tackling mis- and disinformation.10A/HRC/47/25. Freedom of expression experts have stressed that State actors have a particular duty in this context and “should not make, sponsor, encourage or further false information” (A/77/287, para. 45).
The risks inherent in the regulation of expression require a carefully tailored approach that complies with the requirements of legality, necessity and proportionality under human rights law, even when there is a legitimate public interest purpose (ibid., para. 42).
Disinformation is also big business. Both “dark” and mainstream public relations firms, contracted by States, political figures and the private sector, are key sources of false and misleading content.11Stephanie Kirchgaessner and others, “Revealed: the hacking and disinformation team meddling in elections”, The Guardian, 14 February 2023. One tactic, among others, has been to publish content to fake cloned versions of news sites to make articles seem like they are from legitimate sources.12Alexandre Alaphilippe and others, “Doppelganger – media clones serving Russian propaganda”, EU DisinfoLab, 27 September 2022. This shadowy business is extremely difficult to track and research so that the true scale of the problem is unknown. Individuals, too, spread false claims to peddle products or services for profit, often targeting vulnerable groups during times of crisis or insecurity.
A dominant approach in the current business models of most digital platforms hinges on the “attention economy”. Algorithms are designed to prioritize content that keeps users’ attention, thereby maximizing engagement and advertising revenue. Inaccurate and hateful content designed to polarize users and generate strong emotions is often that which generates the most engagement, with the result that algorithms have been known to reward and amplify mis- and disinformation and hate speech.13United Nations Economist Network, “New economics for sustainable development: attention economy”.
Facing a decline in advertising revenue, digital platforms are seeking alternative avenues for profit beyond the attention economy. For example, paid verification plans, whereby accounts can buy a seal of approval previously used to denote authenticity, have raised serious concerns for information integrity given the potential for abuse by disinformation actors.14Twitter, “About Twitter Blue”; and Meta, “Testing Meta Verified to help creators establish their presence”, 17 March 2023.
What is the relevant international legal framework?
The promotion of information integrity must be fully grounded in the pertinent international norms and standards, including human rights law and the principles of sovereignty and non-intervention in domestic affairs. In August 2022, I transmitted to the General Assembly a report entitled “Countering disinformation for the promotion and protection of human rights and fundamental freedoms”.15A/77/287.
In the report, I laid out the international human rights law that applies to dis- information, including the Universal Declaration of Human Rights and the International Covenant on Civil and Political Rights. Under these international legal instruments, everyone has the right to freedom of expression.16As at February 2023, 173 Member States were States parties to the International Covenant on Civil and Political Rights.
Article 19 of the Universal Declaration of Human Rights and article 19 (2) of the Covenant protect the right to freedom of expression, including the freedom to seek, receive and impart information and ideas of all kinds, regardless of frontiers, and through any media. The human right to freedom of expression is not limited to favourably received information (A/77/287, para. 13). Linked to freedom of expression, freedom of information is itself a right.
The General Assembly has stated: “Freedom of information is a fundamental human right and is the touchstone of all the freedoms to which the United Nations is consecrated” (ibid., para. 14). Freedom of expression and access to information may be subject to certain restrictions that meet specific criteria laid out in article 19 (3) of the Covenant.17Limitations on freedom of expression must meet the following well-established conditions: legality, that is, restrictions must be provided by law in a manner that distinguishes between lawful and unlawful expression with sufficient precision; necessity and proportionality, that is, the limitation demonstrably imposes the least burden on the exercise of the right and actually protects, or is likely to protect, the legitimate State interest at issue; and legitimacy, that is, to be lawful, restrictions must protect only those interests enumerated in article 19 (3) of the International Covenant on Civil and Political Rights States cannot add additional grounds or restrict expression beyond what is permissible under international law.
The Rabat Plan of Action on the prohibition of advocacy of national, racial or religious hatred that constitutes incitement to hostility, discrimination or violence, adopted in 2012, provides practical legal and policy guidance to States on how best to implement article 20 (2) of the Covenant and article 4 of the International Convention on the Elimination of All Forms of Racial Discrimination, which prohibit certain forms of hate speech. The Rabat Plan of Action has already been utilized by Member States in different contexts.18These include audiovisual communications in Côte d’Ivoire, Morocco and Tunisia, and monitoring of incitement to violence by the United Nations Multidimensional Integrated Stabilization Mission in the Central African Republic.
Hate speech has been a precursor to atrocity crimes, including genocide. The 1948 Convention on the Prevention and Punishment of the Crime of Genocide prohibits “direct and public incitement to commit genocide”.
In its resolution 76/227, adopted in 2021, the General Assembly emphasized that all forms of disinformation can negatively impact the enjoyment of human rights and fundamental freedoms, as well as the attainment of the Sustainable Development Goals. Similarly, in its resolution 49/21, adopted in 2022, the Human Rights Council affirmed that disinformation can negatively affect the enjoyment and realization of all human rights.
What harm is being caused by online mis- and disinformation and hate speech?
Online mis- and disinformation and hate speech pose serious concerns for the global public. One study of survey data from respondents in 142 countries found that 58.5 per cent of regular Internet and social media users worldwide are concerned about encountering misinformation online, with young people and people in low-income tiers feeling significantly more vulnerable.19Aleksi Knuutila, Lisa-Maria Neudert and Philip N. Howard, “Who is afraid of fake news? Modeling risk perceptions of misinformation in 142 countries”, Harvard Kennedy School (HKS) Misinformation Review, vol. 3, No. 3 (April 2022). Today’s youth are digital natives who are more likely to be connected online than the rest of the population, making them the most digitally connected generation in history.20ITU, Measuring the Information Society (Geneva, 2013). Around the world, a child goes online for the first time every half second, putting them at risk of exposure to online hate speech and harm, in some cases affecting their mental health.21United Nations Children’s Fund (UNICEF), “Protecting children online”, 23 June 2022. Available at www.unicef.org/protection/ violence-against-children-online.
The impacts of online mis- and disinformation and hate speech can be seen across the world, including in the areas of health, climate action, democracy and elections, gender equality, security and humanitarian response. Information pollution was identified as a significant concern by 75 per cent of United Nations Development Programme country offices in a 2021 survey. It has severe implications for trust, safety, democracy and sustainable development, as found in a recent UNESCO-commissioned review of more than 800 academic, civil society, journalistic and corporate documents.22UNESCO, working papers on digital governance and the challenges for trust and safety. Available at www.unesco.org/en/ internet-conference/working-papers.
Mis- and disinformation can be dangerous and potentially deadly, especially in times of crisis, emergency or conflict. During the coronavirus disease (COVID-19) pandemic, a deluge of mis- and disinformation about the virus, public health measures and vaccines began to circulate online.23See Julie Posetti and Kalina Bontcheva, “Disinfodemic: deciphering COVID-19 disinformation” policy brief 1, (Paris, UNESCO, 2020), and “Disinfodemic: dissecting responses to COVID-19 disinformation” policy brief 2, (Paris, UNESCO, 2020). Certain actors exploited the confusion for their own objectives, with anti-vaccine campaigners driving users to sites selling fake cures or preventive measures.24Center for Countering Digital Hate, Pandemic Profiteers: The Business of Anti-Vaxx (2021). Many victims of COVID-19 refused to get vaccinated or take basic health precautions after being exposed to mis- and disinformation online.25Michael A Gisondi and others, “A deadly infodemic: social media and the power of COVID-19 misinformation”, Journal of Medical Internet Research, vol. 24, No. 2 (February 2022).
Disinformation can likewise prove deadly in already volatile societal and political contexts. In a 2022 report, the Special Rapporteur on the promotion and protection of the right to freedom of opinion and expression examined the impact of weaponized information in sowing confusion, feeding hate, inciting violence and prolonging conflict.26A/77/288. In another report issued in 2022, it was found that disinformation can “involve bigotry and hate speech aimed at minorities, women and any so-called ‘others’, posing threats not only to those directly targeted, but also to inclusion and social cohesion. It can amplify tensions and divisions in times of emergency, crisis, key political moments or armed conflict”.27A/77/287, para. 6.
Some of the worst impacts of online harms are in contexts neglected by digital platforms, even where platforms enjoy high penetration rates. Countries in the midst of conflict, or with otherwise volatile contexts that are often less lucrative markets, have not been allocated sufficient resources for content moderation or user assistance. While traditional media remain an important source of news for most people in conflict areas, hatred spread on digital platforms has also sparked and fuelled violence28.In 2018, an independent international fact-finding mission appointed by the Human Rights Council declared Facebook to be “the leading platform for hate speech in Myanmar” (A/HRC/42/50, para. 72). Some digital platforms have faced criticism of their role in conflicts, including the ongoing war in Ukraine.29See United Nations News, “Hate speech: a growing, international threat”, 28 January 2023, and “Digital technology, social media fuelling hate speech like never before, warns UN expert”, 20 October 2022
Similarly, mis- and disinformation about the climate emergency are delaying urgently needed action to ensure a liveable future for the planet. Climate mis- and disinformation can be understood as false or misleading content that undercuts the scientifically agreed basis for the existence of human-induced climate change, its causes and impacts. Coordinated campaign are seeking to deny, minimize or distract from the Intergovernmental Panel on Climate Change scientific consensus and derail urgent action to meet the goals of the 2015 Paris Agreement. A small but vocal minority of climate science denialists30See John Cook, “Deconstructing climate science denial”, in Research Handbook in Communicating Climate Change, David C. Holmes and Lucy M. Richardson, eds. (Cheltenham, United Kingdom, Edward Elgar, 2020). Cook reported that Abraham et al. (2014) had summarized how papers containing denialist claims, such as claims of cooling in satellite measurements or estimates of low climate sensitivity, have been robustly refuted in the scientific literature. Similarly, Benestad et al. (2016) attempted to replicate findings in contrarian papers and found a number of flaws such as inappropriate statistical methods, false dichotomies, and conclusions based on misconceived physics continue to reject the consensus position and command an outsized presence on some digital platforms. For example, in 2022, random simulations by civil society organizations revealed that Facebook’s algorithm was recommending climate denialist content at the expense of climate science.31Global Witness, “The climate divide: how Facebook’s algorithm amplifies climate disinformation”, 28 March 2022. On Twitter, uses of the hashtag #climatescam shot up from fewer than 2,700 a month in the first half of 2022 to 80,000 in July and 199,000 in January 2023. The phrase was also featured by the platform among the top results in the search for “climate”.32Analysis by the Department of Global Communications, using data from Talkwalker. In February 2022, the Intergovernmental Panel on Climate Change called out climate disinformation for the first time, stating that a “deliberate undermining of science” was contributing to “misperceptions of the scientific consensus, uncertainty, disregarded risk and urgency, and dissent”.33Jeffrey A. Hicke and others, “North America”, in Intergovernmental Panel on Climate Change, Climate Change 2022: Impacts, Adaptation and Vulnerability, Contribution of Working Group II to the Sixth Assessment Report of the Intergovernmental Panel on Climate Change (Cambridge, United Kingdom, Cambridge University Press, 2022).
FIGURE I
MONTHLY NUMBER OF USES OF #CLIMATESCAM ON TWITTER
Some fossil fuel companies commonly deploy a strategy of “greenwashing”, misleading the public into believing that a company or entity is doing more to protect the environment, and less to harm it, than it is. The companies are not acting alone. Efforts to confuse the public and divert attention away from the responsibility of the fossil fuel industry are enabled and supported by advertising and public relations providers, advertising tech companies, news outlets and digital platforms.34Mei Li, Gregory Trencher and Jusen Asuka, “The clean energy claims of BP, Chevron, ExxonMobil and Shell: a mismatch between dis- course, actions and investments”, PLOS ONE, issue 17, No. 2 (February 2022).
Advertising and public relations firms that create greenwashing content and third parties that distribute it are collectively earning billions from these efforts to shield the fossil fuel industry from scrutiny and accountability. Public relations firms have run hundreds of campaigns for coal, oil and gas companies.35Robert J. Brulle and Carter Werthman, “The role of public relations firms in climate change politics”, Climatic Change, vol. 169, No. 1–2 (November 2021). According to the Global Disinformation Index, a non-profit watchdog, tech industry advertisers provided $36.7 million to 98 websites carrying climate disinformation in English in 2021. A November 2022 report from the Center for Countering Digital Hate, a campaign group, revealed that, on Google alone, nearly half of the $23.7 million spent on search ads by oil and gas companies in the last two years targeted search terms on environmental sustainability. Research from InfluenceMap found 25,147 misleading ads from 25 oil and gas sector organizations on Facebook’s platforms in the United States of America in 2020, with a total spend of $9,597,376. So far the response has been incommensurate with the scale of the problem.
Mis- and disinformation are having a profound impact on democracy, weakening trust in democratic institutions and independent media, and dampening participation in political and public affairs. Throughout the electoral cycle, exposure to false and misleading information can rob voters of the chance to make informed choices. The spread of mis- and disinformation can undermine public trust in electoral institutions and the electoral process itself – such as voter registration, polling and results – and potentially result in voter apathy or rejection of credible election results. States and political leaders have proved to be potent sources of disinformation, deliberately and strategically spreading falsehoods to maintain or secure power, or undermine democratic processes in other countries.36See General Assembly resolution 76/227; Human Rights Council resolution 49/21; and European Union External Action, “Tackling disin- formation, foreign information manipulation and interference”, 27 October 2021.
Marginalized and vulnerable groups are also frequent targets of mis- and disinformation and hate speech, resulting in their further social, economic and political exclusion. Women candidates, voters, electoral officials, journalists and civil society representatives are targeted with gendered disinformation online.37Lucina Di Meco, “Monetizing misogyny: gendered disinformation and the undermining of women’s rights and democracy globally”, #ShePersisted, February 2023. These attacks undermine political participation and weaken democratic institutions and human rights, including the freedom of expression and access to information of these groups.38See Andrew Puddephatt, “Social media and elections”, Cuadernos de Discusión de Comunicación e Información, No. 14 (Montevideo, UNESCO, 2019); and Julie Posetti and others, “The chilling: global trends in online violence against women journalists”, research discussion paper (UNESCO, 2021). This must be an increasingly urgent priority for the international community, not least because more than 2 billion voters are set to go to the polls around the world in 2024.
Mis- and disinformation also cross-pollinate between and within platforms and traditional media, becoming even more complex to track and address if not detected at the source. Disinformation can be a deliberate tactic of ideologically influenced media outlets co-opted by political and corporate interests.39EU Disinfo Lab, “The role of “media” in producing and spreading disinformation campaigns”, 13 October 2021. At the same time, the rise of digital platforms has precipitated a dramatic decline in trustworthy, independent media. News audiences and advertising revenues have migrated en masse to Internet platforms – a trend exacerbated by the COVID-19 pandemic. “Media extinction” or “news deserts”, where communities lose trustworthy local news sources, can be seen in some regions or countries,40See United Nations News, “Social media poses ‘existential threat’ to traditional, trustworthy news: UNESCO”, 10 March 2022; and Anya Schiffrin and others, “Finding the funds for journalism to thrive: policy options to support media viability”, World Trends in Freedom of Expression and Media Development (Paris, UNESCO, 2022). contributing to pollution of the information ecosystem. “Newswashing” – whereby sponsored content is dressed up to look like reported news stories – is often inadequately sign-posted when posted to digital platforms, lending it a veneer of legitimacy. Once picked up by other media, cited by politicians, or shared widely across platforms, the origin of the information becomes increasingly murky, and news consumers are left unable to distinguish it from objective fact.
FIGURE II
INFORMATION INTEGRITY AND THE SUSTAINABLE DEVELOPMENT GOALS
As shown, threats to information integrity can have a negative impact on the achievement of the Sustainable Development Goals.
- World Health Organization and United Nations Children’s Fund (UNICEF), Progress on Household Drinking Water, Sanitation and Hygiene 2000–2020: Five Years into the SDGs (Geneva, 2021).
- See Roberto Cavazos and CHEQ, “The economic cost of bad actors on the Internet: fake news, 2019”; and London Economics, “The cost of lies: assessing the human and financial impact of COVID-19 related online misinformation on the UK”, December 2020.
- Global Witness, Last Line of Defence: The Industries Causing the Climate Crisis and Attacks against Land and Environmental Defenders (2021)
- A/77/288.
How can we strengthen information integrity?
Mis- and disinformation and hate speech do not exist in a vacuum. They spread when people feel excluded and unheard, when faced with the impacts of economic disparity, and when feeling politically disenchanted. Responses should address these real-world challenges. Efforts to achieve the Sustainable Development Goals are fundamental to building a world in which trust can be restored.
In crafting responses, it is important not to lose sight of the tremendous value digital platforms bring to the world. Platforms have revolutionized real-time mass communication, enabling the spread of life-saving information during natural disasters and pandemics. They have helped to mobilize support for the goals for which the United Nations strives, often proving to be positive forces for inclusion and participation in public life. They have connected geographically disparate communities of people otherwise excluded, including those suffering from rare health conditions, and a vast array of activists working to make the world a better place.
REGULATORY RESPONSES
The question of whether digital platforms can and should be held legally liable for the content that they host has been the subject of lengthy debate. In some contexts, existing laws based on defamation, cyberbullying and harassment have been used effectively to counter threats to information integrity without imposing new restrictions on freedom of expression (A/77/287, para. 44).
In addition, some recent legislative efforts have been made to address the issue at the regional and national levels. These include the framework adopted by the European Union in 2022, comprising the Digital Services Act, the initiative on transparency and targeting of political advertising and the Code of Practice on Disinformation. The Digital Services Act establishes new rules for users, digital platforms and businesses operating online within the European Union. Measures take aim at illegal online content, goods and services and provide a mechanism for users both to flag illegal content and to challenge moderation decisions that go against them. They require digital platforms to improve transparency, especially on the use and nature of recommendation algorithms, and for larger platforms to provide researchers with access to data.
The Code of Practice on Disinformation sets out principles and commitments for online platforms and the advertising sector to counter the spread of disinformation online in the European Union, which its signatories agreed to implement.41European Commission, “Shaping Europe’s digital future: the 2022 Code of Practice on Disinformation”, 4 July 2022. These include voluntary commitments to help demonetize disinformation, both by preventing the dissemination of advertisements containing disinformation and by avoiding the placement of advertisements alongside content containing disinformation. Signatories also agreed to label political advertising more clearly, together with details of the sponsor, advertising spend and display period, and to create searchable databases of political advertisements. Furthermore, they committed to share information about malicious manipulative behaviours used to spread disinformation (such as fake accounts, bot-driven amplification, impersonation and malicious deep-fakes) detected on their platforms and regularly update and implement policies to tackle them. Other commitments are focused on empowering users to recognize, understand and flag disinformation, and on strengthening collaboration with fact-checkers and providing researchers with better access to data. The real test of these new mechanisms will be in their implementation.
One of the key aims of the Code of Practice on Disinformation is to improve platform transparency. In February 2023, signatories to the Code of Practice published their first baseline reports on how they are implementing the commitments. The reports provided insight into the extent to which advertising revenue was prevented from flowing to disinformation actors and other detected manipulative behaviour, including a large-scale coordinated effort to manipulate public opinion about the war in Ukraine in several European countries.42See the remarks by the European Commission Vice-President for Values and Transparency, Věra Jourová, in European Commission, “Code of Practice on Disinformation: new Transparency Centre provides insights and data on online disinformation for the first time”, daily news, 9 February 2023. Available at https://ec.europa.eu/commission/presscorner/detail/en/mex_23_723.
DIGITAL PLATFORM RESPONSES
Digital platforms are very diverse in terms of size, function and structure and have pursued a wide range of responses to tackle harm. Several of the larger platforms have publicly committed to uphold the Guiding Principles on Business and Human Rights,43Available at https://unglobalcompact.org/library/2 but gaps persist in policy, transparency and implementation. Some platforms do not enforce their own standards and, to varying degrees, allow and amplify lies and hate.44Center for Countering Digital Hate and Human Rights Campaign, “Digital hate: social media’s role in amplifying dangerous lies about LGBTQ+ people”, 10 August 2022. Algorithms created to further the platforms’ profit-making model are designed to deliberately maximize engagement and monopolize attention, tending to push users towards polarizing or provocative content.
Most digital platforms have some kind of system of self-regulation, moderation or oversight mechanisms in place, yet transparency around content removal policy and practice remains a challenge.45See Andrew Puddephatt, “Letting the sun shine in: transparency and accountability in the digital age”, World Trends in Freedom of Expression and Media Development (Paris, UNESCO, 2021). Investment in these mechanisms across regions and languages is extremely patchy and largely concentrated in the global North, as is platforms’ enforcement of their own rules. Translation of moderation tools and oversight mechanisms into local languages is incomplete across platforms, a recent survey found46.Whose Knowledge?, Oxford Internet Institute and The Centre for Internet and Society, State of the Internet’s Languages Report (2022). At the same time, moderation is often outsourced and woefully underresourced in languages other than English.47A/HRC/38/35. Testimony from moderators has raised troubling questions related to mistreatment, labour standards and secondary trauma.48Billy Perrigo, “Inside Facebook’s African sweatshop”, Time, 17 February 2022. Moderators report being constantly exposed to violent and disturbing content, and being given a matter of seconds to determine if a reported post violates company policy. Automated content moderation systems can play an essential role, but are exposed to possible bias based on the data and structures used to train them. They also have high rates of error in English and even worse success rates across other languages. A number of digital platforms employ trust and safety, human rights and information integrity teams, yet these experts are often not included at the earliest stages of product development and are often the first jobs cut during cost-saving measures.
DATA ACCESS
Data access for researchers is also an urgent priority on a global scale. Existing research and resources remain heavily skewed towards the United States of America and Europe. With notable exceptions, including the reports on United Nations peacekeeping in Africa and the independent international fact-finding mission on Myanmar,49A/HRC/42/50. and some investigative reporting and coverage by journalists,50Notable examples include Maria Ressa, How to Stand up to a Dictator (New York, HarperCollins, 2022); and Max Fischer, The Chaos Machine (New York, Little, Brown and Company, 2022). limited research has been published about the impact on the rest of the world. This is partly because researchers lack access to platforms and their data. The tools needed for effective research of the limited data provided by the platforms also tend to be designed with marketing in mind and are largely prohibitively expensive. A shift by the platforms from an “access by request” approach to “disclosure by default”, with necessary safeguards for privacy, would allow researchers to properly evaluate harms.
USER EMPOWERMENT
Civil society groups and academics have conducted extensive research on how best to tackle mis- and disinformation and hate speech while protecting freedom of expression. A number have stressed the need for bottom-up solutions that empower Internet users to limit the impact of online harms on their own communities and decentralize power from the hands of the platforms.
Platform users, including marginalized groups, should be encouraged, included and involved in the policy space. Youth in particular have a wealth and depth of expertise. As digital natives, young people, in particular young women, and children are already often the targets of mis- and disinformation and hate speech and will be directly affected by emerging and new platforms. Younger users can speak from experience about the differentiated impact of various proposals and their potential flaws. They have also actively contributed to online advocacy and fact-checking efforts.51See UNICEF, “Young reporters fact-checking COVID-19 information”.
Improved critical thinking skills can make users more resilient against digital manipulation. Specifically, digital literacy teaches users to better evaluate the information that they encounter online and pass it on in a responsible way. A range of United Nations entities have had valuable experiences in this field. The United Nations Verified initiative52See https://shareverified.com/. has successfully deployed a range of tactics, including targeting messaging for users, pre-bunking – warning users about falsehoods before they encounter them – and digital literacy drives.
DISINCENTIVES
The current business models of most digital platforms prioritize engagement above human rights, privacy and safety. This includes the monetization of personal data for profit, despite growing evidence of societal harm caused by this business model.
FIGURE III
UNITED NATIONS CAMPAIGNS EFFECTIVE IN COUNTERING MIS- AND DISINFORMATION
Some civil society groups and researchers have explored avenues for demonetizing and therefore disincentivizing the creation and spread of online mis- and disinformation and hate speech, noting that while freedom of expression is a fundamental human right, profiting from it is not.53The Global Disinformation Index, a non-profit group, tracks advertising placed together with disinformation. The United Nations has been a victim of this practice, with the Global Disinformation Index having found UNICEF advertisements placed alongside anti-vaccine articles, and Office of the United Nations High Commissioner for Refugees advertisements together with anti-refugee content. Proposals seek to address the profitability of disinformation, ensure full transparency around monetization of content and independent risk assessments, and disincentivize those involved in online advertising from enabling disinformation.
Brands that advertise alongside mis- and disinformation and hate speech risk undermining the effectiveness of their campaigns and, ultimately, their reputations. Advertisers can develop clear policies to avoid inadvertently funding and legitimizing mis- and disinformation and hate speech and help make them unprofitable. Implementation measures can include managing up-to-date inclusion and exclusion lists and using advertisement verification tools. Advertisers can also pressure digital platforms to step up action to protect information integrity and can refrain from advertising with media outlets that fuel hatred and spread disinformation.54Conscious Advertising Network, manifestos. Available at www.consciousadnetwork.com/the-manifestos/
INDEPENDENT MEDIA
New measures in dozens of countries continue to undermine press freedom. According to the UNESCO global report for 2022 from its flagship World Trends in Freedom of Expression and Media Development series, 85 per cent of the world’s population experienced a decline in press freedom in their country during the preceding five years.55UNESCO, Journalism is a Public Good: World Trends in Freedom of Expression and Media Development – Global Report 2021/2022 (Paris, 2022). With 2.7 billion people still offline,56ITU, “Facts and figures 2021: 2.9 billion people still offline”, 29 November 2021. The global digital compact to be taken up by Member States at the Summit of the Future, to be held in 2024, will outline shared principles for an open, free and secure digital future for all (see www.un.org/techenvoy/global-digital-compact). a further priority is to strengthen independent media, boost the prevalence of fact-checking initiatives and underpin reliable and accurate reporting in the public interest. Real public debate relies on the facts, told clearly, and reported ethically and independently. Ethical reporters, with quality training and working conditions, have the skills to restore balance in the face of mis- and disinformation. They can offer a vital service: accurate, objective and reliable information about the issues that matter.
FIGURE IV
MONTHS TAKEN TO ADD 100 MILLION MONTHLY ACTIVE USERS FOR CHATGPT AS COMPARED WITH OTHER POPULAR APPS
FUTURE-PROOFING
Even as we seek solutions to protect information integrity in the current landscape, we must ensure that recommendations are future-proof, addressing emerging technologies and those yet to come. Launched in November 2022, Open AI’s ChatGPT-3 platform gained 100 million users by January 2023, making it the fastest-growing consumer app in history,57Krystal Hu, “ChatGPT sets record for fastest-growing user base – analyst note”, Reuters, 2 February 2023. with many other companies racing to develop competing tools.
While holding almost unimaginable potential to address global challenges, there are serious and urgent concerns about the equally powerful potential of recent advances in artificial intelligence – including image generators and video deepfakes – to threaten information integrity. Recent reporting and research have shown that generative artificial intelligence tools generated mis- and disinformation and hate speech, convincingly presented to users as fact.58See Center for Countering Digital Hate, “Misinformation on Bard, Google’s new AI chat”, 5 April 2023; and Tiffany Hsu and Stuart A. Thompson, “Disinformation researchers raise alarms about A.I. chatbots”, The New York Times, 13 February 2023.
My Envoy on Technology is leading efforts to assess the implications of generative artificial intelligence and other emerging platforms. In doing so we must learn from the mistakes of the past. Digital platforms were launched into the world without sufficient awareness or assessment of the potential damage to societies and individuals. We have the opportunity now to ensure that history does not repeat itself with emerging technology. The era of Silicon Valley’s “move fast and break things” philosophy must be brought to a close. It is essential that user privacy, security, transparency and safety by design are integrated into all new technologies and products at the outset.
UNITED NATIONS RESPONSES
Steps are also being taken, including by United Nations peace operations and country offices, to monitor, analyse and respond to the threat that mis- and disinformation pose to the delivery of United Nations mandates. The United Nations Strategy and Plan of Action on Hate Speech sets out strategic guidance for the Organization to address hate speech at the national and global levels. In February 2023, UNESCO hosted the Internet for Trust conference to discuss a set of draft global guidelines for regulating digital platforms, due to be finalized later this year.59The draft guidelines are available at www.unesco.org/en/internet-conference
Together, these initiatives and approaches help to point the way forward for the underlying principles of a United Nations Code of Conduct.
Towards a United Nations Code of Conduct
The United Nations Code of Conduct for Information Integrity on Digital Platforms, which I will put for- ward, would build upon the following principles:
- Commitment to information integrity
- Respect for human rights
- Support for independent media
- Increased transparency
- User empowerment
- Strengthened research and data access
- Scaled up responses
- Stronger disincentives
- Enhanced trust and safety
These principles have been distilled from the core ideas discussed in the present policy brief, and are in line and interlinked with my policy brief on a Global Digital Compact. Member States will be invited to implement the Code of Conduct at the national level. Consultations will continue with stakeholders to further refine the content of the Code of Conduct, as well as to identify concrete methodologies to operationalize its principles.
The Code of Conduct may draw upon the following recommendations:
Commitment to information integrity
- All stakeholders should refrain from using, supporting or amplifying disinformation and hate speech for any purpose, including to pursue political, military or other strategic goals, incite violence, undermine democratic processes or target civilian populations, vulnerable groups, communities or individuals;
Respect for human rights
- Member States should:
- Ensure that responses to mis- and disinformation and hate speech are consistent with international law, including international human rights law, and are not misused to block any legitimate expression of views or opinion, including through blanket Internet shutdowns or bans on platforms or media outlets;
- Undertake regulatory measures to protect the fundamental rights of users of digital platforms, including enforcement mechanisms, with full transparency as to the requirements placed on technology companies;
- All stakeholders should comply with the Guiding Principles on Business and Human Rights;
Support for independent media
- Member States should guarantee a free, viable, independent and plural media landscape with strong protections for journalists and independent media, and support the establishment, funding and training of independent fact-checking organizations in local languages;
- News media should ensure accurate and ethical independent reporting supported by quality training and adequate working conditions in line with international labour and human rights norms and standards;
Increased transparency
- Digital platforms should:
- Ensure meaningful transparency regarding algorithms, data, content moderation and advertising;
- Publish and publicize accessible policies on mis- and disinformation and hate speech, and report on the prevalence of coordinated disinformation on their services and the efficacy of policies to counter such operations;
- News media should ensure meaningful transparency of funding sources and advertising policies, and clearly distinguish editorial content from paid advertising, including when publishing to digital platforms;
User empowerment
- Member States should ensure public access to accurate, transparent, and credibly sourced government information, particularly information that serves the public interest, including all aspects of the Sustainable Development Goals;
- Digital platforms should ensure transparent user empowerment and protection, giving people greater choice over the content that they see and how their data is used. They should enable users to prove identity and authenticity free of monetary or privacy trade-offs and establish transparent user complaint and reporting processes supported by independent, well publicized and accessible complaint review mechanisms;
- All stakeholders should invest in robust digital literacy drives to empower users of all ages to better understand how digital platforms work, how their personal data might be used, and to identify and respond to mis- and disinformation and hate speech. Particular attention should be given to ensuring that young people, adolescents and children are fully aware of their rights in online spaces;
Strengthened research and data access
- Member States should invest in and support independent research on the prevalence and impact of mis- and disinformation and hate speech across countries and languages, particularly in underserved contexts and in languages other than English, allowing civil society and academia to operate freely and safely;
- Digital platforms should:
- Allow researchers and academics access to data, while respecting user privacy. Researchers should be enabled to gather examples and qualitative data on individuals and groups targeted by mis- and disinformation and hate speech to better understand the scope and nature of harms, while respecting data protection and human rights;
- Ensure the full participation of civil society in efforts to address mis- and disinformation and hate speech;
Scaled-up responses
- All stakeholders should:
- Allocate resources to address and report on the origins, spread and impact of mis- and disinformation and hate speech, while respecting human rights norms and standards and further invest in fact-checking capabilities across countries and contexts;
- Form broad coalitions on information integrity, bringing together different expertise and approaches to help to bridge the gap between local organizations and technology companies operating at a global scale;
- Promote training and capacity-building to develop understanding of how mis- and disinformation and hate speech manifest and to strengthen prevention and mitigation strategies;
Stronger disincentives
- Digital platforms should move away from business models that prioritize engagement above human rights, privacy and safety;
- Advertisers and digital platforms should ensure that advertisements are not placed next to online mis- or disinformation or hate speech, and that advertising containing disinformation is not promoted;
- News media should ensure that all paid advertising and advertorial content is clearly marked as such and is free of mis- and disinformation and hate speech;
Enhanced trust and safety
- Digital platforms should:
- Ensure safety and privacy by design in all products, including through adequate resourcing of in-house trust and safety expertise, alongside consistent application of policies across countries and languages;
- Invest in human and artificial intelligence content moderation systems in all languages used in countries of operation, and ensure content reporting mechanisms are transparent, with an accelerated response rate, especially in conflict settings;
- All stakeholders should take urgent and immediate measures to ensure the safe, secure, responsible, ethical and human rights-com- pliant use of artificial intelligence and address the implications of recent advances in this field for the spread of mis- and disinformation and hate speech.
Next Steps
- The United Nations Secretariat will undertake broad consultations with a range of stakeholders on the development of the United Nations Code of Conduct, including mechanisms for follow-up and implementation. This could include the establishment of an independent observatory made up of recognized experts to assess the measures taken by the actors who commit to the Code of Conduct, and other reporting mechanisms.
- To support and inform the Code, the United Nations Secretariat may carry out in-depth studies to enhance understanding of information integrity globally, especially in underresearched parts of the world.
- The Secretary-General will establish dedicated capacity in the United Nations Secretariat to scale up the response to online mis- and disinformation and hate speech affecting United Nations mandate delivery and substantive priorities. Based on expert monitoring and analysis, such capacity would develop tailored communication strategies to anticipate and/or rapidly address threats before they spiral into online and offline harm, and support capacity-building of United Nations staff and Member States. It would support efforts of Member States, digital platforms and other stakeholders to adhere to and implement the Code when finalized.
Conclusion
Strengthening information integrity on digital platforms is an urgent priority for the international community. From health and gender equality to peace, justice, education and climate action, measures that limit the impact of mis- and disinformation and hate speech will boost efforts to achieve a sustainable future and leave no one be- hind. Even with action at the national level, these problems can only be fully addressed through stronger global cooperation. The core ideas outlined in this policy brief demonstrate that the path towards stronger information integrity needs to be human rights-based, multi-stakeholder, and multi-dimensional. They have been distilled into a number of principles to be considered for a United Nations Code of Conduct for Information Integrity on Digital Platforms that would provide a blueprint for bolstering information integrity while vigorously upholding human rights. I look forward to collaborating with Member States and other stakeholders to turn these principles into tangible commitments.
Annex
CONSULTATIONS WITH MEMBER STATES AND OTHER RELEVANT STAKEHOLDERS
The ideas in the present policy brief draw on the proposals outlined in the report entitled “Our Common Agenda” (A/75/982), which benefited from extensive consultations with Member States, the United Nations system, thought leaders, young people and civil society actors from all around the world. The policy brief responds, in particular, to the rich and detailed reflections of Member States and other stakeholders on Our Common Agenda over the course of 25 General Assembly discussions.
In advance of the publication of the present policy brief, consultations were carried out with Member States, including through an informal briefing to the Committee on Information, which all non-Committee members were invited to join. Discussions were also held with civil society partners, academics, experts and the private sector, including technology companies.
Broad consultations will be conducted in the development of the Code of Conduct, ahead of the Summit of the Future.