7,611
Views
1
CrossRef citations to date
0
Altmetric
Articles

How African countries respond to fake news and hate speech

ORCID Icon, ORCID Icon & ORCID Icon
Pages 86-103 | Received 15 Dec 2020, Accepted 07 Oct 2021, Published online: 09 Nov 2021

ABSTRACT

While scholars have already identified and discussed some of the most urgent problems in content moderation in the Global North, fewer scholars have paid attention to content regulation in the Global South, and notably Africa. In the absence of content moderation by Western tech giants themselves, African countries appear to have shifted their focus towards state-centric approaches to regulating content. We argue that those approaches are largely informed by a regime’s motivation to repress media freedom as well as institutional constraints on the executive. We use structural topic modelling on a corpus of news articles worldwide (N = 7′787) mentioning hate speech and fake news in 47 African countries to estimate the salience of discussions of legal and technological approaches to content regulation. We find that, in particular, discussions of technological strategies are more salient in regimes with little respect for media freedom and fewer legislative constraints. Overall, our findings suggest that the state is the dominant actor in shaping content regulation across African countries and point to the need for a better understanding of how regime-specific characteristics shape regulatory decisions.

1. Introduction

Online platforms gained enormous traction in political and social discourse, with platforms like Facebook evolving into transnational companies that are ‘unmatched in their global reach and wealth’ (Gorwa, Citation2019, p. 860). Social media platforms in particular have been blamed for poor efforts to moderate content in many instances around the world, failing to protect users from foreign influence during elections in the USA and France (Walker et al., Citation2019, p. 1532) or to adequately moderate hate speech in Ethiopia inciting violent ethnic protest (Gilbert, Citation2020). ‘We take misinformation seriously,’ Facebook CEO Mark Zuckerberg (Citation2016) wrote just weeks after the 2016 elections in the USA. In the years since, the question of how to counteract the damage done by ‘fake news’ has become a pressing issue both for technology companies and governments across the globe. Indeed, there is a growing debate about how to adequately regulate online content predominantly taking place in Europe and North America (Iosifidis & Andrews, Citation2020).

Yet, how are fake news and hate speech regulated across African countries? In this paper, we use news coverage of fake news and hate speech in Africa to analyse how regulatory strategies are framed, and how these frames are predicted by different regime characteristics. In essence, our analysis of 7′787 news articles covering 47 African countries suggests that in the absence of proactive content moderation by the platforms, discussions regarding the regulation of fake news and hate speech mostly centre on state-centric strategies.

underlines the salience of ‘fake news’, ‘hate speech’, ‘misinformation’, and ‘disinformation’ in news coverage of African countries. The trend over the last five years suggests that these issues have gained increasing importance in public discourse. The enormous spread of misinformation related to the COVID-19 pandemic on Facebook in South Africa and Nigeria (Africa Check, Citation2020; Ahinkorah et al., Citation2020, p. 2) further underlines this trend. The fact that both humans and bots are used in several African countries to spread government-propaganda and discredit public dissent online (Bradshaw & Howard, Citation2019) highlights the challenges related to limiting hate speech and fake news in African contexts. Indeed, in more authoritarian contexts, domestic governments themselves seek to manipulate both information and discourse to ensure their regime’s survival.

Figure 1. Number of articles including the search terms, 2015-2019.

Note: Total number of news articles including the terms “fake news”, “hate speech”, “misinformation”, or “disinformation”, per week between 01.01.2015 and 31.12.2019 (N = 7′787).

Figure 1. Number of articles including the search terms, 2015-2019.Note: Total number of news articles including the terms “fake news”, “hate speech”, “misinformation”, or “disinformation”, per week between 01.01.2015 and 31.12.2019 (N = 7′787).

How African countries respond to fake news and hate speech is a highly relevant question, especially in the absence of content moderation by Western tech giants. While platforms have started to engage in content moderation around the world, they appear comparatively inactive on the African continent. In 2019, upon request from governments, courts, civil society organizations, and ‘members of the Facebook community’ (Facebook, Citationn.d.), Facebook removed content from its platform in several thousands of instances in countries like Pakistan (N = 7′960), Mexico (N = 6′946), Russia (N = 2′958), or Germany (N = 2′182), but it hardly removed any content in Africa. In fact, Morocco had the highest number of content removals, with N = 6 (Facebook, Citationn.d.). Twitter’s transparency reports suggest similar figures for African countries (Twitter, Citationn.d.). The lack of content moderation in African countries seems particularly counterintuitive given the vast use of social media platforms, such as Facebook, on the African continent (e.g., Bosch et al., Citation2020; Nothias, Citation2020).

Theoretically, we build on Lessig (Citation1999) and Boas (Citation2006) framework of the regulation of code as well as recent scholarship on regime survival to explain how regime-specific characteristics shape the prevalence of legal and technological discussions about online content regulation. Empirically, our results underline the importance of regime characteristics to understand debates about online content regulation. Technological approaches to content regulation such as blocking or censoring online content are more commonly discussed in regimes that rely on the repression of media and in which the executive’s actions are less constrained by legislatures. Legal approaches to content regulation, such as legislation passed by parliament, in particular the criminalization of hate speech, are more commonly discussed in countries respecting media and press freedom, yet not necessarily in countries with higher institutional constraints on the executive. Overall, our findings suggest that regime-specific characteristics pave the way for different strategies to regulate content, some of which may have profound consequences for the freedom of expression online.

We proceed with a theoretical section in which we combine insights from internet governance and comparative politics to formulate expectations about the prevalence of technological and legal regulation in public discourse. Subsequently, we present our data and explain how we employ structural topic modelling to identify regulatory frames in our body of collected texts. We then present results from regression analyses and discuss these findings in light of recent regulatory trends in Africa.

2. Theory

Recent scholarly efforts seek to understand determinants and effects of governments’ attempts to control online spaces using censorship (Hellmeier, Citation2016), internet shutdowns (Hassanpour, Citation2014; Freyburg & Garbe, Citation2018; Rydzak et al., Citation2020) or online surveillance (Michaelsen, Citation2018). However, these studies do not take into account the legitimate need for governments to address, prevent and punish the spread of hate speech and fake news. Crucially, the aim and motivation of such regulation can be legitimate as long as it addresses citizens’ needs (Helm & Nasu, Citation2021). By connecting insights from scholarship on internet governance and regime survival, our aim is to explain how variation in (1) regimes’ motivations to control Information and Communication Technologies (ICT) and in (2) institutional constraints on the adoption of regulation result in differences in the framing of regulatory strategies addressing fake news and hate speech.

We assume that regulation of hate speech and fake news covered in news reports can be seen as regulatory frames (Gilardi et al., Citation2021, p. 23) that represent different perspectives on regulation. Following DiMaggio et al. (Citation2013), we consider that news reports offer a useful mirror of societal debates, both because they report on issues when these are under consideration by political institutions, and because they reflect debates among the informed public. Furthermore, by considering not only African but also global news reports, we overcome potential biases in the way regulation is framed in more illiberal countries. We assess how well news reports reflect actual regulatory strategies using information from Freedom on the Net reports, annually released by Freedom House (Citation2015, Citation2016, Citation2017, Citation2018, Citation2019, Citation2020) that cover legal and technological aspects of internet regulation in 16 African countries (see Methods section; Appendix D).

2.1. Legal versus technological approaches controlling online content

Lessig (Citation1999) distinguishes, broadly speaking, between institutional and architectural means (or legal and technological means as we will call them during the remainder of this paper) to regulate online space (cf. Boas, Citation2006, p. 4f.). According to Lessig (Citation2006, pp. 124–125), states can control the technological architecture of the internet through executive decisions, thus influencing or restricting the production of and access to specific content. He (2006, pp. 136–37) argues that technological approaches enable states to regulate online content without having to suffer any political consequence. Legal approaches to content regulation are the predominant institutional strategy to shape access to and production of online content (Lessig, Citation2006, p. 130), including both the formulation of legislation in the form of bills, laws, and acts as well as the judicial review of existing legislation by courts.

Applying this distinction to how governments seek to regulate online content, we argue that the main difference between technological and legal regulation is that technological strategies are an ex-ante approach to prevent the production of online content in the first place, while legal strategies are mostly ex-post, removing and/or punishing harmful content after it was produced or shared (Frieden, Citation2015). These two approaches are not mutually exclusive and are, in fact, often employed in combination with one another.

2.2. Motivations for controlling online content

We acknowledge that a differentiated understanding of regime type is needed when studying politics in Africa. One important and useful distinction with regards to a government’s motivations for controlling the flow of information and communication, is the degree to which a political regime is relying on people’s informed vote and a viable opposition as sources of its legitimacy. Most African regimes qualify as ‘electoral regimes’ (Schedler, Citation2002, p. 36), meaning they hold elections and tolerate some competition but also violate minimal democratic norms so severely and systematically that they cannot be classified as full-fledged democracies. In countries in which the ruler is not (re-)determined by means of free and fair elections, the government usually relies on a whole ‘menu of manipulation’ to stay in power (Schedler, Citation2002). This includes the control of media and civil society actors because a strong and well-informed civil society ‘can hold governments accountable beyond elections’ (Mechkova et al., Citation2019, p. 42). Traditionally, in order to control information and communication, authoritarian rulers rely on manipulation of public discourse through the control of media outlets (Kellam & Stein, Citation2016) or heavy restrictions on civil society (Christensen & Weinstein, Citation2013).

In the digital age, internet and social media provide both civil society and media actors with new means to access and share information (Breuer et al., Citation2015; Eltantawy & Wiest, Citation2011). Authoritarian rulers might therefore require new regulatory strategies to also control the flow of internet-based information and communication. In particular, they need to overcome the challenge posed by some of the decentralized and low-cost features of the internet that facilitate the organization of collective action without formal organization (Bennett & Segerberg, Citation2012). From a regulatory perspective, authoritarian regimes should hence be inclined to use preventive measures to keep civil society and media actors from putting pressure on the incumbent by using ICT for mobilization purposes (Dresden & Howard, Citation2015; Goetz & Jenkins, Citation2005, p. 20). We therefore expect that those regimes that traditionally rely on the repression of media and press freedom are more likely to employ technological ex-ante strategies that prevent the production and sharing of content in the first place. This is likely to affect how regulating online hate speech and fake news is framed in media reports:

H1a: With increasing levels of press and media freedom, the salience of technological regulatory frames decreases.

H1b: With increasing levels of press and media freedom, the salience of legal regulatory frames increases.

2.3. Institutional constraints to controlling online content

The extent to which authoritarian regimes can impose means of regulation that prevent the creation of digital content should not only depend on their tendency to repress press and media freedom in general but also on institutional constraints. We argue that authoritarian regimes can apply more preventive measures of regulation without facing the need for approval by the legislature or the review by the judiciary. They should therefore be more inclined to use technological means of regulation. In turn, in regimes in which the executive faces more constraints by other branches of power, discussions about legal approaches to content regulation should be more prominent.

The separation of powers aims to prevent a government’s abuse of power (Rose-Ackerman, Citation1996). In many authoritarian regimes, institutions such as legislatures or courts serve as a way to co-opt the opposition rather than provide de facto oversight (Gandhi, Citation2008; Rakner & van de Walle, Citation2009; Shen-Bayh, Citation2018). It is therefore important to focus on the de facto capacity of such institutions to constrain executive decisions and hence the government’s capacity to regulate online content. Legislatures can challenge a government through non-confidence votes for example (Mechkova et al., Citation2019). This capacity might be even stronger when opposition actors are represented in the legislature (Herron & Boyko, Citation2015). Some African legislatures have become powerful institutions ‘in terms of checking the executive, contributing to the processes of policy-making, and indeed as a monitor of policy implementation’ (Bolarinwa, Citation2015, p. 20). Independent legislatures are important actors in Africa ‘assessing proposed legislation, drafting amendments, […] asking questions, attending committee and plenary meetings, participating in debates or voting’ (Nijzink et al., Citation2006, p. 315), all of which should be reflected in broader societal debates about different steps in the process of legislation. High courts have the possibility to sanction government actions. Examples from Africa highlight their capacity to challenge even fundamental government decisions such as amendments to the constitution to overcome presidential term limits (Vondoepp, Citation2005).

In a country with independent legislatures and high courts, which effectively constrain the government, the executive is thus more limited in its ability to regulate fake news and hate speech. Ad hoc technological regulation to prevent the circulation of fake news and hate speech appears to be more challenging in such an environment compared to contexts without institutional constraints, as highlighted by an example from Ethiopia. In response to violent protest and the circulation of fake news, the Ethiopian government repeatedly shut down internet access in part of the country. As outlined by Abraha (Citation2017, p. 302) this strategy ‘usually take[s] place in the absence of any specific legislative framework’. We hence argue that discussions about legal strategies to regulate content are more prevalent in regimes where the government is de facto constrained by legislatures and high courts:

H2a: With increasing levels of constraint by legislatures and courts, the salience of legal regulatory frames increases.

H2b: With increasing levels of constraint by legislatures and courts the salience of technological regulatory frames decreases.

3. Methods

We assess legal and technological regulatory frames by analysing how regulation of hate speech and fake news are reported and discussed in news coverage of Africa. Importantly, news items come from both African and non-African publishers. We include news items from non-African publishers as reporting on politically contested issues like misinformation and hate speech might be scarce or biased in more authoritarian countries where news outlets are often owned by government authorities (Stier, Citation2015).

Still, domestic African media outlets are prominent in our sample (like Nigerian Vanguard, The Punch, and The Sun, with 17% of the news stories combined) or African reproduction of media content (like AllAfrica with 13% of the news stories). In contrast to analysing actual regulatory advances, news reporting can provide a sense of debates surrounding regulatory strategies pursued by governments and may provide an indication of regulation even before a law has been formally adopted (DiMaggio et al., Citation2013).

3.1. Corpus

Our data consists of 7′787 English-language news articles from a wide range of news outlets (N = 380), covering both digital and digitalized printed press, in 47 African countries. These articles are sourced from Factiva, containing the terms ‘hate speech’, ‘fake news’, ‘misinformation’ and/or ‘disinformation’ as well as terms related to online activity in the title or article published between 2015 and 2019. The Dow Jones Factiva database is a digital archive of global news content which is frequently used by scholars analysing media reporting on African countries (e.g., Bunce, Citation2016; Obijiofor & MacKinnon, Citation2016). Appendix A provides details of the full Factiva search query, which in total produced 22′457 news stories. To ensure that our corpus only consists of news stories discussing fake news and hate speech in online contexts, we subset the full corpus of news stories, only including articles that mention pre-defined words for online aspects. For each article in the final corpus, we only keep those paragraphs in which our key online terms are mentioned.Footnote1 provides an overview of the final corpus of 7′787 news stories.

Table 1. Description of the textual corpus.

3.2. Structural topic model

In order to analyse how regulatory strategies are framed, and to test our hypotheses about how these frames are predicted by different regime characteristics, we apply structural topic modelling (STM) (Roberts et al., Citation2019). We first estimate topic models ranging from 10 to 50 topics per model using the stm package in R (Roberts et al., Citation2019). We choose the 35-topic model as the most meaningful in terms of topic quality, based on quantitative measures for exclusivity and semantic coherence, and qualitative evaluation of the topics’ interpretability (see Appendix B). Because the STM analysis relies on the probabilistic topic model technique Latent Dirichlet Allocation (see Blei, Citation2012), a technique which uses word counts and not the order of words, it is up to the researcher to infer meaning from the words and topics that appear, rather than assert it (Grimmer & Stewart, Citation2013, p. 272).

Based on the words in each topic and a close reading of the twenty most representative articles, we identify two topics as indicators for the framing of technological and legal approaches to content regulation. Representative articles can be found using the findThoughts function of the stm package, which provides documents highly associated with particular topics (Roberts et al., Citation2019, p. 14). To validate our interpretation and labelling of the selected topics, four human coders read and manually coded a sample of the most representative texts for each topic. The coders’ agreement with the assignment of the structural topic model is around 70–75 percent (see Appendix C).

We further assess how well the identified topics for ‘legal’ and ‘technological’ approaches to content regulation in news coverage capture actual regulatory steps undertaken by African governments (see Appendix F). Specifically, we compare our country-year mean topic proportions with data from the Freedom on the Net reports (Freedom House, Citation2015, Citation2016, Citation2017, Citation2018, Citation2019, Citation2020), first through a t-test and then by investigating four cases more qualitatively. According to the results, news reports provide a fair indication of different legal and technological regulations by African governments For the remainder of this study, we use the expected proportion of each topic as dependent variable.

3.3. Covariates

To predict the expected proportion of each topic, we use three different indicators from the Varieties of Democracy (V-Dem) project, version 10 (Coppedge et al., Citation2020). First, we use an aggregated index to assess media and press freedom (v2x_freexp_altinf) that ranges from 0 to 1 assessing the extent to which citizens are able to ‘make an informed choice based on at least some minimal possibilities for collective deliberation’ (Teorell et al., Citation2019). Second, following Mechkova et al. (Citation2019), we use two different indicators to assess de facto rather than de jure accountability mechanisms through legislatures (v2xlg_legcon) and high courts (v2x_jucon) both of which range from 0 to 1. For both indicators, higher values indicate more freedom and more constraints on governments, respectively. In addition, we include a variable on state ownership of the telecom sector per country and year to control for a government’s capacity to block internet access (Freyburg & Garbe, Citation2018). Specifically, the variable indicates the proportion of the telecom sector that is majority state-owned. Here, data comes from the Telecommunications Ownership and Control Dataset (Freyburg et al., Citation2021).

3.4. Methods

To estimate the effects of press and media freedom and institutional constraints on the proportion of the three selected topics, we use Linear Mixed Models (LMM; Baayen, Citation2008) and include country as random intercepts to acknowledge that the articles are nested in countries and time fixed effects. We use the logarithm of the proportion of topics as the distribution of these variables is right-skewed. All predictors are standardized. After fitting the model, we check whether the assumptions of normally distributed and homogeneous residuals are fulfilled. Appropriate tests indicate no substantial deviations from these assumptions. Finally, collinearity determined for a standard linear model without random effects, appeared to be no major issue (maximum Variance Inflation Factor: 5; Field, Citation2009).

4. Results

We identify two topics that reflect the two dominant state-centric regulatory strategies technological and legal approaches: Topic 31, which we label ‘technological approaches’, represents regulatory frames of governments using technological means to block, manipulate, or censor specific online content; and Topic 5, which we label ‘legal approaches, reflects legislative strategies to regulate the production of fake news and hate speech. We illustrate how each of those topics reflects different types of regulatory strategies with excerpts from representative news articles from the corpus .

Table 2. Topics related to state-centric online regulation in news coverage

4.1. Technological approaches

Topic 31 appears to be related to more technological approaches to content regulation, with terms including ‘shutdown, ‘access, ‘blackout, and ‘block’. Both representative articles below point to the problem that fake news and hate speech might often be used as a pretence to prevent opposition actors from accessing specific content or sharing information. One article exemplifies how a government, here the Ethiopian government, uses the blocking of specific websites as a means to prevent the spread of ‘rumours’:

‘Amid reports of violent clashes that have led to at least 15 deaths, the Ethiopian government has partially blocked internet access […]. The government has justified such action in the past as a response to unverified reports and rumors, noting that social media become flooded with unconfirmed claims and misinformation when violence erupts.’ (Solomon, Citation2017)

Another representative article focusing on government blocking in Cameroon underlines the preventive character of such measures:

‘[Cameroon] endured at least two Internet cuts since January last year with government saying the blackouts were among ways of preventing the spread of hate speech and fake news as the regime tried to control misinformation by separatists groups in the Northwest and Southwest.’ (The Citizen, Citation2018)

4.2. Legal approaches

Topic 5 seems to be concerned mostly with legal processes, as shown by the combination of terms like ‘legisl[ation]’, ‘bill’, ‘regul[ation]’, ‘fine’ or ‘prosecut[ion]’. A closer look at a representative article exemplifies that this topic embraces news coverage of specific legislation such as in Kenya:

Kenyan President Uhuru Kenyatta signed a lengthy new Bill into law, criminalising cybercrimes including fake news […] The clause says if a person ‘intentionally publishes false, misleading or fictitious data or misinforms with intent that the data shall be considered or acted upon as authentic,’ they can be fined up to 5 000 000 shilling (nearly R620 000 [43′865 USD]) or imprisoned for up to two years. (Mail & Guardian, Citation2018)

Before Kenyatta signed the Bill, there were demands to have Parliament review the law to make sure that it does not violate the right to media freedom and expression. Similar laws or proposed laws are discussed in news articles covering Nigeria (Adegbo, Citation2019), Botswana (The Botswana Gazette, Citation2017), or Ethiopia (Ethiopian News Agency, Citation2019), all of which focus on criminalizing the publication of fake news or hate speech and holding to account the individual. In few cases, articles also point to legal approaches that would make it possible to hold internet service providers (ISPs) liable who fail to moderate content appropriately as indicated by a bill discussed in South Africa (Eloff, Citation2019).

4.3. The influence of media freedom and institutional constraints on regulation

We use our topics ‘technological approaches’ and ‘legal approaches’ as dependent variables and estimate the effect of a country’s press and media freedom as well as institutional constraints on the expected proportion of each of these topics. depicts the results; more detailed results from the linear mixed models can be found in Appendix D.

Figure 2. Marginal effects of covariates on expected topic proportions.

Note: Marginal effects are calculated using the ggpredict() function (Lüdecke, Citation2021); points indicate the expected topic proportion per text; in contrast to the statistical models (see Appendix D), the marginal effects are calculated using models in which the variables are not logarithmized and/or standardized to facilitate interpretation.

Figure 2. Marginal effects of covariates on expected topic proportions.Note: Marginal effects are calculated using the ggpredict() function (Lüdecke, Citation2021); points indicate the expected topic proportion per text; in contrast to the statistical models (see Appendix D), the marginal effects are calculated using models in which the variables are not logarithmized and/or standardized to facilitate interpretation.

First, results from the linear mixed models reveal a differential impact of a country’s level of press and media freedom on the expected reporting of technological and legal approaches to regulation. Increasing levels of press and media freedom are associated with decreasing levels of technological regulatory frames (B = −0.55, SE = 0.09). This supports Hypothesis 1a that countries traditionally relying on the repression of press and media are more likely to appear in frames related to technological strategies of content regulation. Furthermore, press and media freedom is positively associated with legal frames (B = 0.17, SE = 0.09), supporting Hypothesis 1b that countries respecting press and media freedom are more likely to be associated with legal strategies of content regulation.

Second, institutional constraints vary in their effects on the expected proportion of technological and legal frames. Legislative constraints are associated negatively with the expected proportion of technological frames (B = −0.44, SE = 0.08) whereas judicial constraints are positively associated with the expected proportion of technological frames (B = 0.27, SE = 0.08). Overall, this suggests that only legislative but not judicial constraints are negatively associated with frames of technological regulation and lends mixed support for Hypothesis 2b. Furthermore, legislative constraints are negatively associated with legal frames (B = −0.37, SE = 0.09), whereas judicial constraints are positively associated with legal frames (B = 0.17, SE = 0.08). Overall, this provides mixed support for Hypothesis 2a, suggesting that in contrast to judicial constraints, legislative constraints are not necessarily associated with the reporting of legal strategies.

4.4. Discussion & limitations

Our investigation suggests that both traditional restrictions on press and media freedom as well as institutional constraints influence the salience of different regulatory strategies as reported in news outlets. Given that our indicators for different regulatory strategies are informed by computer-assisted text analysis rather than in-depth analysis of all articles in the corpus, we discuss our findings in light of country-specific examples.

The salience of regulatory frames of technological approaches to address fake news and hate speech appears to be higher in countries that traditionally restrict press and media freedom and that are less constrained by legislative institutions. This is also reflected by the growing trend across authoritarian African rulers to block internet access during elections (Freyburg & Garbe, Citation2018; Garbe, Citation2020). Indeed, our findings suggest that those rulers who have strong incentives to prevent the production and spread of content – often because it is considered potentially harmful to the regime – are more likely to use technological means to restrict access to and production of online content. Our findings further suggest that strong legislative constraints on the executive might prevent governments from using technological means of blocking. On the other hand, our findings indicate that regimes with strong(er) judicial constraints on the executive may still revert to technological strategies of content regulation. This is exemplified by the shutdown of social media in Zimbabwe amid protests in 2019 which was later challenged by Zimbabwe’s high court (Asiedu, Citation2020). Given that governments who seek to fundamentally restrict access to and production of content online often do so in response to pressing political issues, ex-post judicial review of such measures might not deter governments from doing so. While some observers recognize the legitimate aim to contain the spread of fake news (Madebo, Citation2020), there is also widespread concern about the potential harm of such preventive measures in over-censoring potentially important information such as news related to Covid-19 (Nanfuka, Citation2019).

Our findings further indicate that legal approaches to regulating fake news or hate speech, i.e., media coverage of the introduction of bills, laws, and acts, are not more prevalent in those regimes with strong legislative constraints. This may reflect the increasing importance of law-making as a political tool of power consolidation and illiberal practices, also known as ‘autocratic legalism’ (Scheppele, Citation2018, p. 548). In fact, many African rulers started introducing legislation on the production and spread of content online. Kenya’s Computer Misuse and Cybercrime Act, for instance, criminalizes the ‘publication of false information in print, broadcast, data or over a computer system’ (Citation2018, Art 22, 23) and also explicitly refers to the publication of ‘hate speech’. Digital human rights defenders have observed many of the changes in the legal landscape in both authoritarian and democratic countries with worry. Regulations specifically criminalizing online content that is regarded as misinformation or hate speech are often ‘inherently vague, and […] create a space for abuse of the law to censor speech’ (Taye, Citation2020). While Helm and Nasu (Citation2021) argue that criminal sanctions can be an effective way to counter hate speech, they underline that it is necessary to find an appropriate balance between censoring information and respecting freedom of expression. In addition to bills criminalizing the publication or spread of fake news, authoritarian regimes also seem to develop more indirect means of legislation that can be described as ex ante measures to prevent the production of fake news and hate speech. For instance, Tanzania, Lesotho, and Uganda, all introduced laws that indirectly prevent people from sharing content online either through fees on social media use itself or fees that are required from online bloggers (Karombo, Citation2020). Overall, our STM approach is limited in grasping more nuanced types of legislation and further qualitative work is needed to better understand how regimes differ in their legislative approach to regulating fake news and hate speech and to what extent legislatures affect this process. In addition, the increasing use of bots by African governments can also be seen as regulatory strategy in itself requiring more fine-grained approaches to investigate differences across countries (Nanfuka, Citation2019).

We acknowledge that there might be non-state solutions to regulation as, for instance, reflected in Topic 14 (see Appendix C). While governments appear to be the most prevalent actors emerging from our analysis of media coverage on hate speech and fake news in Africa, news reporting also points to other approaches, such as bottom-up initiatives to improve fact-checking skills, to regulate fake news and hate speech. This might reflect the fact that, facing increasing pressure on fundamental rights, civil society in Africa is advocating for online platforms to meaningfully invest in content moderation in Africa and collaborate with local civil society (Owono, Citation2020; Dube et al., Citation2020).

Finally, we want to highlight three limitations of our study. First, our approach using news coverage of African countries enabled us to assess the public discourse surrounding the regulation of fake news and hate speech. This has the advantage that we also include discussions about the regulation of fake news and hate speech, often before they translate into actual legislation. However, it is unclear to what extent news reports reflect actual regulation across African countries. As our validation highlights, news reports provide a good indication of technological approaches to regulate fake news (see Appendix D). Yet, it is less clear how well news reports reflect legal approaches to content regulation. The fact that most data sources on African legislation do not directly assess the extent to which legislation regulating the digital space is meant to address fake news, makes it difficult to validate the fit of news reports. Empirical studies comparing actual laws explicitly addressing fake news as well as technological manipulation of online activity beyond shutdowns are hence encouraged. Second, our sample is biased towards large African countries and countries with a high degree of digitalized press, like South Africa and Nigeria, which represent up to 20 and 40 percent in our sample respectively. As both are prominent and dominant countries on the continent, however, we can assume that they are important actors in both driving and shaping discussions on how online content should be regulated. Third, the salience of the two strategies is subject to variation over time (see Appendix F) and highlights that especially legal frames have only recently gained importance in the African context.

5. Conclusion

Our study contributes to the growing discussion on content regulation in two ways. Theoretically, we add to the understanding of online regulation by showing that regime-specific characteristics can alter a government’s choice of different regulatory strategies. Empirically, we find that public discourse on online content regulation in Africa points to the relevance of technological and legal strategies pursued by governments. Discourse on technological approaches to content regulation is more prominent in coverage of countries with lower levels of media and press freedom and legislative constraints. Our analysis further suggests that legal frames are more dominant in coverage of countries with judicial constraints, but not in coverage of countries with more legislative constraints. More qualitative insights suggest that criminalization is among the dominant legal strategies. While criminal regulation can be an effective strategy to counter hate speech and fake news, Helm and Nasu (Citation2021, p. 327) also warn about the potential for abuse of laws to supress dissent in more authoritarian regimes.

Overall, our analysis points to the central actors when it comes to content regulation in Africa: African governments. While theory has so far tended to either follow Lessig (Citation1999, Citation2006) and focus on content regulation in democracies, or to focus on censorship in authoritarian regimes (Stoycheff et al., Citation2020; Keremoğlu & Weidmann, Citation2020), our analysis demonstrates that these issues cannot always be easily separated. News reporting on African countries underlines that all regimes face issues of fake news and hate speech and seek to find solutions to manage them. While technological strategies to address fake news and hate speech (including shutting down internet and blocking specific content) appear to be more prominent in regimes with low respect for media and press freedom and fewer institutional constraints, our results indicate that the same regimes may also revert to more legal means to regulate content. Not only technological but also legal strategies of content regulation may have severe implications for freedom of speech (Helm & Nasu, Citation2021), especially, but not only, in countries facing weak institutional constraints. Overall, our paper highlights that the regulation of fake news and hate speech are also pressing issues beyond the Western world. In turn, the current prevalence of state regulation addressing problems of fake news and hate speech points to a need to strive for multi-stakeholder approaches across continents.

Author contributions

All authors contributed equally to the design and implementation of the research, to the analysis of the results and to the writing of the manuscript.

Supplemental material

Supplementary Material.docx

Download MS Word (337.6 KB)

Acknowledgements

We wish to extend special thanks to Daniela Stockmann, Lance Bennett, and Tina Freyburg for valuable comments in developing this paper. We thank Mikael Johannesson and Theo Toppe for inputs on the data preparation and analysis, and Caroline Borge Bjelland and Jean-Baptiste Milon for their research assistance, partly funded by the Norwegian Research Council project ‘Breaking BAD: Understanding the Backlash Against Democracy in Africa’ (#262862). We also thank the research groups at the Comparative Politics Department at the University of St.Gallen and the Department of Administration and Organization Theory of the University of Bergen, as well as the two anonymous reviewers for their constructive and encouraging comments.

Disclosure statement

No potential conflict of interest was reported by the author(s).

Data availability statement

The data that support the findings of this study are available from the Factiva Global News Monitoring & Search. Restrictions apply to the availability of these data, which were used under license for this study. R scripts for data preparation and analysis are available from the authors in the Github repository https://github.com/lisagarbe/ContentRegulationAfrica

Additional information

Funding

This work was supported by the Norwegian Research Council [Grant numbers 288489, 262862].

Notes on contributors

Lisa Garbe

Lisa Garbe is a postdoctoral researcher at the WZB Berlin Social Science Center. Her research focuses on inequalities in internet provision and use with a focus on authoritarian-developing contexts.

Lisa-Marie Selvik

Lisa-Marie Selvik is a PhD Candidate at the Department of Comparative Politics, University of Bergen. Her work focuses on information rights, with a focus on contentious processes of advocacy on freedom of information laws in Africa.

Pauline Lemaire

Pauline Lemaire is a doctoral researcher at Chr. Michelsen Institute, and a PhD candidate at the Department of Comparative Politics, University of Bergen. Her research focuses on regime - youth interactions as mediated by social media.

Notes

1 In preparing for the textual analysis, the body of textual data was properly pre-processed, white space, punctuation, and so-called ‘stopwords’ (the, is, are, etc.) were removed, as well as the search terms we used to delimit our body of texts (Benoit et al., Citation2018). Our full script for importing, preprocessing and analysing the corpus is available on GitHub: https://github.com/lisagarbe/ContentRegulationAfrica.

References