The digital gender divide refers to the gap in access and use of the internet between women* and men, which can perpetuate and exacerbate gender inequalities and leave women out of an increasingly digital world. Despite the rapid growth of internet access around the globe (97% of people now live within reach of a mobile cellular network), women are still 17% less likely to use the internet compared to men; a gap that is actually widening in many low- and middle-income countries (LMICs) where women are 8% less likely than men to own a mobile phone and 20% less likely to actually access the internet on a mobile device.
Civil society leader in La Paz, Honduras. The gender digital divide affects every aspect of women’s lives. Photo credit: Honduras Local Governance Activity / USAID Honduras.
Though it might seem like a relatively small gap, because mobile phones and smartphones have surpassed computers as the primary way people access the internet, that statistic translates to 390 million fewer women online in LMICs than men. Without access to the internet, women cannot fully participate in different aspects of the economy, join educational opportunities, and fully utilize legal systems.
The digital gender divide does not just stop at access to the internet however; it is also the gap in how women and men use the internet once they get online. Studies show that even when women own mobile phones, they tend to use them less frequently and intensively than men, especially for more sophisticated services, including searching for information, looking for jobs, or engaging in civic and political spaces. Additionally, there is less locally relevant content available to women internet users, because women themselves are more often content consumers than content creators.
The digital gender divide is also apparent in the exclusion of women from leadership or development roles in the information and communications technology (ICT) sector. In fact, the proportion of women working in the ICT sector has been declining over the last 20 years. In the United States alone, women only hold around 18% of programming and software development jobs, down from 37% in the 1980s. This helps explain why software, apps, and tools do not often reflect the unique needs that women have, further alienating them. Apple, for instance, whose tech employees are 77% male, did not include a menstrual cycle tracker in its Health app until 2019, five years after it was launched (though it did have a sodium level tracker and blood alcohol tracker during that time).
Ward nurses providing vaccines in Haiti. Closing the gender digital divide is key to global public health efforts. Photo credit: Karen Kasmauski / MCSP and Jhpiego
A NOTE ON GENDER TERMINOLOGY All references to “women” (except those that reference specific external studies or surveys, which has been set by those respective authors) is gender-inclusive of girls, women, or any person or persons identifying as a woman.
Why is there a digital gender divide?
At the root of the digital gender divide are entrenched traditional gender inequalities, including gender bias, socio-cultural norms, lack of affordability and digital literacy, digital safety issues, and women’s lower (compared to men’s) confidence levels navigating the digital world. While all of these factors play a part in keeping women from achieving equity in their access to and use of digital technologies, the relative importance of each factor depends largely on the region or country.
Affordability
In developing countries especially, the biggest barrier to access is simple: affordability. While the costs of internet access and of devices have been decreasing, they are often still too expensive for many people. While this is true for both genders, women tend to face secondary barriers that keep them from getting access, such as not being financially independent, or being passed over by family members in favor of a male relative. In Rwanda, an evaluation of The Digital Ambassador Programme pilot phase found that the costs of data bundles and/or access to devices were prohibitively expensive for a large number of potential women users, especially in the rural areas.
Education
Education is another major barrier for women all over the world. According to the Web Foundation, women who have some secondary education or have completed secondary school are six times more likely to be online than women with primary school or less.
Further, digital skills are also required to meaningfully engage with the Internet. While digital education varies widely by country (and even within countries), girls are still less likely to go to school over all, and those that do tend to have “lower self-efficacy and interest” in studying Science, Technology, Engineering and Math (STEM) topics, according to a report by UNICEF and the ITU. While STEM subjects are not strictly required to use digital technologies, these subjects can help to expose girls to ICTs and build skills that help them be confident in their use of new and emerging technologies. Without encouragement and confidence in their digital skills, women may shy away or avoid opportunities that are perceived to be technologically advanced, even when they do not actually require a high level of digital knowledge.
Social Norms
Social norms have an outsized impact on many aspects of the digital gender divide because they can also be a driving factor vis-à-vis other barriers. Social norms look different in different communities; in places where women are round-the-clock caregivers, they often do not have time to spend online, while in other situations women are discouraged pursuing STEM careers. In other cases, the barriers are more strictly cultural. For example, a report by the OECD indicated that in India and Egypt around one-fifth of women considered that the Internet “was not an appropriate place for them” due to cultural reasons.
Online Violence
Scarcity of content that is relevant and empowering for women and other barriers that prevent women from speaking freely and safely online are also fundamental aspects of the digital gender divide. Even when women access online environments, they face a disproportionate risk of gender-based violence (GBV) online: digital harassment, cyberstalking, doxxing, and the nonconsensual distribution of images (e.g., “revenge porn”). Gender minorities are also targets of online GBV. Trans activists, for example, have experienced increased vulnerability in digital spaces, especially as they have become more visible and vocal. Cyber harassment of women is so extreme that the UN’s High Commissioner for Human Rights has warned, “if trends continue, instead of empowering women, online spaces may actually widen sex and gender-based discrimination and violence.”
“…if trends continue, instead of empowering women, online spaces may actually widen sex and gender-based discrimination and violence.”
How is the digital gender divide relevant in civic space and for democracy?
The UN understands the importance of women’s inclusion and participation in a digital society. The fifth Sustainable Development Goal (SDG) calls to “enhance the use of enabling technology, in particular information and communications technology, to promote the empowerment of women.” Moreover, women’s digital inclusion and technological empowerment are relevant to achieving quality education, creating decent work and economic growth, reducing inequality, and building peaceful and inclusive institutions. While digital technologies offer unparalleled opportunities in areas ranging from economic development to health improvement, education, cultural development and political participation, gaps in the access to and use of these technologies and heightened safety concerns hinder women’s ability to access resources and information that are key to improve their lives and the wellbeing of their communities, and exacerbate gender inequalities.
Further, the ways in which technologies are designed and employed, and how data are collected and used impact men and women differently. Whether using technologies to develop artificial intelligence systems and implement data protection frameworks or just for the everyday uses of social media, gender considerations should be at the center of decision–making and planning in the democracy, rights and governance space.
Students in Zanzibar. Without access to the internet, women and girls cannot fully engage in economies, participate in educational opportunities or access legal systems. Photo credit: Morgana Wingard / USAID.
Initiatives that ignore gender disparities in access to the Internet and ownership and use of mobile phones will exacerbate the gender inequalities already experienced by the most vulnerable and marginalized populations. In the context of the Covid-19 pandemic and increasing GBV during lockdown, millions of women and non-binary individuals have been left with limited access to help, whether via instant messaging services, calls to domestic abuse hotlines, or discreet apps that provide disguised support and information to survivors in case of surveillance by abusers.
Most importantly, initiatives in the civic space must recognize women’s agency and knowledge and be gender-inclusive from the design stage. Women must participate as co-designers of programs and engaged as competent members of society with equal potential to devise solutions rather than perceived as passive victims.
There are a number of different areas to engage in that can have a positive impact in closing the digital gender divide. Read below to learn how to more effectively and safely think about some areas that your work may already touch on (or could include).
Widening job opportunities
According to the ITU, 90% of future jobs will require ICT skills. As traditional analog jobs in which women are overrepresented (such as in the manufacturing, service, and agricultural sectors) are replaced by automation, it is more vital than ever that women learn ICT skills to be able to compete for jobs. While digital literacy is becoming a requirement for many sectors, new, more flexible job opportunities are also becoming more common, and are eliminating traditional barriers to entry, such as age, experience, or location. Digital platforms can enable women in rural areas to connect with cities, where they can more easily sell goods or services. And part-time, contractor jobs in the “gig economy” (such as ride sharing, food delivery, and other freelance platforms) allow women more flexible schedules that are often necessitated by familial responsibilities.
Increasing access to financial services
The majority of the world’s unbanked population is comprised of women. Women are more likely than men to lack credit history and the mobility to go to the bank. As such, financial technologies can play a large equalizing role, not only in terms of access to tools but also in terms of how financial products and services could be designed to respond to women’s needs. In the MENA region, for example, where 52% of men but only 35% of women have bank accounts, and up to 20 million unbanked adults in the region send or receive domestic remittances using cash or an over-the-counter service, opportunities to increase women’s financial inclusion through digital financial services are promising.
Policy change for legal protections
There are few legal protections for women and gender-diverse people who seek justice for the online abuse they face. According to the UN Broadband Commission, only one in five women live in a country where online abuse is likely to be punished. In many countries, perpetrators of online violence act with impunity, as laws have not been updated for the digital world, even when online harassment results in real-world violence. In the Democratic Republic of Congo (DRC), for instance, there are no laws that specifically protect women from online harassment, and women who have brought related crimes to the police risk being prosecuted for “ruining the reputation of the attacker.” Sometimes cyber legislation even results in the punishment of victimized women: women in Uganda have been arrested under the Anti-Pornography Act after ex-partners released nude photos of them online. As many of these laws are new, and technologies are constantly changing, there is a need for lawyers and advocates to understand these laws and gaps in legislation to propose policies and amend existing laws that allow women to be truly protected online and safe from abuse.
Digital security education and digital literacy training
Digital-security education can help women (especially those at higher risk, like HRDs and journalists) stay safe online and attain critical knowledge to survive and thrive politically, socially, and economically in an increasingly digital world. However, there are not enough digital-safety trainers that understand the context and the challenges at-risk women face; there are few digital-safety resources that provide contextualized guidance around the unique threats that women face or have usable solutions for the problems they need to solve; and social and cultural pressures can prevent women from attending digital- safety trainings. Women can and will be content creators and build resources for themselves and others, but they first must be given the chance to learn about digital safety and security as part of a digital- literacy curricula. Men and boys, too, need training on online harassment and digital-safety education.
Connecting and campaigning on issues that matter
Digital platforms enable women to connect with each other, build networks and organize on justice issues. For example, the #MeToo movement against sexual misconduct in the media industry, which became a global movement, has allowed a multitude of people to participate in activism previously bound to a certain time and place. Read more about digital activism in the Social Media primer.
Young women at a digital inclusion center in the Peruvian Amazon. Cost remains the greatest barrier for women to access the Internet and own a mobile phone. Photo credit: Jack Gordon / USAID / Digital Development Communications.
There are many ways in which the digital gender divide is widening, even with external intervention. Read below to learn about some of the factors that are accelerating this gap, as well as how to mitigate for unintended – and intended – consequences.
Considering the gender digital divide as “women’s issue”
It is fundamental to consider the gender digital divide as a cross-cutting and holistic issue, affecting countries, societies, communities, and families, and not just as a “women’s issue.” As such, closing the gender gap in access, use and development of technology demands the involvement of societies as a whole. Approaches to close the divide must be holistic and take into account context-specific power and gender dynamics and include active participation of men in the relevant communities to make a sustained difference.
Further, the gender digital divide should not be understood as restricted to the technology space, but as a social, political, and economic issue with far-reaching implications.
National disasters intensify the education gap for women
Lockdowns and school closures due to the Covid-19 pandemic are contributing to the gender gap in education, especially in the most vulnerable contexts. According to UNESCO, more than 111 million girls who were forced out of school in March 2020 live in the world’s least-developed countries where gender disparities in education are already the highest. In Mali, Niger and South Sudan, countries with some of the lowest enrolment and completion rates for girls, closures left over 4 million girls out of school. Increasing domestic and caregiving responsibilities, a shift towards income generation, and gaps in digital-literacy skills mean that many girls will stop receiving an education, even where access to the Internet and distance-learning opportunities are available. In Ghana, for example, 16% of adolescent boys have digital skills compared to only 7% of girls.
Online violence increases self-censorship
Online GBV has proven an especially powerful tool for undermining women and women-identifying human-rights defenders, civil society leaders and journalists, leading to self-censorship, weakening women’s political leadership and engagement, and restraining women’s self-expression and innovation. According to UNESCO, 73% of women have been or will be exposed to some form of cyber violence in their lifetimes, and 52% of women feel the internet is not a safe place to express their opinions. If these trends are not addressed, closing the digital divide will never be possible, as many women who do get online will be pushed off because of the threats they face there. Women journalists, activists, politicians, and other female public figures receive the brunt of online harassment, through threats of sexual violence and other intimidation tactics. Online violence against journalists leads to journalistic self-censorship, affecting the quality of the information environment and democratic debate.
Solutions include education (training women to use computers and men and boys on how not to behave in online environments), policy change (advocating for the adoption of policies that address online harassment and protect women’s rights online), and technology change (encouraging more women to be involved in the creation of tech will help ensure that the tools and software that are available serve their needs, as women).
In 2019, UNESCO released “I’d blush if I could”, a research paper whose title was based on the response given by Siri when a human user addressed “her” in an extremely offensive manner. The paper noted that although the system was updated in April 2019 to reply to the insult more flatly (“I don’t know how to respond to that”), “the assistant’s submissiveness in the face of gender abuse remain[ed] unchanged since the technology’s wide release in 2011.” UNESCO suggested that by rendering the voices as women-sounding by default, tech companies were preconditioning users to rely on antiquated and harmful perceptions of women as subservient and failed to build in proper safeguards against abusive, gendered language.
Further, machine-learning systems rely on data that reflect larger gender biases. A group of researchers from Microsoft Research and Boston University trained a machine learning algorithm on Google News articles, and then asked it to complete the analogy: “Man is to Computer Programmer as Woman is to X.” The answer was “Homemaker,” reflecting the stereotyped portrayal and the deficit of women’s authoritative voices in the news. (Read more about bias in artificial intelligence systems in the Artificial Intelligence and Machine Learning Primer section on Bias in AI and ML).
In addition to expanding job opportunities, higher participation of women in tech leadership and development can help add a gender lens to the field and enhance the ways in which new technologies can be used to improve women’s lives.
New technologies allow for the increased surveillance of women
Research conducted by Privacy International shows that there is a uniqueness to the surveillance faced by women and gender non-conforming individuals. From data privacy implications related to menstrual-tracker apps, which might collect data without appropriate informed consent, to the ability of women to privately access information about sexual and reproductive health online, to stalkerware and GPS trackers installed on smartphones and internet of things (IoT) devices by intimate partners, pervasive technology use has exacerbated privacy concerns and the surveillance of women.
Research conducted by the CitizenLab, for example, highlights the alarming breadth of commercial software that exists for the explicit purpose of covertly tracking another mobile device activities, remotely and in real-time. This could include monitoring someone’s text messages, call logs, browser history, personal calendars, email accounts and/or photos.
Increased technological unemployment
Job losses caused by the replacement of human labor with automated systems lead to “technological unemployment,” which disproportionately affects women, the poor and other vulnerable groups, unless they are re-skilled and provided with adequate protections. Automation also requires skilled labor that can operate, oversee and/or maintain automated systems, eventually creating jobs for a smaller section of the population. But the immediate impact of this transformation of work can be harmful for people and communities without social safety nets or opportunities for finding other work.
Consider these questions when developing or assessing a project proposal that works with women or girls:
Have women been involved in the design of your project?
Have you considered the gendered impacts and unintended consequences of adopting a particular technology in your work?
How are differences in access and use of technology likely to affect the outcomes of your project?
Are you employing in your project technologies that could reinforce harmful gender stereotypes or fail the needs of women participants?
Are women and/or non-binary people exposed to additional safety concerns brought about by the use of the tools and technologies adopted in your project?
Have you considered gaps in sex- or gender-disaggregated data in the dataset used to inform the design and implementation of your project? How could these gaps be bridged through additional primary or secondary research?
How can your project meaningfully engage men and boys to address the gender digital divide?
How can your organization’s work help mitigate and eventually close the gender digital divide?
There are many examples of programs that are engaging with women to have a positive effect on the digital gender divide. Find out more about a few of these below.
USAID’s WomenConnect Challenge
In 2018, USAID launched the WomenConnect Challenge to enable women’s access to, and use of, digital technologies. The first call for solutions brought in more than 530 ideas from 89 countries, and USAID selected nine organizations to receive $100,000 awards. In the Republic of Mozambique, the development-finance institution GAPI is lowering barriers to women’s mobile access by providing offline Internet browsing, rent-to-own options, and tailored training in micro-entrepreneurship for women by region. Another awardee, AFCHIX, creates opportunities for rural women in Kenya, Namibia, and Sénégal and Morocco to become network engineers and build their own community networks or Internet services. The entrepreneurial and empowerment program helps women establish their own companies, provides important community services, and positions these individuals as role models.
Safe Sisters – Empowering women to take on digital security
In 2017, Internews and DefendDefenders piloted the Safe Sisters program in East Africa to empower women to protect themselves against online GBV. Safe Sisters is a digital-safety training-of-trainers program that provides female human rights defenders and journalists who are new to digital safety with techniques and tools to navigate online spaces safely, assume informed risks, and take control of their lives in an increasingly digital world. Through the program, which was created and run entirely by women, for women, participants learn digital-security skills and get hands-on experience by training their own at-risk communities.
In building the Safe Sisters model, Internews has verified that women will dive into improving their understanding of digital safety and share this information in their communities and get new job opportunities—if given the chance. Women can also create context- and language-specific digital-safety resources and will fight for policies that protect their rights online and deter abuse. There is strong evidence of the lasting impact of the Safe Sisters program: two years after the pilot cohort of 13 women, 80% are actively involved in digital safety; 10 have earned new professional opportunities because of their participation; and four have changed careers to pursue digital security professionally.
Internet Saathi
In 2015, Google India and Tata Trusts launched Internet Saathi, a program designed to equip women (known as Internet Saathis) in villages across the country with basic Internet skills and provide them with Internet-enabled devices. The Saathis then train other women in digital literacy skills, based on ‘train the trainer’ model. As of April 2019, there were more than 81,500 Internet Saathis who helped over 28 million women learn about the Internet across 289,000 villages. Read more about the Saathis here.
*A NOTE ON GENDER TERMINOLOGY All references to “women” (except those that reference specific external studies or surveys, which has been set by those respective authors) is gender-inclusive of girls, women, or any person or persons identifying as a woman.
Social media provide spaces for people and organizations to share and access news and information, communicate with beneficiaries and advocate for change. Social-media content includes the text, photos, videos, infographics, or any other material placed on a blog, Facebook page, Twitter account, etc. for the audience to consume, interact with, and circulate. Content is curated by platforms and delivered to users according to what is most likely to attract their attention. There is an ever-expanding amount of content available on these platforms.
Digital inclusion center in the Peruvian Amazon. For NGOs, social media platforms can be useful to reach new audiences and to raise awareness of services. Photo credit: Jack Gordon for USAID / Digital Development Communications.
Theoretically, through social media, everyone has a way to speak out and reach audiences across the world, which can be empowering and bring people together. At the same time, much of what is shared on social media can be misleading, hateful, and dangerous , which theoretically imposes a level of responsibility by the owners of platforms to moderate content.
How does social media work?
Social media platforms are owned by private companies, with business models usually based on advertising and monetization of users’ data. This affects the way that content appears to users, and influences data-sharing practices. Moderating content on these social-media spaces brings its own challenges and complications because it requires balancing multiple fundamental freedoms. Understanding the content moderation practices and business models of the platforms is essential to reap the benefits while mitigating the risks of using social media.
Business Models
Most social-media platforms rely on advertising. Advertisers pay for engagement, such as clicks, likes and shares. Therefore, sensational and attention-grabbing content is more valuable. This motivates platforms to use automated-recommendation technology that relies on algorithmic decision-making to prioritize content likely to grab attention. The main strategy of “user-targeted amplification” shows users content that is most likely to interest them based on detailed data that are collected about them. See more in the Risk section under Data Monetization by social media companies and tailored information streams.
The Emergence of Programmatic Advertising
The transition of advertising to digital systems has dramatically altered the advertising business. In an analog world, advertising placements were predicated on aggregate demographics, collected by publishers and measurement firms. These measurements were rough, capable at best of tracking subscribers and household-level engagement. Advertisers hoped their ads would be seen by enough of their target demographic (for example, men between 18 and 35 with income at a certain level) to be worth their while. Even more challenging was tracking the efficacy of the ads. Systems for measuring if an ad resulted in a sale were limited largely to mail-in cards and special discount codes.
The emergence of digital systems changed all of that. Pioneered for the most part by Google and then supercharged by Facebook in the early years of the 21st century, a new promise emerged: “Place ads through our platform, and we can put the right ad in front of the right person at the right time. Not only that, but we can report back to you (advertiser) who saw the ad, if they clicked on it, and if that click led to a ‘conversion’ or a sale.”
But this promise has come with significant unintended consequences. The way that the platforms—and the massive ad tech industry that has rapidly emerged alongside them—deliver on this promise requires a level of data gathering, tracking and individual surveillance unprecedented in human history. The tracking of individual behaviors, preferences and habits powers the wildly profitable digital advertising industry, dominated by platforms that can control these data at scale.
Managing huge consumer data sets at the scale and speed required to deliver value to advertisers has come to mean a heavy dependence on algorithms to do the searching, sorting, tracking, placement and delivery of ads. This development of sophisticated algorithms led to the emergence of programmatic advertising, which is the placement of ads in real time on websites with no human intervention. Programmatic advertising made up roughly two thirds of the $237 billion global ad market in 2019.
The digitization of the advertising market, particularly the dominance of programmatic advertising, has resulted in a highly uneven playing field. The technology companies enter with a significant advantage: they built the new structures and set the terms of engagement. What began as a value add in the new digital space— “We will give advertisers efficiency and publishers new audience and revenue streams”—has evolved to disadvantage both.
One of the primary challenges is in how audience engagement is measured and tracked. The primary performance indicators in the digital world are views and clicks. As mentioned above (and well documented in the literature), an incentive structure based on views and clicks (engagement) tends to favor sensational and eye-catching content. In the race for engagement, misleading or false content, with dramatic headlines and incendiary claims, consistently wins out over more balanced news and information. See also the section on digital advertising in the disinformation resource.
Advertising motivated content
Platforms leverage tools like hashtags and search engine optimization (SEO) to rank and cluster content around certain topics. Unfortunately, automated content curation motivated by advertising does not tend to prioritize healthful, educational, or rigorous content. Instead, frivolous, distracting, potentially untrue or even harmful content tends to spread more widely: conspiracy theories, shocking or violent content and “click-bait” (misleading phrases designed to entice viewing). Many platforms have features of upvoting (“like” buttons) which, similar to hashtags and SEO, influence the algorithmic moderation and promote certain content to circulate more widely. These features together cause “virality,” one of the defining features of the social-media ecosystem: the tendency of an image, video, or piece of information to be circulated rapidly and widely.
In some cases, virality can spark political activism and raise awareness (like the #MeToo hashtag), but it can also amplify tragedies and spread inaccurate information (anti-vaccine information and other health rumors, etc.). Additionally, the business models of the platforms reward quantity over quality (number of “likes”, “followers”, and views), encouraging a growth logic that has led to the problem of information saturation or information overload, overwhelming users with seemingly infinite content. Indeed, design decisions like the “infinite scroll” intended to make our social media spaces ever larger and more entertaining have been associated with impulsive behaviors, increased distraction, attention-seeking behavior, lower self-esteem, etc.
Many digital advertising strategies raise risks regarding access to information, privacy, and discrimination, in part because of their pervasiveness and subtlety. Influencer marketing, for example, is the practice of sponsoring a social media influencer to promote or use a certain product by working it into their social-media content, while native advertising is the practice of embedding ads in or beside other non-paid content. Most consumers do not know what native advertising is and may not even know when they are being delivered ads.
It is not new for brands to strategically place their content. However, today there is much more advertising, and it is seamlessly integrated with other content. In addition, the design of platforms makes content from diverse sources—advertisers and news agencies, experts and amateurs—indistinguishable. Individuals’ right to information and basic guarantees of transparency are at stake if advertisements are placed on equal footing with desired content.
Content Moderation
Content moderation is at the heart of the service that social-media platforms provide: the hosting and curation of the content uploaded by their users. Content moderation is not just the review of content, but every design decision made by the platforms, from the Terms of Service and their Community Guidelines, to the algorithms they use to rank and order content, to the types of content they allow and encourage through design features (“like”, “follow”, “block”, “restrict”, etc.).
Content moderation is particularly challenging because of the issues it raises around freedom of expression. While it is necessary to address massive quantities of harmful content that circulate widely, educational, historic, or journalistic content is often censored by algorithmic moderation systems. In 2016, Facebook took down a post with a Pulitzer Prize-winning image of a naked 9-year-old girl fleeing a napalm bombing and suspended the account of the journalist who had posted it.
Though nations differ in their stances on freedom of speech, international human rights provide a framework for how to balance freedom of expression against other rights, and against protections for vulnerable groups. Still, content-moderation challenges increase as content itself evolves, for instance through increase of live streaming, ephemeral content, voice assistants, etc. Moderating internet memes is particularly challenging, for instance, because of their ambiguity and ever-changing nature; and yet meme culture is a central tool used by the far right to share ideology and glorify violence. Some communications manipulations are also intentionally difficult to detect, for example, “dog whistling” (sending coded messages to subgroups of the population) and “gaslighting” (psychological manipulation to make people doubt their own knowledge or judgement).
Automated moderation
Content moderation is usually performed by a mix of humans and artificial intelligence , with the precise mix dependent on the platform and the category of content. The largest platforms like Facebook and YouTube use automated tools to filter content as it is uploaded. Facebook, for example, claims it is able to detect up to 80% of hate speech content in some languages as it is posted, then submitting it for human review. Though the working conditions for the human moderators have been heavily critiqued, the accuracy and transparency of the algorithms are also disputed and, expectedly, have some concerning biases. Humans are of course subject to biases as well, but algorithmic bias in content moderation poses more serious threats around equity and freedom of expression.
The complexity of content-moderation decisions does not lend itself easily to automation, and the porosity between legal and illegal/permissible and impermissible content leads to both cases of legitimate content being censored, and cases of harmful and illegal content passing through the filter (cyberbullying, defamation, etc.).
The moderation of content posted to social media has been increasingly important during the COVID-19 pandemic, when access to misleading and inaccurate information about the virus can result in death. The current moderation strategy of Facebook, for example, has been described as creating “a platform that is effectively at war with itself: the News Feed algorithm relentlessly promotes irresistible click-bait about Bill Gates, vaccines, and hydroxychloroquine; the trust and safety team then dutifully counters it with bolded, underlined doses of reality.”
Addressing harmful content
In some countries, local laws may address content moderation, but they relate mainly to child abuse images or illegal content that incites violence. Most platforms also have community standards or safety and security policies that state the kind of content allowed, and that sets the rules for harmful content. Enforcement of legal requirements and the platforms’ own standards relies primarily on content being flagged by social media users. The social-media platforms are only responsible for harmful content shared on their platforms once it has been reported to them.
Some platforms have established mechanisms that allow civil society organizations (CSOs) to contribute to the flagging process by becoming a so-called trusted flagger. With Facebook, this allows verification of accounts for civic organizations, provides higher levels of protection, faster response for incident reports, and accounts are less easily automatically disabled. For example, Access Now’s Digital Security Helpline is a trusted partner, and Facebook also offers access to its Trusted Partners Program to partners of members of the Design 4 Democracy Coalition. However, this program does not compensate for the limited accessibility to the platform for CSOs that encounter problems.
How is social media relevant in civic space and for democracy?
The flow of information on social media involves many fundamental human rights and supports the functioning of a democracy, to the extent that it allows freedom of expression, democratic debate, and civic participation. Social-media platforms have become core communication channels for CSOs. As more aspects of our lives take place within digital environments, social media becomes critical to aspects as fundamental as access to education, work and livelihood, health, and other services.
For example, citizen journalism, which has flourished through social media, has allowed internet users across the world to supplement the mainstream media with facts and perspectives ‘from the ground’ that might otherwise be overlooked or misrepresented. In some contexts, civic space actors rely on social-media platforms to produce and disseminate critical information during humanitarian crises or emergencies.
Digital inclusion center in the Peruvian Amazon. The business models and content moderation practices of social media platforms directly affect the content displayed to users. Photo Credit: Chandy Mao, Development Innovations.
However, the information shared over social media is mediated by private companies and governments, who possess new tactics for censorship, control and information manipulation . Censorship is no longer necessarily the denial of information but can be the denial of attention or credibility. Further, the porosity of online and offline space can be dangerous for individuals and for democracy, as harassment, hate speech, and “trolling” behaviors offer new methods for violence, including organized violence. Doxxing and targeted digital attacks have also been used to intimidate journalists and political minorities or opponents. Read more about online violence and targeted digital attacks in the Risks section .
Social-media content is also pervasive and harmonized—in our personalized news feeds, information shared by amateurs, advertisers, or for political objectives can be difficult to distinguish from quality news, giving rise to a range of information disorders, from the accidental forwarding of inaccurate information to the intentional sharing of harmful content, as explored in the disinformation primer .
The platforms’ algorithms are designed to reward quantity over quality (number of “likes,” “followers,” and views), which has led to the problem of information saturation or information overload, overwhelming users with seemingly infinite content. Indeed, design decisions like the “infinite scroll,” which intended to make our social-media spaces ever larger and more entertaining, have been associated with impulsive behaviors, increased distraction, attention-seeking behavior, lower self-esteem, etc.
Furthermore, social-media platforms have become gatekeepers of information and connections. It has become harder to work and live without these platforms: those not using social media may miss important public announcements, events, community information or even family updates.
Students from the Kandal Province, Cambodia. Social media platforms have opened up new platforms for video storytelling. Photo credit: Chandy Mao, Development Innovations.
Social media can have positive impacts when used to further democracy, human rights and governance issues. Read below to learn how to more effectively and safely think about social media use in your work.
Citizen Journalism
Social media has been credited with providing channels for citizens, activists, and experts to report instantly and directly—from disaster settings, during protests, from within local communities, etc. Citizen journalism, also referred to as participatory journalism or guerrilla journalism, does not have a definite set of principles and should not be considered as a replacement for professional journalism, but it is an important supplement to mainstream journalism. Collaborative journalism, the partnership between citizen and professional journalists, as well as crowdsourcing strategies, are further techniques permitted by social media that have enhanced journalism, helping to promote voices from the ground and to magnify diverse voices and viewpoints. The outlet France 24 has developed a network of 5,000 contributors, the “observateurs,” who are able to cover important events directly by virtue of being on scene at the time, as well as to confirm the accuracy of information.
Social-media platforms as well as blogging tools have allowed for the decentralization of expertise, bridging elite and non-elite forms of knowledge. Without proper fact-checking or supplementary sources and proper context, citizen reporting carries risks— including security risks to the authors themselves—but it is an important democratizing force and source of information.
Crowdsourcing
In crowdsourcing, the public is mobilized to share data together to tell a larger story or accomplish a greater goal. Crowdsourcing can be a method for financing, for journalism/reporting, or simply for gathering ideas. Usually some kind of software tool or platform is put in place that the public can easily access and contribute to. Crisis mapping, for example, is a type of crowdsourcing through which the public shares data in real time during a crisis (a natural disaster, an election, a protest, etc.). These data are then ordered and displayed in a useful way. For instance, crisis mapping can be used in the wake of an earthquake to show first responders the areas that have been hit and need immediate assistance. Ushahidi is an open-source crisis-mapping software developed in Kenya after the violent outbreak following the election in 2007. The tool was first created to allow Kenyans to flag incidents, to form a complete and accurate picture of the situation on the ground, to share with the media, outside governments, and relevant civil society and relief organizations. In Kenya, the tool gathered texts, tweets, and photos and created crowdsourced maps of incidents of violence, election fraud, and other abuse. Ushahidi now has a global team and works in 30 different languages.
Digital Activism
Social media has allowed local and global movements to spring up overnight, inviting broad participation and visibility. Twitter hashtags in particular have been instrumental for coalition building, coordination, and for raising awareness among international audiences, media and government. Researchers began to take note of digital activism around the 2011 “Arab Spring,” when movements in Tunisia, Morocco, Syria, Libya, Egypt and Bahrain, among others countries, leveraged social media and were quickly followed by the Occupy Wallstreet movement in the United States. Ukranian’s Euromaidan movement in late 2013 and the Hong Kong protests in 2019 are also examples of political movements that used social media to galvanize support.
In 2013, the acquittal of George Zimmerman in the death of unarmed 17-year-old Trayvon Martin inspired the creation of the #BlackLivesMatter hashtag. This movement grew stronger in response to the tragic killing of Michael Brown. The hashtag, at the front of an organized national protest movement, provided an outlet for people to join an online conversation and articulate alternative narratives in real time about subjects that the media and the rest of the United States (and more recently with the killing of George Floyd, the world) had not paid sufficient attention to: police brutality, systemic racism, racial profiling, inequality, etc.
The #MeToo movement against sexual misconduct in the media industry, which also became a global movement, has allowed a multitude of people to participate in activism previously bound to a certain time and place.
Some researchers and activists fear “slacktivism” and the effect of social media giving people an excuse to stay at home rather than make a more dynamic response, and some fear that the tools of social media are ultimately insufficient for enacting meaningful social change, which requires nuanced political arguments. (Interestingly, a 2018 Pew Research survey on attitudes toward digital activism showed that just 39% of white Americans believed social media was an important tool to use to express themselves, while 54% percent of Black people said that it was an important tool for them.)
Social media has enabled new online groups to gather together and to express a common sentiment as a form of solidarity or as a means to protest. Especially since the COVID-19 pandemic broke out, many physical protests have been suspended or cancelled, and virtual protests proceeded.
Expansion and engagement with international audience at low costs
Social media provides a valuable opportunity for CSOs to reach their goals and engage with existing and new audiences. A good social-media strategy is best underpinned by a permanent staff position to grow a strong and consistent social-media presence based on the organization’s purpose, values, and culture. This person should know how to seek information, be aware of both the risks and benefits of sharing information online and understand the importance of using sound judgment when posting. The USAID ‘Social Networking: A Guide to Strengthening Civil Society through Social Media’ provides a set of questions as guidance to develop a sound social-media policy, asking organizations to think about values, roles, content, tone, controversy and privacy.
Increased awareness of services
Social media can be integrated into programmatic activities to strengthen the reach and impact of the program, for example, by generating awareness of an organization’s services to a new demographic. Organizations can promote their programs and services while responding to questions and fostering open dialogue. Widely used social media platforms can be useful to reach new audiences for training and consulting activities through webinars or individual meetings designed for NGOs.
Opportunities for Philanthropy and Fundraising
Social-media fundraising presents an important opportunity for non-profits, but organizations should carefully consider the type of campaign and platforms they choose. TechSoup, a non-profit providing tech support for NGOs, offers advice and an online course on fundraising with social media for non-profits.
After the blast in Beirut’s harbor in the summer of 2020, many Lebanese people started online fundraising pages for their organizations. Social-media platforms were used extensively to share funding suggestions to the global audience watching the disaster unfold, reinforced by traditional media coverage.
Emergency communication
In some contexts, civic actors rely on social media platforms to produce and disseminate critical information, for example, during humanitarian crises or emergencies. Even in a widespread disaster, the internet often remains a significant communication channel, which makes social media a useful complementary means for emergency teams and the public. Reliance on the internet, however, increases vulnerability in case of network shutdowns.
In Kyiv, Ukrainian students share pictures at the opening ceremony of a Parliamentary Education Center. Photo credit: Press Service of the Verkhovna Rada of Ukraine, Andrii Nesterenko.
The use of emerging technologies can also create risks in civil society programming. Read below on how to discern the possible dangers associated with social media platforms in DRG work, as well as how to mitigate for unintended – and intended – consequences.
Polarization and Ideological Segregation
The ways in which content flows and is presented in social media due to the platforms’ business models risk limiting our access to information, particularly to information that challenges our preexisting beliefs, by exposing us to content likely to attract our attention and support our views. The concept of the filter bubble refers to the filtering of information by online platforms and through our own intellectual biases that worsen polarization by allowing us to live in echo chambers. This is easily witnessed in a YouTube feed: when you search for a song by an artist, you will likely find more songs by the same artist, or similar ones—the algorithms are designed to prolong your viewing, and assume you want more of something similar. The same has been identified for political content. Social media algorithms encourage confirmation bias, exposing us to content we will agree with and enjoy, often at the expense of the accuracy, rigor, educational or social value of that content.
The massive and precise data amassed by advertisers and social media companies about our preferences and opinions permit the practice of micro-targeting, which involves the display of tailored content based on data about users’ online behaviors, connections, and demographics, among others, as will be further explained below.
The increasingly tailored distribution of news and information is a threat to political discourse, diversity of opinions and democracy. Users can become detached even from factual information that disagrees with their viewpoints and isolated within their own cultural or ideological bubbles.
Because tailoring news and other information on social media is driven largely by nontransparent, opaque algorithms that are owned by private companies, it is hard for users to avoid these bubbles. Access to and intake of the very diverse information available on social media, with its many viewpoints, perspectives, ideas and opinions, requires an explicit effort by the individual user to go beyond passive consumption of the content that is presented to them.
Misinformation and Disinformation
The internet and the dominant online platforms provide new tools that amplify and alter the danger presented by false, inaccurate or taken out of context information. The online space increasingly drives discourse and is where much of today’s disinformation takes root. Refer to the Disinformation resource for a detailed overview of these problems.
Online Violence and Targeted Digital Attacks
Social media facilitates a number of violent behaviors such as defamation, harassment, bullying, stalking, “trolling” and “doxxing.” Cyberbullying among children, like traditional offline bullying, can harm students’ performance in school and causes real psychological damage. Cyberbullying is particularly harmful because victims experience the violence alone, isolated in cyberspace. They often do not seek help from parents and teachers, who they believe are not able to intervene. Cyberbullying is also difficult to address because it can move across social-media platforms, beginning on one and moving to another. Like cyberbullying, cyber harassment and cyberstalking have very tangible offline effects. Women are most often the victims of cyber harassment and cyberviolence, sometimes through the use of stalkerware installed by their partners to track their movements. A frightening cyber harassment trend has accelerated in France during the COVID-19 confinement in the form of “fisha” accounts, where bullies, aggressors or jilted ex-boyfriends will publish and circulate naked photos of teenage girls without their consent.
Journalists, women in particular, are often subject to cyber harassment and threats, particularly those who write about socially sensitive or political topics. Online violence against journalists can lead to journalistic self-censorship, affecting the quality of the information environment and democratic debate. Online tools provide new ways to spread and amplify hate speech and harassment. The use of fake accounts, bots, and even bot-nets (automated networks of accounts) allow perpetrators to attack, overwhelm, and even disable the social media accounts of their victims. Doxxing, by revealing sensitive information about journalists, is another strategy that can be used for censorship.
The 2014 case of Gamergate, when several women video-game developers were attacked by a coordinated harassment campaign that included doxxing and threats of rape and death, illustrates the strength and capacity of loosely-connected hate groups online to rally together, inflict real violence, and even drown out criticism. Many of the actions of the most active Gamergate trolls are illegal, but their identities are unknown. Importantly, it has been suggested by supporters of Gamergate that the most violent trolls were a “smaller, but vocal minority”—evidence of the magnifying power of internet channels and their use for coordinated online harassment.
Online hoaxes, scams, and frauds, like in their traditional offline forms, usually aim to extract money or sensitive information from the target. The practice of phishing is increasingly common on social media: an attacker pretends to be a contact or a reputable source in order to send malware or to extract personal information or account credentials. Spearphishing is a targeted phishing attack that leverages information about the recipient and details related to the surrounding circumstances. Coronavirus-related spearphishing attacks have increased steadily during the pandemic.
Data monetization by social media companies and tailored information streams
Most social-media platforms are free to use. Users simply register and then are able to use the platform. Social-media platforms do not receive revenue directly from users, like in a traditional subscription service; rather they generate profit primarily through digital advertising. Digital advertising is based on the collection of users’ data by social-media companies, which allows advertisers to target their ads to specific users and types of users. Social-media platforms monitor their users and build detailed profiles that they sell to advertisers. The data tracked includes information about the user’s connections and behavior on the platform, such as friends, posts, likes, searches, clicks and mouse movements. Data are also extensively collected outside platforms, including information about users’ location, webpages visited, online shopping and banking behavior. Additionally, many companies regularly access the contact book and photos of their users.
In the case of Facebook, this has led to a long-held and widespread conspiracy theory that the company listens to conversations to serve tailored advertisements. No one has ever been able to find clear evidence that this is actually happening. Research has shown that a company like Facebook does not need to listen in to your conversations, because it has the capacity to track you in so many other ways: “Not only does the system know exactly where you are at every moment, it knows who your friends are, what they are interested in, and who you are spending time with. It can track you across all your devices, log call and text metadata on phones, and even watch you write something that you end up deleting and never actually send.”
The massive and precise data amassed by advertisers and social-media companies about our preferences and opinions permit the practice of micro-targeting, that is, displaying targeted advertisements based on what you have recently purchased, searched for or liked. But just as online advertisers can target us with products, political parties can target us with more relevant or personalized messaging. Studies are currently trying to determine the extent to which political micro-targeting is a serious concern for the functioning of democratic elections. The question has also been raised by researchers and digital rights activists as to how micro-targeting may be interfering with our freedom of thought.
The increasingly tailored distribution of news and information is a threat to political discourse, diversity of opinions and democracy. Users can become detached from information that disagrees with their viewpoints and isolated within their own cultural or ideological bubbles. This is easily witnessed in a YouTube feed: when you search for a song by an artist, you will likely be shown more by the same artist, or similar ones—the algorithms are designed to prolong your viewing, and assume you want more of something similar. The same has been identified for ideological content.
Government surveillance and access to personal data
The content shared over social media is monitored by governments, who use social media for censorship, control and information manipulation. Many democratic governments are known to engage in extensive social-media monitoring for law enforcement and intelligence-gathering purposes. These practices should be guided by robust legal frameworks to safeguard individuals’ rights online, such as privacy and data-protection laws , but many countries have not yet enacted these types of laws.
There are also many examples of authoritarian governments using personal and other data harvested through social media to intimidate activists, silence opposition, and bring development projects to a halt. The information shared on social media often allows bad actors to build extensive profiles of individuals that enable targeted online and offline attacks, often through social engineering techniques. For example, a phishing email can be carefully crafted based on social-media data to trick an activist to click on a malicious link that provides access to the device, documents or social-media accounts.
Sometimes, however, a strong, real-time presence on social media can protect a prominent activist against threats by the government. A disappearance or arrest will be immediately noticed by followers or friends of a person who suddenly becomes silent on social media.
Market power and differing regulation
We rely on social-media platforms to help fulfill our fundamental rights (freedom of expression, assembly, etc.). However, these platforms are massive global monopolies and have been referred to as “the new governors.” This market concentration is troubling to national and international governance mechanisms. Simply breaking up the biggest platform companies will not fully solve the information disorders and social problems presented by social media. Civil society and governments also need visibility of the design choices made by the platforms to understand how to address their negative aspects.
The growing influence of social-media platforms has given many governments reasons to impose laws on online content. There is a surge in laws across the world regulating illegal and harmful content, such as incitement to terrorism or violence, false information, and hate speech. These laws often criminalize speech and contain punishments of jail terms or high fines for something like a retweet on Twitter. Even in countries where the rule of law is respected, legal approaches to regulating online content may be ineffective due to the many technical challenges of content moderation. There is also a risk of violating internet users’ freedom of expression by reinforcing imperfect and non-transparent moderation practices and over-deletion. Lastly, they constitute a challenge to social media companies to navigate between compliance with local laws and defending international human rights law.
Amplification/virality
As noted above, virality is one of the defining features of the social-media ecosystem: the tendency of an image, video, or piece of information to circulate rapidly and widely. In some cases, virality can spark political activism and raise awareness (like the #MeToo hashtag), but it can also amplify tragedies (the video of the Christchurch massacre in New Zealand) and spread inaccurate information.
Impact on journalism
Social media has had a profound impact on the field of journalism. While it has allowed the emergence of the citizen-journalist, locally-reported and crowd-sourced information, social-media companies have displaced the relationship between advertising and the traditional newspaper, and created a rewards system for sensationalist, click-bait content that will attract the widest attention globally, over quality journalism that may be pertinent to local communities. In many places, these monopolies have also been partially responsible for the collapse of local news.
The reason for this impact is that, while advertising has successfully made the transition to digital, with global revenues currently at $247 billion and growing by 4% year over year, very little of that revenue is making its way to publishers. The ad tech supply chain, dominated by Google and Facebook, now consumes 90% of all new growth in the world’s major markets and 61 cents of every dollar spent on digital advertising worldwide.
The disruption of the publishing business model has been a slow-motion disaster for news organizations around the world. Since 2012, digital newspaper ad revenues worldwide have grown from an anemic $7.3 billion to $9.95 billion in 2016, while the Google/Facebook duopoly will earn $174 billion or 61% of the global digital advertising market.
In addition, the way search tools work dramatically affects local publishers, as search is a powerful vector for news and information. Researchers have found that search rankings have a marked impact on our attention. Not only do we tend to think information that is ranked more highly is more trusted and relevant, but we tend to click on top results more often than lower ones. The Google search engine concentrates our attention on a narrow range of news sources, a trend that works against diverse and pluralistic media outlets. It also tends to work against the advertising revenue of smaller and community publishers, which is based on user attention and traffic. It is a downward spiral: search results favor larger outlets, those results drive more user engagement which makes their inventory more valuable in the advertising market, those publishers grow larger driving more favorable search results and onward we go.
To understand the implications of social media information flows and choice of platforms used in your work, ask yourself these questions:
Does your organization have a social-media strategy? What does your organization hope to achieve through social media use?
Do you have staff who can oversee and ethically moderate your social-media accounts and content?
Which platform do you intend to use to accomplish your organization’s goals? What is the business model of that platform? How does this business model affect you as a user?
How is content ordered and moderated on the platform used (humans, volunteers, AI, etc.)? Can content go viral?
Where is the platform legally headquartered? What jurisdiction and legal frameworks does it fall under?
Do the platforms chosen have mechanisms for users to signal harassment and hate speech for review and possible removal?
Do the platforms have mechanisms for users to be heard when content is unfairly taken down or accounts unfairly blocked?
Are the platforms collecting data about users? Who else has access to collected data and how is it being used?
How does the platform involve their community of users and civil society (for instance, in flagging dangerous content, in giving feedback on design features, in fact-checking information, etc.)? Are there local representatives?
Do the platforms chosen have privacy features like encryption? If so, what level of encryption and for what precise services (for example, only on the app, only in private message threads)? What are the default settings?
“Crowdsourced mapping has become an integral part of humanitarian response, with high profile deployments of platforms following the Haiti and Nepal earthquakes, and the multiple projects initiated during the Ebola outbreak in North West Africa in 2014, being prominent examples. There have also been hundreds of deployments of crowdsourced mapping projects across the globe that did not have a high profile. This paper, through an analysis of 51 mapping deployments between 2010 and 2016, complimented with expert interviews, seeks to explore the organisational structures that create the conditions for effective mapping actions, and the relationship between the commissioning body, often a non-governmental organisation (NGO) and the volunteers who regularly make up the team charged with producing the map.”
Scaling Social Movements Through Social Media: The Case of Black Lives Matter
Marcia Mundt, Karen Ross , and Charla M Burnett, Social Media and Society, 2018
“Drawing on a case study of Black Lives Matter (BLM) that includes both analysis of public social media accounts and interviews with BLM groups, [the authors] highlight possibilities created by social media for building connections, mobilizing participants and tangible resources, coalition building, and amplifying alternative narratives. [They] also discuss challenges and risks associated with using social media as a platform for scaling up. Our analysis suggests that while benefits of social media use outweigh its risks, careful management of online media platforms is necessary to mitigate concrete, physical risks that social media can create for activists.”
Facebook’s Role in the Genocide in Myanmar: New Reporting Complicates the Narrative
“Members of the Myanmar military have systematically used Facebook as a tool in the government’s campaign of ethnic cleansing against Myanmar’s Rohingya Muslim minority, according to an incredible piece of reporting by the New York Times on Oct. 15. The Times writes that the military harnessed Facebook over a period of years to disseminate hate propaganda, false news and inflammatory posts. The story adds to the horrors known about the ongoing violence in Myanmar, but it also should complicate the ongoing debate about Facebook’s role and responsibility for spreading hate and exacerbating conflict in Myanmar and other developing countries…”
How the Chinese Government Fabricates Social Media Posts for Strategic Distraction, not Engaged Argument
“The Chinese government has long been suspected of hiring as many as 2,000,000 people to surreptitiously insert huge numbers of pseudonymous and other deceptive writings into the stream of real social media posts, as if they were the genuine opinions of ordinary people. Many academics, and most journalists and activists, claim that these so-called “50c party” posts vociferously argue for the government’s side in political and policy debates. As we show, this is also true of the vast majority of posts openly accused on social media of being 50c. Yet, almost no systematic empirical evidence exists for this claim, or, more importantly, for the Chinese regime’s strategic objective in pursuing this activity. In the first large scale empirical analysis of this operation, we show how to identify the secretive authors of these posts, the posts written by them, and their content. We estimate that the government fabricates and posts about 448 million social media comments a year. In contrast to prior claims, we show that the Chinese regime’s strategy is to avoid arguing with skeptics of the party and the government, and to not even discuss controversial issues. We show that the goal of this massive secretive operation is instead to distract the public and change the subject, as most of these posts involve cheerleading for China, the revolutionary history of the Communist Party, or other symbols of the regime. We discuss how these results fit with what is known about the Chinese censorship program and suggest how they may change our broader theoretical understanding of “common knowledge” and information control in authoritarian regimes.”
Environmental campaigning: Earth Hour
The World Wide Fund For Nature (WWF) launched Earth Hour in 2007. This social media campaign calls for everyone – individuals and businesses alike – to switch off their lights for one hour. In 2017, the campaign’s 10th anniversary, millions of people and thousands of landmarks around the world turned their lights off for a single hour. The WWF uses the #EarthHour hashtag (amongst others) to galvanize its followers. Elements of this successful campaign are the limited time frame which makes engagement actionable, the push to share individual actions on multiple platforms, the enticing language and countdown timers.
The Cyber Harassment Helpline
The Cyber Harassment Helpline is an accessible toll-free Helpline for victims and survivors of online harassment and violence in Pakistan. Digital Rights Foundation founded this helpline in 2016 in response to the increasing examples of harassment of social media users and in particular women. The helpline focuses in particular on marginalized groups in Pakistan and prefers not to communicate via social media platforms for reasons of privacy and confidentiality.
BellingCat: An independent international collective of researchers, investigators and citizen journalists using open source and social media investigation.
Documentary “The Social Dilemma.” Preview available here.
Graphika: an investigative research company that leverages AI to study online communities, analyzing how online social networks form, evolve, and are manipulated.