A look at AI filters, bias, exploitation and what Africans can do about it
By Rebecca Nanono, Contributor
Introduction
Artificial intelligence systems ,from generative text to
face filters on apps , are only as smart as the data they are trained on. That
means the billions of images, videos, recordings, messages, and internet
activity that exist online influence how AI understands the world. And
increasingly, African digital content , especially from young, creative
users is being sucked into global AI models.
This is not just a technical issue but a digital rights
and power issue.
In this blog, we explore the following.
- How
African data fuels AI
- Why
African women’s images and voices often appear in AI systems
- The
risks of this dynamic
- What
communities and policymakers can do
The Data Behind the AI Curtain
Modern AI systems, like those powering TikTok’s face filters or global large language models (LLMs), rely on large datasets drawn from the internet. These datasets come from the following.
- Social
media posts
- Videos
and images
- Text
scraped from public websites
- Publicly
available videos, comments, metadata
AI companies often build models by scraping public content without
explicit individual consent, relying instead on “publicly available” status
as justification. While this is common across the tech industry, it raises
deep ethical questions about African data sovereignty and consent. CIPIT
Why African Data Gets Picked Up (and Used by AI)
African users, particularly on apps like TikTok,
create huge volumes of engaging visual content , dancing, storytelling, beauty
content, comedy, and more. This makes African digital footprints rich fodder
for AI models because:
- High
engagement = data visibility
Algorithms amplify content that gets likes, comments, and shares ,meaning more African content gets indexed and becomes part of what AI systems see. - Mobile-first
and video-first behaviour
Many Africans access the internet primarily through mobile video platforms, generating images and videos every day ,a gold mine for image-based AI training. - Lack
of robust consent regimes
In many countries, data protection laws are weak or unevenly enforced, allowing companies to collect and move data across borders with little oversight or accountability. African Commission on Human Rights
As a result, a lot of African user data , including
images and videos of African women , finds its way into global AI training
pipelines, often without people realising it.
TikTok and AI Filters: A Case in Point
TikTok’s AI systems analyse all user-generated videos to do
the following.
- Personalise
content recommendations
- Create
and refine filters and effects
- Understand
user preferences
Critics have noted that TikTok’s AI can reinforce harmful
social biases by feeding users more of what the algorithm thinks they want
, including beauty standards and aesthetic filters that may not reflect
African diversity. AIAA IC
While there’s no public evidence that social media or other
tech companies single out African women specifically for AI training, the
very nature of algorithmic learning reinforces majority trends from what it
sees online, for better or worse.
This can lead to the following.
- Algorithmic
bias in facial filters
- Reinforcement
of narrow beauty standards
- Reduced
visibility for minority creators
- Harm
to self-esteem, especially among young users whose images are repeatedly
processed and showcased by AI. AIAA IC
The Problem of Representation and Power
Oddly, while African content is widely available online
and thus ingested into models, much of global AI’s core training data still
under-represents African contexts in structured ways seen below.
- Major
AI systems rely heavily on data from Europe and North America; African
representation in high-quality training datasets is small. TechCabal
- Many
African languages, cultural narratives, and contextual nuances don’t
appear in the text corpora used to train LLMs (massive language models) ,
meaning the AI does not understand African contexts well. Carnegie Endowment
This creates a paradox.
African data,especially social content
is used to tune AI behaviours, but AI systems are not designed for
African realities.
Why This Matters
This dynamic has real implications seen below.
1. Lack of Consent & Data Sovereignty
People rarely know when their posts, photos, or videos are
ingested into datasets used to teach AI , especially outside strong
data-protection regimes. This raises classic concerns about autonomy over
personal information. African Commission on Human Rights
2. Bias and Misrepresentation
AI models trained on Western-centric data can
misinterpret or misrepresent African features, languages, and cultural contexts
, leading to algorithmic bias. Wikipedia
3. Digital Colonialism
According to the African Union and human rights experts,
this dynamic resembles data colonialism: foreign companies benefiting
from African digital content while Africans have limited say over how their
data is used. African Commission on Human Rights
4. Economic Inequity
Data created by Africans enriches foreign AI companies that
monetise AI outputs ,but African communities rarely see economic benefit or
control from this value creation.
What Can Be Done? (Solutions and Interventions)
1. Strengthen Data Protection Laws
African governments need robust, enforceable laws that
- Require
consent for data use in AI
- Restrict
cross-border transfer without safeguards
- Give
individuals control over how their data is used
Many countries are already working on or implementing such frameworks. African Commission on Human Rights
2. Data Sovereignty Initiatives
Countries and regional blocs can do the following.
- Store
and govern data within Africa
- Build
local AI datasets that reflect African languages and cultures
- Negotiate
fair terms with tech companies
This helps ensure AI serves local needs, not just global
profit. Carnegie Endowment
3. Educate Users About Digital Rights
According to digital rights organisations, users should
be sensitised about their data rights , what consent means and how content
might be used. African Commission on Human Rights
4. Support Ethical Data Stewardship
Community-led initiatives can do the following.
- Curate
open, consent-based datasets
- Provide
transparent governance
- Reward
content creators whose data adds value to AI
5. Advocate for Algorithmic Transparency
Tech companies should be urged , through policy and public
pressure, to disclose:
- What
data goes into their AI systems
- How
models treat data from underrepresented populations
Toward Fairer AI
AI is not inherently bad .It has the potential to transform
healthcare, education, commerce, and creativity across Africa. But fairness,
consent, and representation must be central.
If AI is going to learn from African people, then Africans should
- Understand
how their digital footprints contribute to that learning
- Have
legal safeguards protecting their data
- Be
equitably involved in shaping AI systems
The future of AI is not just about technology but
about whose stories and faces the technology learns from, and who benefits
from it.
Reading references for the blog
AI, Data Governance & African Policy
- African
Commission on Human and Peoples’ Rights AI Study
Study on human rights, AI and data governance in Africa, highlighting the need for data sovereignty and concerns about unrepresentative data.
https://achpr.au.int/sites/default/files/files/2025-04/draft-achpr-ai-study-march-2025.pdf African Commission on Human Rights - Pan-African
Parliament on Data Sovereignty & Ethical AI
Press release on Africa’s push for data sovereignty, ethical AI, and governance frameworks.
https://pap.au.int/en/news/press-releases/2025-07-25/pan-african-parliament-champions-africas-quest-data-sovereignty-and? Pan-African Parliament - African
Union — Africa Declares AI a Strategic Priority
AU Declaration emphasising the need to protect data ownership and ethical AI development for inclusive growth.
https://au.int/en/pressreleases/20250517/africa-declares-ai-strategic-priority-investment-inclusion-and-innovation African Union - Carnegie
Endowment — Understanding Africa’s AI Governance Landscape
Analysis of data ownership, lack of African datasets, and importance of localized AI models.
https://carnegieendowment.org/posts/2025/09/understanding-africas-ai-governance-landscape-insights-from-policy-practice-and-dialogue - CIPIT
— AI Governance in East Africa
Overview of AI governance efforts and ethical frameworks being developed across the region.
https://cipit.strathmore.edu/ai-governance-landscape-in-the-east-african-region/ - CIPIT
— The State of AI in Africa Report
Report mapping AI and data governance, including data sovereignty and indigenous knowledge systems.
https://revamp.cipit.org/the-state-of-ai-in-africa-report/
AI Bias, Ethics & Policy
- IAPP
— Data Protection Authorities and AI Regulation in Africa
Explains how data protection authorities (DPAs) are engaging with AI regulation and automated decision-making challenges.
https://iapp.org/news/a/dpas-and-ai-regulation-in-africa/ - UNESCO
— African Guidelines for Information Integrity
UNESCO article on consultations for guidelines to monitor platform accountability and data integrity support.
https://www.unesco.org/en/articles/consultations-launched-african-guidelines-ensuring-information-integrity-tech-platforms
Digital Rights & Data Protection Context
- CIPESA
— The Impact of Artificial Intelligence on Data Protection in Africa
Brief on AI’s risks to privacy, bias, misinformation, and recommendations on balancing innovation and privacy.
https://cipesa.org/2024/05/the-impact-of-artificial-intelligence-on-data-protection-and-privacy-in-africa/

Comments
Post a Comment