CRAFT 61: Big Data and AI in the Global South
Notes from the workshop to continue the conversation.
I took notes for myself during the CRAFT session on June 21st and have since realized that these may be more useful if shared broadly. Any mistakes are my own. I have removed contact information from here because I do not intend to share it without their permission. I’m sharing their names and their work because it is present on the website for the workshop.
High Level Thoughts
I was honored to be able to listen, learn, and engage with the keynote speakers and presenters. They brought forth excellent and diverse points. As one of a few white people who attended the CRAFT session I have to acknowledge two things:
- the timing of the event reinforces entrenched, asymmetrical power dynamics
- as a white man who works for a tech company, I am closer to the center of power than the presenters are–even as I feel powerless to affect change on any given day
The event was at a convenient time for me (12:30pm to 4pm PDT). It was inconvenient, even unreasonable, for almost all other attendees and the presenters who had to be up in the middle of the night for hours to participate. I am disappointed in the timing of this event and have voiced my concerns to the conference. I hope future conferences learn from this and find timing that is more convenient for the speakers. The irony is not lost on me that this session enacted the very power asymmetry it was critiquing.
The second point is something I am struggling to put into words more clearly. I exist on the periphery of power within every organization I belong to. I consistently advocate for others who are excluded to have a voice and challenge dominant power structures. This tends to make me seem like a threat to my managers and executives because I am not okay with just doing the work to make the company money. I believe I have to do it in a way that promotes human flourishing.
Reflecting on everything that was said during the workshop and the issues at play with internet quality, timing, etc. I realized how I may (or “must”, I’m unsure about this) come across to others. It was truly humbling and reinforces how much I consider myself peripheral to the work of bringing diverse, under appreciated voices towards the center.
The critical themes I’m taking from the workshop:
- The social construction, and usage, of data demands that we refrain from the impulse to universalize and focus on relational meaning
- Theory and practice both need to originate from the Global South to be truly beneficial to those communities
- Inequality, power asymmetry, and colonization are likely to grow with greater adoption of AI until we can achieve the first two points
- My own addition: I believe the field of AI needs to draw from the depth of thought in the humanities, acknowledging that they are also centered in the Global North, to avoid recreating the wheel and further eroding language with more jargon or siloing
Keynotes
Dr. Nimmi Rangaswamy
She gave a talk on centering human and cultural experience in our discussions of AI. She focused on some key points that I’ll list below. Sadly, her talk started a bit earlier than the scheduled time so I missed a portion.
Dr. Ragaswamy discussed:
- Legacy Systems
- Problems of data needing to be based in context
- Offshoring AI
- Inclusive AI
My key takeaway
We need to view data as socially constructed rather than as “being real” in itself. To me this means we need to challenge the concept of most conceptions of “ground truth,” because people often fail to critically reflect on what that means. They tend to generalize from the “ground truth” as if it represents some universal form of truth.
Dr. Nick Couldry
He gave a talk largely centered in his book, The Costs of Connection, and he talked about how we as practitioners need to recognize that data is a continuation of the colonial enterprise. It is not something new, or “capitalism gone wrong”, but rather it is an extension of hundreds of years of history using a new tool for control.
He directly argues against Shoshana Zuboff’s book, The Age of Surveillance Capitalism, considering his work to go beyond what she offers. I do not agree with his critiques on the surface (having recently read her work and not yet gotten to read his). I am going to capture a few points here and expand on them after I’ve read both works.
- Against Surveillance Capitalism
- He feels Zuboff’s work agrees with a lot of the points of Data Colonialism but doesn’t theorize colonialism on any level
- He points out that his work addresses two critical things not covered:
- Scope - time and sheer scale of what’s going on
- Depth - Forces us to think about how colonialism changed the world
- Other works recommended:
- “Your Privacy is Important to Us! Restoring Human Dignity in Data-Driven Marketing” Jan Trzaskowski
My Key Takeaway
I think I fundamentally disagree with Couldry’s critiques of Surveillance Capitalism. I do think that his work adds another layer to what she’s exploring. Both appear to offer deep insight into how data exists in the world and I think they may be more complementary rather than exclusive. Couldry’s perspective appears to look at it from a historical perspective, while I would say Zuboff is theorizing a new type of power (e.g. Instrumentarian Power) that the tech companies and governments are using to shape reality.
Presenters and Organizers
Deepa Singh
Her talk was about reimagining AI Ethics in India. She talked about focusing on situated ethics (within cultures and contexts) to ensure we’re not trying to create false universals. Her focus was on breaking away from binary distinctions because “our ethical and moral actions happen in a particular context” and what we see as ethical may change based on who we are, where we are, or what is happening around us.
Other key things she mentioned:
- Ontological Shift
- Knowledge attuned to resonance (rather than a universal truth)
- Rationality as a false universal
- What is a rational account of the world?
- Whose ideas and values are you carrying forward?
- Construction and upholding of these values are inherently acts of power
- Centers of Hegemonic Power - Global North
My Key Takeaway
As an existentialist I find that Deepa’s perspective resonates with what I differentiate as “T” Truth and “t” truth, which focuses on how we exist and what we can know. Her words provided a powerful reason to push us towards a less externalized understanding of truth–there is no “objective” truth–and towards relational truth. How are we present in relationship with others? How do the ideas we carry become embedded in AI systems?
Rokeya Akter
Rokeya presented on building an OCR model for text extraction from images within the specific context of Bangladesh. They discussed the challenges they faced, which represented a good insight into the technical hurdles of working in data-scarce environments.
Some challenges
- There’s no data in a Bangladeshi context
- She described her method for the collection of the images
- The difficulty of finding ones with the date pattern on the image
- There does not appear to be crowdsourcing for images in Bangladesh
- Someone asked: Are there platforms that allow you to find the images you need?
My Key Takeaway
I had two takeaways from this. The first was how technical literacy/familiarity and internet access create enormous disparities in how data science can be done in specific contexts. The second was, unrelated to the talk, about how these developing countries still have time to create laws to protect the biometrics of their citizens while more developed ones that have weak laws may be too late.
Chinasa T. Okolo
Chinasa wrote a survey of explainable AI (XAI) in the Global South for the conference. You can find her paper here. She discussed a lot of the issues with how AI is being brought to and used in the context of the Global South.
- Little work has been done to assess workers’ understanding of AI
- They are novices in how it works and what it can do
- Due to this it can exacerbate the inequalities, issues, and workload in these communities
- Chinasa’s ideas for the future:
- Centers Humans - acknowledging constraints
- Engaging with communities - fieldwork and ethnography
- Converging on a Research Agenda - to address and resolve inequities
The audience had some great questions (Note: I’ve removed who asked questions because I don’t want to make their information public without their permission.)
One attendee asked about the tension in balancing the organizational and resource constraints with the need to engage communities.
- Only 1 intervention actually used XAI in the field of the 16 papers assessed
- 3/16 are planning on deploying in the field
Another attendee asked whether there are good models in technology prior to AI that are potentially useful for Chinasa’s work.
- All the papers are within 2019 to 2021.
- Looking into other tech deployments (earlier) may be very useful.
- Ex. Google AI and Google Health - pushed AI to address Diabetes in Thailand
- Highlighted several problems on the ground, such as insufficient lighting
My Key Takeaway
Chinasa’s paper represents an excellent opportunity to reflect on how XAI is being used in the Global South by people who may not be aware of or intentional about the needs of those in the context they are working. It seems like a great starting point for having a conversation about what it means to “bring AI to underserved people” and how researchers need to consider how their work capitalizes, reinforces, or creates inequities.
Aprajita Bhandari
From Cell Phones to Crisis Mapping: Datafication of Humanitarian Aid as Enactments of Neo-Colonial Power
She discussed her work researching how data can be used during and after disasters to create surveillance of citizens and then export that data for the benefit of others. The context was an earthquake in Nepal in 2015, where thousands of people died. A non-profit (flowminder.org) partnered with the service provider (Ncell) to do a data collaboration. They shared the data with the UN in the name of aiding those affected by the earthquake.
Questions she asks
- What liberties are eroded in times of crisis?
- Who makes these decisions?
- Who is impacted?
She grounds her work in Post-Colonial Theorists
- Manidou - Techno-Colonialism
- Glissant - Right to Opacity
- Fanon - Occupied Breathing
She’s currently examining new projects with Flowminder.
- How is data-driven humanitarianism proliferating across the globe?
- These require surveillance technology to work
- Reinforce top-down solutions
Her goal
- It is not to provide blanket criticism
- Capture the complexities of data collection and how they become embedded in larger colonial systems
A few of the questions from the group
- Have you thought about or read research on futuring and speculative futures?
- What are the potential positive paths forward?
- With crisis there’s an interesting dynamic, the people who are thinking about a potential crisis are likely not the population?
- How would you even think about bringing power back to the people?
My Key Takeaway
Aprajita’s work highlights the criticality of protecting decision rights for individuals and communities. When other “interested” parties get involved without engaging the communities throughout the process there will always be issues of power and authority. One of the biggest things I felt upon listening to her work was how reactive all of these implementations of AI are and therefore how critical it is to have protections in place ahead of disasters. It reinforces the need to drive policy to protect people proactively.
In Washington state there is a very real fear of a major earthquake in the next 5-15 years, if that event occurs it could destroy large parts of Seattle and the coastal residences around the Puget Sound. Washington has failed to pass consumer privacy acts that would protect people’s personal information. In a situation like the one Aprajita describes, “well intentioned” organizations could co-opt “aid” to gather a tremendous amount of personal data.
Questions I have:
- Are residents of WA more vulnerable due to their proximity to Tech companies like Amazon, Microsoft, Google, AT&T, T-Mobile, Verizon, and the like?
- Are they more vulnerable due to the number of devices they have connected to the internet?
- Can we use this situation as a thought experiment to alter the focus, asking the question “how would we want this to be used on us?” Would we then be able to bring a new perspective to working with communities in the Global South that becomes more of a dialogue and less of an imposition?
Abdullah Hasan Safir
Digital Divide Using Data Science Techniques
Abdullah explored internet access using visualization methods to highlight some aspects of how data visualization may reinforce the inequalities in the data. He applied K-means clustering using 213 countries of the world into three clusters. His argument was that visualizations and analytical techniques can recreate inequalities.
Data
- GapMinder from UN (ITU)
- Lots of missing data in the dataset
- Used nearby available data (most recent year of available data)
- Requires manual adjustment of country names
- Data Rich vs Data Poor regions (lots of issues in Africa and 0 in Europe)
Visualizations
- Dot Plot by Continent
- Choropleth of the World Map
- Line Graph showing trend of continents by year over 29 years
- Ignores specific countries and context
- K-Means Clustering in a simple Table of 3 columns
- Applied a cluster against the world average
- Then show the clusters on the world map
My Key Takeaway
Data visualization is both an art and a science. Some of the points made about visual analysis are well known in the field, and require individual creators to make choices that best fit their data, audience, and intentions. One of the others, specifically the choropleth of the world map, reiterated some things I’ve struggled with:
- Who determines what countries are recognized, and how they are labeled, within products?
- Since most are created in the Global North, there is a fundamental imposition of values and worldviews that propagate any time we use these technologies. Does your map recognize Congo as two nations?
- If so, what names does it use?
- For other countries with complex colonial histories, what name is used and does it match the way the people who live there refer to themselves?
Mohammad Rashidujjaman Rifat
He explored what it was like working with religious communities around technology and AI. He brought up two case studies: 2017 in Bangladesh and 2020 (I didn’t capture where). I have captured some of his statements and quotes of participants to the best of my ability. Consider any quotes to be highly suspect unless you verify them directly with him.
Case 1
- 2017 - work trying to understand how mosques are using technology and their attitudes in Bangladesh
- 1 person said television is a “box of satan”.
- paraphrase ‘The larger trend is that there are ethical concerns around letting children see television programs’
- 1 person said “social media will destroy their children”.
- paraphrase ‘Moral opinion is that they should not let their kids interact with people who are from different perspectives’
Case 2
- 2020 - an ethnography of people recording Islamic public preaching and uploading them
- Focused on Facebook & TikTok
- Someone said they don’t watch videos on YouTube. Their reasoning is that there are a lot of videos on islamic on YouTube but he doesn’t want the commercial interference from unexpected ads in the middle of his video
- Auto play leads to other types of content are brought into his view unexpectedly
- Someone said they don’t watch videos on YouTube. Their reasoning is that there are a lot of videos on islamic on YouTube but he doesn’t want the commercial interference from unexpected ads in the middle of his video
- Others don’t use AI because they “predict” the future
- Only God can predict the future
Questions
- What are the assumptions people bring into their views or uncertainties about using technology?
- How do people manage the uncertainties they have around these technologies?
- How does this connect to the ideas of justice and fairness in AI?
My Key Takeaway
Religion and technology are two things I’m fairly familiar with. When listening to Mohammad speak I kept focusing on the ideas of insider and outsider (aka. emic and etic), as it exists in linguistics and sociology. Who holds the truth in these situations is a fundamental question for qualitative research that continues to be debated in these fields.
For the work of justice and fairness in AI, as posed in the last question, I find that it relates back to what Dr. Rangaswamy mentioned at the end of her keynote on understanding all data as socially created. To find solutions for communities such as the ones described within Mohammad’s case studies one must ask whether to privilege the views of the community or the observer.
One must negotiate the needs of all parties involved to identify something that is both practical and minimizes harm. This is something I aim to explore more.
Sharifa Sultana
Big Data, AI, and Misinformation in the Global South
Sharifa discussed some ethnographic techniques to work with people in Bangladesh. They studied fact checking and social effects in rural Bangladesh villagers in the Covid-19 pandemic and what is or is not “misinformation”. An example of misinformation was someone “selling amulets on the street claiming they could cure Covid-19”. It is hard to check facts because in Bangladesh only 3 entities are authorized for fact-checking. Very few people engage with them; it is mostly news agencies. Therefore, rural villagers rely on social experts’ opinions to create their understanding and gauge veracity.
- Since these are not categorized as “rational” how do AI systems account for these?
- How can we make these systems more tolerant of alternate ways of seeing the world?
Questions
- Is misinformation associated with radical groups in Bangladesh (such as the far-right in the US)?
- She has a hunch that they might be but did not actually examine this directly
- Can we suggest a different train of thought, rather than “non-rational” is not included? What if we consider “rational first” is embedded in the data systems? (Changes the question to a bit more positive framing.)
- Because of the way data and comp sci is taught the local person may not code their understanding into the system.
My Key Takeaway
Much of what Sharifa said resonated with Mohammad’s observations and Dr. Rangaswamy’s perspective. By forcing the Global North’s universals onto communities and populations they are inherently subverting the decision rights and autonomy of those groups. We as researchers, global citizens, and technologists need to understand this and find ways to ask how we can build systems that support other perspectives by default and allow us to achieve our ends without imposing our worldview as the “definitive” truth or some idealization of rationality.
Yousif Hassan
What does it mean to decolonize AI ethics?
Focuses on development of AI in the African context. He points out that AI Guidelines are being developed by largely western nations (‘just restating the obvious to make sure we’re all on the same page’). Then the Global South attempts to adopt these frameworks (embedded western epistemologies).
His overarching questions
- How should researchers and policy makers approach AI governance in Africa?
- AI4D (AI for Development) - ACTS in Kenya, Ghana AI Ethics Lab
- AfriLabs in Nigeria
Main Argument
- Researchers lack the African Context
- Decolonization is put forward as the “solution”, which he refutes as being truly useful
- He argues these economies purposefully exclude the imaginations of marginalized communities
- 3 main points
- Seems to only broaden the critique of AI
- Doesn’t problematize the historical origins of ethics and intelligence
- Doesn’t center the future in the African Context
- His way forward
- Move the project of theory making to the south
- Examine new conceptions of blackness under regimes of techno-scientific capitalism
- Engage with processes of co production in the margins
Questions
- How do you collect your data?
- Semi-structured interviews (40) with people in Canada, Kenya, Ghana, and Nigeria
- Are you familiar with Tyson Yunkaporta’s book on Sand Talk?
- It’s not a rejection of Western philosophy but the need to bring it into conversations with local contexts
- Are we exporting Canadian values and projects to the local context that may not be appropriate? Is it a form of neo-colonialism?
- One of the things that was looked at is the asymmetries of power, they approve the projects and success criteria for them
My Key Takeaway
I appreciated Yousif’s call to “move the project of theory making to the south” as I think there is a great deal of knowledge that is trapped at the periphery, or excluded entirely, because it doesn’t fit dominant narratives. When I heard his talk I was immediately reminded of the Stanford HAI’s recent talk with Sabelo Sethu Mhlambi, The Movement to Decolonize AI: Centering Dignity Over Dependency. I believe that the idea of “decolonializing AI” takes more than just changing the focus or allowing communities to participate. I feel it requires us to approach communities in a new way.
I’m also reminded of Tyson Yunkaporta’s book, Sand Talk, because it discusses a way of approaching people that keeps them in focus the entire time. In short it’s a methodology of approaching a community using a framework of Respect –> Connect –> Reflect –> Direct.
- Respect first by acknowledging, accepting, and supporting their autonomy and way of thinking.
- Connect with the community, and its individual members. The focus must be on what they need, want, and can sustain.
- Reflect on what, if anything, you have to offer them and whether or not they want you to be a part of a “solution”.
- Direct the creation of any solution alongside the community in an act of partnership and mutual interest.