Artificial Intelligence: A Silver Bullet or Scrap Metal for Global Development?

February 6, 2019 Global Data Policy
Beverley Hatcher-Mbu
Explainer, Innovation

When someone mentions artificial intelligence (AI), it’s easy to conjure up two conflicting images: the first, killer robots whizzing past, replacing human jobs, daily tasks, and social interactions in a post-apocalyptic world; the second, a C-3PO-esque personality revolutionizing our health and food systems. Pondering this, we are also inclined to explore the question, where does global development fit in within this futuristic, Star Wars-inspired universe?

Our recent work collaborating with AidData to design a machine learning model predicting humanitarian trends, and our participation in an Open Data & Artificial Intelligence Roundtable at the Open Gov Hub, has reinforced for us that AI isn’t a “silver bullet” for global development. However, when used ethically, it can be (another) tool in the toolbox for tackling persistent global challenges.

ai

So What’s the Skinny on AI?

AI is the training of computers to automate decisions in order to simulate human-like decision making, while machine learning (ML) is a subset of AI that teaches computers to recognize patterns in data, using those patterns to make future predictions. Already, both are rife with possibilities for better development outcomes. AI is currently being used as a tool to improve health service delivery, implemented through mobile phones to make health screening more widely available for patients in rural, often underserved areas.

However, the downside is that AI has come under pressure for perpetuating racial discrimination in facial recognition technology. The same technology has also appeared in some countries as a tool to identify (and jail) political dissidents. While these varying outcomes are not unique to AI (plenty of repressive regimes have targeted other digital tools to perpetuate discrimination), this does mean that international actors hold significant ethical responsibility as they increasingly turn to AI to enhance programs on everything from agricultural value chains to health and wellness.

 

Factors to Consider

Pick your problems.

Not all problems can, or should, be solved by AI. Why? Because the use of AI is heavily shaped by the context surrounding the problem: the culture and processes already in existence, and the cost-benefit analysis of redesigning the culture so that AI can be used effectively. For example, for a health program operating in a rural area where community relationships are highly valued, AI could speed up the time needed to read x-rays, but it cannot effectively replace the program’s human elements – for example, the relationship between local health extension workers and community members that drives preventive care, such as motivation to get x-rays.

The ideal mix is a blend of humans and machines working together, resulting in the slow, time-adjusted reorganization of work that reflects which tasks can be completed by a machine, and which should be completed by a human.

AI requires investment.

AI is not like a microwave – you can’t “set and forget it.” AI models are affected by “concept drift,” meaning they have to be routinely re-calibrated with new data so that they continue to produce “correct” decisions. At DG, we learned this first hand in developing the Autogeocoder tool which uses machine learning to read through project documents and pinpoint specific activity locations. We continually updated our model to address questions such as, what happens when a project’s activities occur in multiple locations at once? Or when the document mentions other, unrelated locations such as the implementing partner’s mailing address? The Autogeocoder model needed to be trained, consistently and by many users over time, to become effective at making these important distinctions.

Mind the accountability/privacy gap.

AI can help increase the quality of public service delivery, but it can also exacerbate a lack of transparency around decision making. As with the open data world, basic principles and guidelines do exist that outline how algorithms can become accountable. Additionally, there is a developing methodology that breaks down how algorithms come to make their decisions, which can help identify bias. In the push for transparency in AI practices, France is leading the way by making public all algorithms its government uses.

Closely related to transparency concerns is the need to protect individual data, Vulnerable groups remain at risk in digital development spaces – highlighting just how essential it is that safety-oriented, post-colonial, and gender-focused concerns frame how, when, and why individual data is used to feed AI models.

 

Some Best Practices for Using AI

AI is already showing incredible ability to answer complex development questions, so it’s not all doom and gloom! As a result, we’ve identified three approaches that can help anchor AI models as solutions-oriented tools to tackle global challenges:

Talk to people.

Conduct an assessment before you build the AI model to gain a basic idea of the availability, quality, and timeliness of data that will feed your model. We developed CALM as a method to navigate assessing data ecosystems, to then build tools and processes that address the most pressing decisions.

Remember partnerships.

Consensus building through human partnerships remains essential, from accessing the data needed to teach the model, to ensuring that the model ultimately solves the problem raised by citizens/end users.

Draw parallels.

The medical industry has frequently balanced open data, informed consent, and restricted access, offering examples of how the data for development community could also weigh innovation with individual rights.

 

Where do we go next?

There is no “one size fits all” answer to solving global development challenges. Ultimately, the power of AI rests with the humans behind it, and hinges on their ability to connect AI to other digital and analog tools – that, collectively, can address complex problems. Effective use of AI will require proponents and skeptics alike to proceed both cautiously and with an open mind, to ensure that our analog weaknesses don’t become our digital ones.

 

Image credit: Mike MacKenzie, (CC-BY-2.0)

Share This Post

Related from our library

Demystifying interoperability: Key takeaways from our new white paper

This blog post gives an overview on our latest paper on interoperability, implementing interoperable solutions in partnership with public administrations. Based on over 20 years of DG’s experience, the paper demystifies key components needed to build robust, resilient, and interoperable data systems, focusing on the “how” of data standardization, data governance, and implementing technical infrastructure.

November 14, 2024 Agriculture, Digital Public Infrastructure
More Smoke, More Stroke

In honor of this year’s World Stroke Day, observed annually on October 29th, this piece aims to raise awareness of the substantial burden of non-communicable diseases–particularly stroke incidents–using the case study of Nigeria, one of the main tobacco production hubs on the continent, in addition to Kenya.

October 29, 2024 Health
Healthy Farming, Healthy Planet: The Environmental Case Against Tobacco Farming

While all agriculture has an environmental impact, tobacco is unique in that every stage of the tobacco lifecycle–from the production and consumption of tobacco to farming and disposal of the final product–wreaks havoc on the environment. In this piece, we’ll introduce the lifecycle of producing and using tobacco and explore the requisite environmental impact.

September 6, 2024 Agriculture, Health