
This article is brought to you in association with the European Commission.
- The humanitarian sector is increasingly trying to anticipate needs, including by using AI.
- But mostly AI tools are built by and for large organizations, not local actors.
- Humanitarian AI should be designed with crisis-affected communities.
When it comes to managing a crisis, timing is everything. Getting ahead of a disaster through predictions, early warning systems or receiving timely and accurate information about the impact of a crisis can make all the difference to hundreds of lives. That’s why the humanitarian sector is increasingly promoting anticipatory action as a key way to reduce humanitarian needs. One of the emerging tools at the heart of this change is artificial intelligence (AI).
In 2021, a World Bank analysis of AI innovation in disaster risk management, highlighted that AI is being used for a wide range of tasks including vulnerability mapping, modelling population movement, predicting risks and supporting damage assessment. These tools have unprecedented potential to equip both responders and communities affected by crises with the timely and relevant information they need to make smarter decisions.
But deploying humanitarian AI presents several challenges. Some of these, such as the risks of data bias, threats to privacy and poor interpretability of models, are common to all high-stakes domains, from medicine to social services. Others, such as the lack of accountability and the danger of reinforcing old power dynamics between international and local actors pose a unique threat.
The risks posed by AI are in direct opposition to the Core Humanitarian Standard, the charter to which the aid sector strives to adhere, and also undermine sector commitments to give more agency to local organizations. An analysis of predictive analytics in the humanitarian sector showed that these tools are mostly used by large international agencies or industry. Our own research into existing AI tools for humanitarian action found that even when AI models are trained using locally generated data or insights, less than one-third are built to be used by local organizations or crisis-affected populations.
Unsurprisingly, AI tools are rarely designed or developed with the participation of locals on the ground who would be impacted by them. This puts communities at risk: AI models are notoriously bad at transferring to a new context. If they are built by teams thousands of miles away without meaningful local input, they are likely to miss out relevant insights about which problems to address, what data could prove useful, and how these technologies should be used.
The solution to this challenge lies in building locally developed and locally owned humanitarian AI, using participatory AI methodologies. Tools should be created that harness the collective intelligence of crisis-affected communities, for use by local organizations and communities who are typically the first responders to any crisis.
With a grant from the UK Humanitarian Innovation Hub, Nesta’s Centre for Collective Intelligence Design and the International Federation of the Red Cross & Crescent Societies, recently tested out this new approach, developing two new tools in Nepal and Cameroon to improve frontline crisis response. This project included working with communities to build new, more diverse datasets on which to train AI models – helping to minimize the risk of biased data. It also meant working with communities and frontline responders to agree how the tools would be evaluated – ensuring the AI model was optimized for the outcomes that mattered to them. In addition, communities were involved in refining the AI models during the development phase. As a result of community workshops in Nepal, ethnicity was removed as one of the model inputs; while in Cameroon, Red Cross volunteers identified a blind spot with the way the model worked.
Though challenging, projects like this one prove that it is possible to build local AI with local data, local infrastructure and local talent. And that it is possible to build AI that responds to local priorities and values. In short, taking a participatory AI approach that foregrounds local context and affected communities can help create more robust and socially acceptable AI tools.
Making inclusive AI a mainstream practice rather than an experimental one-off will require a shift across the sector from funders, agencies and tech innovators. Investing in the right technical skills at a local level is of course an obvious starting point. In addition, we need technologists everywhere to develop AI models that work in resource-constrained, data-scarce settings (which is the reality for most countries facing humanitarian crises), rather than building large, supervised, data-hungry models.
We need humanitarian funders to invest in a coordinated approach to filling data gaps, and build open datasets that are important to the operations of local humanitarian responders and communities, not just the big agencies. And we really need to see more efforts to design AI tools with the participation and oversight of crisis-affected communities. To achieve this, it will be critical to upskill community engagement teams to become participatory AI practitioners – and develop the skills and tools that can help bridge between technologists and local communities.
Discover
How is the World Economic Forum ensuring the ethical development of artificial intelligence?
The World Economic Forum’s Platform for Shaping the Future of Artificial Intelligence and Machine Learning brings together global stakeholders to accelerate the adoption of transparent and inclusive AI, so the technology can be deployed in a safe, ethical and responsible way.
- The Forum created a toolkit for human resources to promote positive and ethical human-centred use of AI for organizations, workers and society.
- From robotic toys and social media to the classroom and home, AI is part of life. By developing AI standards for children, the Forum is creating actionable guidelines to educate, empower and protect children and youth.
- The Forum is bringing together over 100 companies, governments, civil society organizations and academic institutions in the Global AI Action Alliance to accelerate the adoption of responsible AI in the global public interest.
- The Forum’s Empowering AI Leadership: AI C-Suite Toolkit provides practical tools to help companies better understand the ethical and business impact of their AI investment. The Model AI Governance Framework features responsible practices of leading companies from different sectors that organizations can adopt in a similar manner.
- In partnership with the UK government, the Forum created a set of procurement recommendations designed to unlock public-sector adoption of responsible AI.
- The Centre for the Fourth Industrial Revolution Rwanda is promoting the adoption of new technologies in the country, driving innovation on data policy and AI – particularly in healthcare.
Contact us for more information on how to get involved.
Changing the prevailing logic of humanitarian tech development won’t be easy, but what if every agency or funder simply started asking the question: How will affected communities participate in the design and oversight of this technology? This is surely the first move in creating humanitarian AI that is truly intelligent.
Speak your Mind Here