17 December 2024
- RSIS
- Publication
- RSIS Publications
- IP24105 | AI in Humanitarian Action: Understanding the Digital Divide and Humanitarian Principles
SYNOPSIS
Artificial Intelligence (AI) is positioned to revolutionise humanitarian assistance and disaster relief in disaster-prone Southeast Asia with transformative prediction, forecast, and analytic capabilities. However, the digital divide and the risk to the humanitarian principles pose significant challenges. Hence, robust governance and strategies aligning AI with the humanitarian principles are essential to empower communities and build resilience.
COMMENTARY
Southeast Asia is one of the world’s most disaster-prone regions. The Philippines alone faced six tropical cyclones in the span of a month (October to November), causing widespread flooding, significant damage to infrastructure, and affecting millions of people. Mainland Southeast Asia was also severely impacted. NASA characterised the event as unusual with the Japan Meteorological Agency reporting that four simultaneously active cyclones in the Pacific basin, have not been observed since 1951. This is a glimpse of the disaster complexity across the region, which includes not only cyclones and its related hazards (flooding, rain-induced landslides, storm surge, strong winds), but also droughts, earthquakes, volcanic eruptions, and tsunamis.
Given such complexity, ASEAN has recognised the need for innovative solutions such as advancing the integration of Artificial Intelligence (AI) into its humanitarian assistance and disaster relief (HADR) strategies. Current plans, like the ASEAN Agreement on Disaster Management and Emergency Response (AADMER) Work Programme 2021-2025 and the ASEAN Coordinating Centre for Humanitarian Assistance on disaster management (AHA Centre) Work Plan 2025, include developing data-driven decision-making capabilities, fostering innovation hubs, and exploring AI applications. ASEAN should carefully tread the path of integrating AI to future-proof its disaster management as outlined and guided by the ASEAN Disaster Resilience Outlook 2025.
AI offers immense potential to transform HADR in the region enabling capabilities such as detecting emerging threats, enhancing the forecasting of typhoons and volcanic eruptions, providing early warnings to at-risk populations to facilitate pre-emptive evacuations, optimising resource mobilisation, predicting conflicts, and identifying migration streams. However, this potential is tempered by the region’s significant digital divide with a third of the Southeast Asian adult population still lacking access to digital technologies according to a 2021 report by Roland Berger. Additionally, the region averages only 56.22 out of 100 across six different pillars for the ASEAN Digital Integration Index, an index that measures digital integration. AI risks exacerbating this divide.
The challenge lies in leveraging AI to empower vulnerable communities while safeguarding the humanitarian principles. Striking this balance is essential to ensure that technological innovation fosters inclusivity, ethical applications, and long-term resilience.
Making the Case for AI in HADR
Weather prediction using AI is a competitive race and greatly substantiates AI’s impact. The Fengwu model of the Shanghai Artificial Intelligence Laboratory, Google DeepMind’s GraphCast, and Microsoft’s Aurora model are some of the notable AI models that showcase respective strengths in lead time and predicting track, intensity, and other aspects of weather systems over traditional prediction models. This application is pivotal for early warning systems and enables timely HADR to mitigate disaster risks.
The application of AI in conflict prediction and the analysis of migration streams has significantly enhanced humanitarian operations, particularly by forecasting the displacement of internally displaced persons (IDPs) and refugees. AI tools, leveraging vast datasets like satellite imagery, geospatial analytics, and real-time social media activity, enable precise predictions of conflict escalations and migration flows. For instance, the UN Global Pulse project has utilised machine learning to analyse sentiment and conflict indicators derived from social media, improving pre-emptive humanitarian responses. Similarly, NATO’s AI in strategic warning systems aims to analyse complex datasets for early detection of potential conflicts, enhancing crisis preparedness and response. Moreover, the US military’s Integrated Crisis Early Warning System (ICEWS) combines political event databases with advanced analytical models to deliver actionable conflict predictions.
However, a critical challenge arises with the involvement of non-humanitarian stakeholders, including private tech firms, military entities, and governmental actors, in developing and deploying these systems. Unlike traditional humanitarian organisations, these actors are not bound by the principles of humanity, impartiality, neutrality, and independence, leading to ethical dilemmas and potential misuse of predictive insights. An illustrative example is the collaboration between migration-focused NGOs and private analytics firms like Palantir, where advanced predictive analytics raised concerns about data privacy and alignment with humanitarian goals.
While these innovations underscore AI’s transformative potential, their ethical deployment necessitates rigorous governance frameworks that ensure adherence to the humanitarian principles, safeguarding the rights and dignity of affected populations.
AI and the Digital Divide
The AI trend is pushing organisations to implement the technology without ensuring its tangible benefits. This includes for HADR purposes. While AI holds promise for improving disaster response, its technical complexities can alienate the communities it aims to serve, exacerbating the digital divide or even vulnerabilities as crisis situations are high-stake environments not ideal for testing or deploying AI solutions. In the humanitarian domain, several different kinds of the digital divide become pertinent:
- Access divide: Beyond digital literacy, this refers to the disparity in access to digital infrastructure, such as the internet, AI systems, and devices, especially in remote or underserved areas. It also includes energy reliability and connectivity, which are crucial for AI deployment.
- Knowledge divide: A gap in understanding the ethical, operational, and strategic use of AI. This includes disparities in awareness of AI risks and benefits among stakeholders, including communities affected by disasters.
- Governance divide: A lack of robust frameworks to regulate AI use in humanitarian contexts, which can disadvantage organisations unable to navigate complex legal or ethical landscapes.
- Data divide: Inequalities in access to high-quality data for training AI models. Organisations with insufficient resources may lack the data diversity necessary for accurate predictions or solutions tailored to local contexts.
- Sustainability divide: Differences in the capacity to maintain and scale AI systems over time, considering ongoing costs, software updates, and adapting to evolving needs.
These forms of the digital divide highlight the multifaceted challenges in aligning AI with the humanitarian principles and emphasise the need for equitable governance and capacity-building initiatives.
The push for AI should not be obligatory but rather a nudge to understand risks, address fears and uncertainties, and make sense of AI for humanitarian purposes. It must be tailored to the context, focusing on empowering vulnerable populations rather than multiplying risks. For those in remote and inaccessible areas that lack technology or even basic necessities, AI should deliver tangible benefits such as improved food distribution or targeted evacuation plans — ensuring no one is left behind. Striking this balance between technological innovation and inclusivity is vital to achieve impactful HADR outcomes.
Risk to the Humanitarian Principles
The humanitarian principles of humanity, impartiality, neutrality, and independence are fundamental to humanitarian action, ensuring that aid is provided solely based on need, without discrimination, and free from political or military influence. These principles are crucial for maintaining trust and access in conflict and crisis situations.
The integration of AI into conflict prediction and migration analysis presents several risks to these principles. Regarding Impartiality, AI can introduce and perpetuate biases present in their training data, leading to skewed predictions that may favour certain groups over others. This undermines the principle of impartiality, as aid distribution could become unequal, not solely based on need. For instance, according to a study on AI for humanitarian action, if an AI system is trained on data that underrepresents certain populations, it may fail to predict crises affecting those groups, resulting in inadequate humanitarian response.
Neutrality is equally at stake when it comes to the misuse of data. Collaborations with non-humanitarian stakeholders, such as private tech firms or military entities, can compromise neutrality. Data collected for humanitarian purposes might be repurposed for military or political objectives, eroding the perceived neutrality of humanitarian organisations and potentially making them targets in conflicts. For example, partnerships between humanitarian agencies and private companies like Palantir have raised concerns about data being used beyond humanitarian aims, potentially for surveillance or military operations.
Independence is aimed to prevent undue external influence. However, the reliance on AI technologies developed by external entities can threaten exactly that independence of humanitarian organisations. Dependence on proprietary AI systems may subject humanitarian actions to the interests and agendas of technology providers, which may not align with humanitarian goals. This could lead to situations where aid delivery is influenced by external political or economic pressures, compromising the organisation’s autonomy.
To mitigate these risks, it is essential to establish robust data governance frameworks that uphold the humanitarian principles, ensuring that AI applications in humanitarian contexts are ethical, unbiased, and solely focused on alleviating human suffering. Furthermore, we might need to think to adapt the humanitarian principles to the 21st century’s data and AI focus by creating a set of digital humanitarian principles that protect humanitarians on the battlefield of the 5th domain of warfare (besides land, water, air and space), namely cyberspace.
Towards Inclusive AI in HADR
The integration of AI into HADR offers transformational opportunities to enhance disaster preparedness, response, and recovery. However, the implementation of AI must navigate significant challenges, including the digital divide, risks to the humanitarian principles, and ethical concerns. In regions like Southeast Asia, where natural hazards are frequent and devastating, the digital divide threatens to exclude populations from the benefits of AI-driven solutions. Therefore, ensuring that these AI solutions are inclusive, applicable, and context-appropriate is essential.
AI solutions must not undermine the humanitarian principles to not exacerbate inequalities. Robust governance frameworks and the respect for data privacy, intellectual property rights, and relevant guidelines should guide AI integration and are essential to safeguard data, address biases, and ensure that AI aligns with the core values of humanitarian action.
AI should serve as a catalyst — bridging gaps rather than widening them and empowering communities and stakeholders rather than alienating them. By fostering collaboration, innovation, and inclusivity, the humanitarian sector can harness AI for enhanced and sustainable resilience.
Andrej Janko Zwitter is Professor of Political Theory and Governance and Director of the Centre for Innovation, Technology and Ethics at the Rijksuniversiteit Groningen, Netherlands. Keith Paolo C. Landicho is an Associate Research Fellow of the Humanitarian Assistance and Disaster Relief [HADR] Programme, Centre for Non-Traditional Security Studies, S. Rajaratnam School of International Studies, Nanyang Technological University, Singapore.
SYNOPSIS
Artificial Intelligence (AI) is positioned to revolutionise humanitarian assistance and disaster relief in disaster-prone Southeast Asia with transformative prediction, forecast, and analytic capabilities. However, the digital divide and the risk to the humanitarian principles pose significant challenges. Hence, robust governance and strategies aligning AI with the humanitarian principles are essential to empower communities and build resilience.
COMMENTARY
Southeast Asia is one of the world’s most disaster-prone regions. The Philippines alone faced six tropical cyclones in the span of a month (October to November), causing widespread flooding, significant damage to infrastructure, and affecting millions of people. Mainland Southeast Asia was also severely impacted. NASA characterised the event as unusual with the Japan Meteorological Agency reporting that four simultaneously active cyclones in the Pacific basin, have not been observed since 1951. This is a glimpse of the disaster complexity across the region, which includes not only cyclones and its related hazards (flooding, rain-induced landslides, storm surge, strong winds), but also droughts, earthquakes, volcanic eruptions, and tsunamis.
Given such complexity, ASEAN has recognised the need for innovative solutions such as advancing the integration of Artificial Intelligence (AI) into its humanitarian assistance and disaster relief (HADR) strategies. Current plans, like the ASEAN Agreement on Disaster Management and Emergency Response (AADMER) Work Programme 2021-2025 and the ASEAN Coordinating Centre for Humanitarian Assistance on disaster management (AHA Centre) Work Plan 2025, include developing data-driven decision-making capabilities, fostering innovation hubs, and exploring AI applications. ASEAN should carefully tread the path of integrating AI to future-proof its disaster management as outlined and guided by the ASEAN Disaster Resilience Outlook 2025.
AI offers immense potential to transform HADR in the region enabling capabilities such as detecting emerging threats, enhancing the forecasting of typhoons and volcanic eruptions, providing early warnings to at-risk populations to facilitate pre-emptive evacuations, optimising resource mobilisation, predicting conflicts, and identifying migration streams. However, this potential is tempered by the region’s significant digital divide with a third of the Southeast Asian adult population still lacking access to digital technologies according to a 2021 report by Roland Berger. Additionally, the region averages only 56.22 out of 100 across six different pillars for the ASEAN Digital Integration Index, an index that measures digital integration. AI risks exacerbating this divide.
The challenge lies in leveraging AI to empower vulnerable communities while safeguarding the humanitarian principles. Striking this balance is essential to ensure that technological innovation fosters inclusivity, ethical applications, and long-term resilience.
Making the Case for AI in HADR
Weather prediction using AI is a competitive race and greatly substantiates AI’s impact. The Fengwu model of the Shanghai Artificial Intelligence Laboratory, Google DeepMind’s GraphCast, and Microsoft’s Aurora model are some of the notable AI models that showcase respective strengths in lead time and predicting track, intensity, and other aspects of weather systems over traditional prediction models. This application is pivotal for early warning systems and enables timely HADR to mitigate disaster risks.
The application of AI in conflict prediction and the analysis of migration streams has significantly enhanced humanitarian operations, particularly by forecasting the displacement of internally displaced persons (IDPs) and refugees. AI tools, leveraging vast datasets like satellite imagery, geospatial analytics, and real-time social media activity, enable precise predictions of conflict escalations and migration flows. For instance, the UN Global Pulse project has utilised machine learning to analyse sentiment and conflict indicators derived from social media, improving pre-emptive humanitarian responses. Similarly, NATO’s AI in strategic warning systems aims to analyse complex datasets for early detection of potential conflicts, enhancing crisis preparedness and response. Moreover, the US military’s Integrated Crisis Early Warning System (ICEWS) combines political event databases with advanced analytical models to deliver actionable conflict predictions.
However, a critical challenge arises with the involvement of non-humanitarian stakeholders, including private tech firms, military entities, and governmental actors, in developing and deploying these systems. Unlike traditional humanitarian organisations, these actors are not bound by the principles of humanity, impartiality, neutrality, and independence, leading to ethical dilemmas and potential misuse of predictive insights. An illustrative example is the collaboration between migration-focused NGOs and private analytics firms like Palantir, where advanced predictive analytics raised concerns about data privacy and alignment with humanitarian goals.
While these innovations underscore AI’s transformative potential, their ethical deployment necessitates rigorous governance frameworks that ensure adherence to the humanitarian principles, safeguarding the rights and dignity of affected populations.
AI and the Digital Divide
The AI trend is pushing organisations to implement the technology without ensuring its tangible benefits. This includes for HADR purposes. While AI holds promise for improving disaster response, its technical complexities can alienate the communities it aims to serve, exacerbating the digital divide or even vulnerabilities as crisis situations are high-stake environments not ideal for testing or deploying AI solutions. In the humanitarian domain, several different kinds of the digital divide become pertinent:
- Access divide: Beyond digital literacy, this refers to the disparity in access to digital infrastructure, such as the internet, AI systems, and devices, especially in remote or underserved areas. It also includes energy reliability and connectivity, which are crucial for AI deployment.
- Knowledge divide: A gap in understanding the ethical, operational, and strategic use of AI. This includes disparities in awareness of AI risks and benefits among stakeholders, including communities affected by disasters.
- Governance divide: A lack of robust frameworks to regulate AI use in humanitarian contexts, which can disadvantage organisations unable to navigate complex legal or ethical landscapes.
- Data divide: Inequalities in access to high-quality data for training AI models. Organisations with insufficient resources may lack the data diversity necessary for accurate predictions or solutions tailored to local contexts.
- Sustainability divide: Differences in the capacity to maintain and scale AI systems over time, considering ongoing costs, software updates, and adapting to evolving needs.
These forms of the digital divide highlight the multifaceted challenges in aligning AI with the humanitarian principles and emphasise the need for equitable governance and capacity-building initiatives.
The push for AI should not be obligatory but rather a nudge to understand risks, address fears and uncertainties, and make sense of AI for humanitarian purposes. It must be tailored to the context, focusing on empowering vulnerable populations rather than multiplying risks. For those in remote and inaccessible areas that lack technology or even basic necessities, AI should deliver tangible benefits such as improved food distribution or targeted evacuation plans — ensuring no one is left behind. Striking this balance between technological innovation and inclusivity is vital to achieve impactful HADR outcomes.
Risk to the Humanitarian Principles
The humanitarian principles of humanity, impartiality, neutrality, and independence are fundamental to humanitarian action, ensuring that aid is provided solely based on need, without discrimination, and free from political or military influence. These principles are crucial for maintaining trust and access in conflict and crisis situations.
The integration of AI into conflict prediction and migration analysis presents several risks to these principles. Regarding Impartiality, AI can introduce and perpetuate biases present in their training data, leading to skewed predictions that may favour certain groups over others. This undermines the principle of impartiality, as aid distribution could become unequal, not solely based on need. For instance, according to a study on AI for humanitarian action, if an AI system is trained on data that underrepresents certain populations, it may fail to predict crises affecting those groups, resulting in inadequate humanitarian response.
Neutrality is equally at stake when it comes to the misuse of data. Collaborations with non-humanitarian stakeholders, such as private tech firms or military entities, can compromise neutrality. Data collected for humanitarian purposes might be repurposed for military or political objectives, eroding the perceived neutrality of humanitarian organisations and potentially making them targets in conflicts. For example, partnerships between humanitarian agencies and private companies like Palantir have raised concerns about data being used beyond humanitarian aims, potentially for surveillance or military operations.
Independence is aimed to prevent undue external influence. However, the reliance on AI technologies developed by external entities can threaten exactly that independence of humanitarian organisations. Dependence on proprietary AI systems may subject humanitarian actions to the interests and agendas of technology providers, which may not align with humanitarian goals. This could lead to situations where aid delivery is influenced by external political or economic pressures, compromising the organisation’s autonomy.
To mitigate these risks, it is essential to establish robust data governance frameworks that uphold the humanitarian principles, ensuring that AI applications in humanitarian contexts are ethical, unbiased, and solely focused on alleviating human suffering. Furthermore, we might need to think to adapt the humanitarian principles to the 21st century’s data and AI focus by creating a set of digital humanitarian principles that protect humanitarians on the battlefield of the 5th domain of warfare (besides land, water, air and space), namely cyberspace.
Towards Inclusive AI in HADR
The integration of AI into HADR offers transformational opportunities to enhance disaster preparedness, response, and recovery. However, the implementation of AI must navigate significant challenges, including the digital divide, risks to the humanitarian principles, and ethical concerns. In regions like Southeast Asia, where natural hazards are frequent and devastating, the digital divide threatens to exclude populations from the benefits of AI-driven solutions. Therefore, ensuring that these AI solutions are inclusive, applicable, and context-appropriate is essential.
AI solutions must not undermine the humanitarian principles to not exacerbate inequalities. Robust governance frameworks and the respect for data privacy, intellectual property rights, and relevant guidelines should guide AI integration and are essential to safeguard data, address biases, and ensure that AI aligns with the core values of humanitarian action.
AI should serve as a catalyst — bridging gaps rather than widening them and empowering communities and stakeholders rather than alienating them. By fostering collaboration, innovation, and inclusivity, the humanitarian sector can harness AI for enhanced and sustainable resilience.
Andrej Janko Zwitter is Professor of Political Theory and Governance and Director of the Centre for Innovation, Technology and Ethics at the Rijksuniversiteit Groningen, Netherlands. Keith Paolo C. Landicho is an Associate Research Fellow of the Humanitarian Assistance and Disaster Relief [HADR] Programme, Centre for Non-Traditional Security Studies, S. Rajaratnam School of International Studies, Nanyang Technological University, Singapore.