Categories

Archives

Data science could help battle wildfires

fire

Guest comment by David Wild, Indiana University

A major wildfire spread through the western US state of Colorado, and I spent long hours locating shelters, identifying evacuation routes and piecing together satellite imagery.

As the Fourmile Canyon Fire devastated areas west of the city of Boulder, ultimately destroying 169 homes and causing US$217 million in damage, my biggest concerns were ensuring that people could safely evacuate and first responders had the best chance of keeping the fire at bay.

The oddest thing about that September 7, 2010?

I spent it sitting comfortably in my home in Bloomington, Indiana, 1600 kilimetres away from the action.

I was a volunteer, trying to help fire victims. I had created a webpage to aggregate data about the fire, including the location of shelters and the latest predictions of fire spread. I shared it on Twitter in the hope that someone would find it useful; according to the usage statistics, over 40,000 people did.

Today, researchers like me are finding transformative new ways to use data and computational methods to help planners, leaders and first responders tackle disasters like wildfires from afar.

A growing problem

The kind of work I do is increasingly necessary.

As I write this, wildfires are threatening homes across California. Vast areas are without electricity, due to the power company PG&E taking extreme measures to prevent downed power lines from igniting new fires, cutting off power to more than 2 million people.

Fueled by strong winds and dry conditions, these fires are a product of climate change.

It’s not just California where crisis is the new normal. Areas hit by hurricanes in 2017, such as Puerto Rico and the US Virgin Islands, are still struggling to recover. Here in the Midwest, we are dealing with unprecedented floods every year, caused by extreme rainfalls driven by climate change.

Federal aid agencies just can’t scale their response fast enough to meet the needs of disaster response and recovery at the level of annual disaster that Americans now face.

Even if world leaders take concrete steps to reduce carbon emissions, everyone on the planet is going to be confronting tough consequences for decades to come.

Data-driven solutions

But I am optimistic. A whole world of new possibilities have been opened up by an explosion of data. Artificial intelligence lets computers predict and find insights from that data.

Governmental and nongovernmental organisations are starting to recognise these possibilities. For example, in 2015, the Federal Emergency Management Agency (FEMA) appointed a chief data officer to “free the data” within the organisation.

This year, I helped found the Crisis Technologies Innovation Lab at Indiana University, specifically to harness the power of data, technology and artificial intelligence to respond to and prepare for the impacts of climate change.

Through a grant from the federal Economic Development Administration, we are building tools to help federal agencies like FEMA as well as local planners learn how to rebuild communities devastated by wildfires or hurricanes.

By analysing historical disaster information, publicly available census data and predictive models of risk and resilience, our tools will be able to identify and prioritise key decisions, like what kinds of infrastructure investments to make.

We are also partnering directly with first responders to create new kinds of disaster visualisations that fuse together thousands of data points about weather, current conditions, power outages and traffic conditions in real time. Only recently have such capabilities become possible in the field due to improvements in public safety communications infrastructure, such as FirstNet.

We hope that this will help the incident commanders and emergency managers make more informed decisions in high-stress situations.

Other researchers are already demonstrating the power of technology to help in wildfires and other disasters, including using drones to send back streaming video from the air; using artificial intelligence to predict the impact of disasters at a hyper-local level; and tracking the changes in air quality during wildfires much more accurately using sensors.

Resilience2 allows rapid analysis of disaster resilience at the state and county level.
David Wild, CC BY-SA

Looking ahead

The research we all are doing demonstrates ways that the powerful capabilities of data science and artificial intelligence could help planners, first responders and governments adapt to the huge challenges of climate change.

But there are barriers to overcome. Climate change and disasters are complex and difficult to model precisely.

What’s more, disaster response technology has to be specifically designed for high-stress, difficult environments. It has to be physically robust, able to operate in adverse environments with broken infrastructure. We need safe spaces to test and innovate new capabilities in simulated environments where failure does not result in real deaths.

My hope is that many more data and technology researchers will consider redirecting their research to the urgent problems of climate change.

The Conversation


Photo: AP Photo/Marcio Jose Sanchez

This article is republished from The Conversation under a Creative Commons license.

Under

Data, Predictive Analytics

Share

Tags

Related articles

AI governance: what is it? 
  By Briar Christensen, Principal Advisory Consultant AI governance: what is it?    AI Governance:…
keyboard_arrow_up