The work presented herein proposes DisasterReliefGPT, a multimodal AI system for automation in the areas of crisis communication and post-disaster assessment. The system integrates three tightly coupled components: a vision module called DisasterOCS for structural damage detection in satellite images, a Large Vision–Language Model (LVLM) for enhanced visual understanding and contextual reasoning, and a Large Language Model (LLM) to produce detailed, clear assessment reports. DisasterOCS relies on a ResNet34-based encoder with partial weight sharing and event-specific decoders, coupled with a custom MultiCrossEntropyDiceLoss function for multi-class segmentation on pre- and post-disaster image pairs. On the benchmark xBD dataset, the developed system reaches a high score of 78.8% in identifying F1-damage, making correct identifications of destroyed buildings with 81.3% precision, while undamaged structures are found with a very high value of 90.7%. From a combination of these components, emergency responders can immediately provide reliable and readable assessments of damage that can be used to directly support urgent decision-making.

DisasterReliefGPT: Multimodal AI for Autonomous Disaster Impact Assessment and Crisis Communication / Reghunath, L. C.; Abhishek, A. S.; Changat, A.; Unnikrishnan, A.; Rai, A. K.; Napoli, C.; Randieri, C.. - In: TECHNOLOGIES. - ISSN 2227-7080. - 14:3(2026). [10.3390/technologies14030179]

DisasterReliefGPT: Multimodal AI for Autonomous Disaster Impact Assessment and Crisis Communication

Napoli C.
;
Randieri C.
2026

Abstract

The work presented herein proposes DisasterReliefGPT, a multimodal AI system for automation in the areas of crisis communication and post-disaster assessment. The system integrates three tightly coupled components: a vision module called DisasterOCS for structural damage detection in satellite images, a Large Vision–Language Model (LVLM) for enhanced visual understanding and contextual reasoning, and a Large Language Model (LLM) to produce detailed, clear assessment reports. DisasterOCS relies on a ResNet34-based encoder with partial weight sharing and event-specific decoders, coupled with a custom MultiCrossEntropyDiceLoss function for multi-class segmentation on pre- and post-disaster image pairs. On the benchmark xBD dataset, the developed system reaches a high score of 78.8% in identifying F1-damage, making correct identifications of destroyed buildings with 81.3% precision, while undamaged structures are found with a very high value of 90.7%. From a combination of these components, emergency responders can immediately provide reliable and readable assessments of damage that can be used to directly support urgent decision-making.
2026
damage assessment; disaster response AI; emergency management; large vision–language model; multimodal learning; semantic segmentation; vision–language models
01 Pubblicazione su rivista::01a Articolo in rivista
DisasterReliefGPT: Multimodal AI for Autonomous Disaster Impact Assessment and Crisis Communication / Reghunath, L. C.; Abhishek, A. S.; Changat, A.; Unnikrishnan, A.; Rai, A. K.; Napoli, C.; Randieri, C.. - In: TECHNOLOGIES. - ISSN 2227-7080. - 14:3(2026). [10.3390/technologies14030179]
File allegati a questo prodotto
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1765420
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact