Effective Damage Data Generation by Fusing Imagery with Human Knowledge Using Vision-Language Models
By: Jie Wei , Erika Ardiles-Cruz , Aleksey Panasyuk and more
Potential Business Impact:
Helps rescue teams quickly see disaster damage.
It is of crucial importance to assess damages promptly and accurately in humanitarian assistance and disaster response (HADR). Current deep learning approaches struggle to generalize effectively due to the imbalance of data classes, scarcity of moderate damage examples, and human inaccuracy in pixel labeling during HADR situations. To accommodate for these limitations and exploit state-of-the-art techniques in vision-language models (VLMs) to fuse imagery with human knowledge understanding, there is an opportunity to generate a diversified set of image-based damage data effectively. Our initial experimental results suggest encouraging data generation quality, which demonstrates an improvement in classifying scenes with different levels of structural damage to buildings, roads, and infrastructures.
Similar Papers
Automated Wildfire Damage Assessment from Multi view Ground level Imagery Via Vision Language Models
CV and Pattern Recognition
Helps quickly see fire damage from pictures.
Structural Damage Detection Using AI Super Resolution and Visual Language Model
CV and Pattern Recognition
Helps drones see damage after disasters.
Efficient Few-Shot Learning in Remote Sensing: Fusing Vision and Vision-Language Models
CV and Pattern Recognition
Finds planes in pictures better, even blurry ones.