Ear-CCPM-Net: a cross-modal collaborative perception network for early accident risk prediction

Early traffic accident risk prediction in complex road environments poses significant challenges due to the heterogeneous nature and incomplete semantic alignment of multimodal data. To address this, we propose a novel Early Accident Risk Cross-modal Collaborative Perception Mechanism Network (EAR-C...

Full description

Saved in:
Bibliographic Details
Main Authors: Sun, Wei, Abdullah, Lili Nurliyana, Khalid, Fatimah, Sulaiman, Puteri Suhaiza
Format: Article
Language:en
Published: Multidisciplinary Digital Publishing Institute 2025
Online Access:http://psasir.upm.edu.my/id/eprint/121780/1/121780.pdf
http://psasir.upm.edu.my/id/eprint/121780/
https://www.mdpi.com/2076-3417/15/17/9299
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Early traffic accident risk prediction in complex road environments poses significant challenges due to the heterogeneous nature and incomplete semantic alignment of multimodal data. To address this, we propose a novel Early Accident Risk Cross-modal Collaborative Perception Mechanism Network (EAR-CCPM-Net) that integrates hierarchical fusion modules and cross-modal attention mechanisms to enable semantic interaction between visual, motion, and textual modalities. The model is trained and evaluated on the newly constructed CAP-DATA dataset, incorporating advanced preprocessing techniques such as bilateral filtering and a rigorous MINI-Train-Test sampling protocol. Experimental results show that EAR-CCPM-Net achieves an AUC of 0.853, AP of 0.758, and improves the Time-to-Accident (TTA0.5) from 3.927 s to 4.225 s, significantly outperforming baseline methods. These findings demonstrate that EAR-CCPM-Net effectively enhances early-stage semantic perception and prediction accuracy, providing an interpretable solution for real-world traffic risk anticipation.