AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (1.3 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Research Article | Open Access | Just Accepted

CogDrive: Cognition-driven multimodal prediction-planning  
fusion for safe autonomy

Heye Huang1,2Yibin Yang3Mingfeng Fan4Haoran Wang5Xiaocong Zhao5( )Jianqiang Wang3

1 Singapore-MIT Alliance for Research and Technology, Singapore 138602, Singapore.

2 Department of Urban Studies and Planning, Massachusetts Institute of Technology, Cambridge MA 02139, USA.

3 School of Vehicle and Mobility, Tsinghua University, Beijing 100084, China.

4 Department of Mechanical Engineering, National University of Singapore, Singapore 117576, Singapore.

5 Key Laboratory of Road and Traffic Engineering, Ministry of Education, Tongji University, Shanghai 200092, China. 

Show Author Information

Abstract

Safe autonomous driving in mixed traffic requires a unified understanding of multimodal interactions and dynamic planning under uncertainty. Existing learning-based methods often fail to capture rare but safety-critical behaviors, while rule-based systems lack adaptability in complex interactions. To address these limitations, we propose CogDrive, a cognition-driven multimodal prediction–planning fusion framework that integrates explicit modal reasoning with safety aware decision optimization. The prediction module introduces cognitive representations of interaction modes based on topological motion semantics and nearest-neighbor relational encoding. By incorporating a differentiable modal loss and multimodal Gaussian decoding, CogDrive effectively learns sparse and unbalanced interaction behaviors, improving long-tail trajectory prediction accuracy. The planning module builds upon an emergency-response concept and develops a safety-stabilized trajectory tree optimization. Short-term consistent root trajectories ensure safety within replanning cycles, while long-term branches provide smooth and collision-free avoidance under low-probability or rapidly switching modes. Experiments on Argoverse2 and INTERACTION datasets show that CogDrive achieves state-of-the-art performance, reducing minADE and miss rate while maintaining smoothness. Closed-loop simulations further confirm stable and adaptive behavior across strong-interaction scenarios such as merging and intersections. By coupling cognitive multimodal prediction with safety-oriented planning, CogDrive establishes an interpretable and reliable paradigm for safe autonomy in complex traffic. 

References

【1】
【1】
 
 
Communications in Transportation Research

{{item.num}}

Comments on this article

Go to comment

< Back to all reports

Review Status: {{reviewData.commendedNum}} Commended , {{reviewData.revisionRequiredNum}} Revision Required , {{reviewData.notCommendedNum}} Not Commended Under Peer Review

Review Comment

Close
Close
Cite this article:
Huang H, Yang Y, Fan M, et al. CogDrive: Cognition-driven multimodal prediction-planning  
fusion for safe autonomy.
Communications in Transportation Research, 2026, https://doi.org/10.26599/COMMTR.2026.9640016

320

Views

50

Downloads

0

Crossref

0

Web of Science

0

Scopus

Received: 31 October 2025
Revised: 10 December 2025
Accepted: 03 March 2026
Available online: 04 March 2026

© The author(s) 2026

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).