4.8 Article

GSAN: Graph Self-Attention Network for Learning Spatial-Temporal Interaction Representation in Autonomous Driving

Journal

IEEE INTERNET OF THINGS JOURNAL
Volume 9, Issue 12, Pages 9190-9204

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/JIOT.2021.3093523

Keywords

Task analysis; Adaptation models; Trajectory; Predictive models; Autonomous vehicles; Internet of Things; Modeling; Autonomous driving; graph attention network; interaction; lane changing; self-attention mechanism; trajectory prediction

Funding

  1. Hong Kong Research Grant Council [NSFC/RGC N_CityU 140/20]
  2. Science and Technology Innovation Committee Foundation of Shenzhen [JCYJ20200109143223052]

Ask authors/readers for more resources

Modeling interactions among vehicles is critical for improving efficiency and safety in autonomous driving. Most existing works consider interaction information implicitly and do not explore shared interaction representations. This article proposes a general graph self-attention network to learn interaction representations and utilizes pretraining and fine-tuning steps.
Modeling interactions among vehicles is critical in improving the efficiency and safety of autonomous driving since complex interactions are ubiquitous in many traffic scenarios. To model interactions under different traffic scenarios, most existing works consider interaction information implicitly in their specific tasks with hand-crafted features and predefined maneuvers. Extracting interaction representation, which can be commonly used among different downstream tasks, is not explored. In this article, we propose a general and novel graph self-attention network (GSAN) to learn the spatial-temporal interaction representation among vehicles by a framework consisting of pretraining and fine-tuning. Specifically, in the pretraining step, we construct the GSAN module based on a graph self-attention layer and a gated recurrent unit layer, and use trajectory autoregression to learn the interaction information among vehicles. In the fine-tuning step, we propose two different adaptation schemes to utilize the learned interaction information in various downstream tasks and fine-tune the entire model with only a few steps. To illustrate the effectiveness and generality of our spatial-temporal interaction model, we conduct extensive experiments on two typical interaction-related tasks, namely, lane-changing classification and trajectory prediction. The experiment results demonstrate that our approach significantly outperforms the state-of-the-art solutions of these two tasks. We also visualize the impact of surrounding vehicles on the ego vehicle in different interaction scenes. The visualization offers an intuitive explanation on how our model captures the dynamic changing interactions among vehicles and makes good predictions in various interaction-related tasks.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available