Qing Xu’s research while affiliated with Horizon Research and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (2)


VAD: Vectorized Scene Representation for Efficient Autonomous Driving
  • Conference Paper

October 2023

·

40 Reads

·

133 Citations

Bo Jiang

·

Shaoyu Chen

·

Qing Xu

·

[...]

·


Figure 2. Overall architecture of VAD. The full pipeline of VAD is divided into four phases. Backbone includes an image feature extractor and a BEV encoder to project the image features to the BEV features. Vectorized Scene Learning aims to encode the scene information into agent queries and map queries, as well as represent the scene with motion vectors and map vectors. In the inferring phase of planning, VAD utilizes an ego query to extract map and agent information via query interaction and outputs the planning trajectory (represented as ego vector). The proposed vectorized planning constraints regularize the planning trajectory in the training phase.
Ablation for module necessity. "Motion" and "Map" indicate whether VAD adopts the motion and map modules to per- form the tasks of motion prediction and online mapping, as well as utilizing the module outputs for planning.
VAD: Vectorized Scene Representation for Efficient Autonomous Driving
  • Preprint
  • File available

March 2023

·

219 Reads

Autonomous driving requires a comprehensive understanding of the surrounding environment for reliable trajectory planning. Previous works rely on dense rasterized scene representation (e.g., agent occupancy and semantic map) to perform planning, which is computationally intensive and misses the instance-level structure information. In this paper, we propose VAD, an end-to-end vectorized paradigm for autonomous driving, which models the driving scene as fully vectorized representation. The proposed vectorized paradigm has two significant advantages. On one hand, VAD exploits the vectorized agent motion and map elements as explicit instance-level planning constraints which effectively improves planning safety. On the other hand, VAD runs much faster than previous end-to-end planning methods by getting rid of computation-intensive rasterized representation and hand-designed post-processing steps. VAD achieves state-of-the-art end-to-end planning performance on the nuScenes dataset, outperforming the previous best method by a large margin (reducing the average collision rate by 48.4%). Besides, VAD greatly improves the inference speed (up to 9.3x), which is critical for the real-world deployment of an autonomous driving system. Code and models will be released for facilitating future research.

Download

Citations (1)


... Research in autonomous driving has moved away from the traditional strategy of integrating sequential, independently trained models [5,24,28,31,33,47,53], and moved towards developing multi-task systems that are trained in an end-to-end manner [4,6,8,18,22,51]. These approaches demonstrate improved performance and interpretability while being highly efficient. ...

Reference:

Distilling Multi-modal Large Language Models for Autonomous Driving
VAD: Vectorized Scene Representation for Efficient Autonomous Driving
  • Citing Conference Paper
  • October 2023