Home  >  Article  >  Technology peripherals  >  Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

WBOY
WBOYforward
2023-04-12 23:40:011702browse

As autonomous driving technology continues to iterate, vehicle behavior and trajectory prediction are of extremely important significance for efficient and safe driving. Although traditional trajectory prediction methods such as dynamic model deduction and accessibility analysis have the advantages of clear form and strong interpretability, their modeling capabilities for the interaction between the environment and objects are relatively limited in complex traffic environments. Therefore, in recent years, a large number of research and applications have been based on various deep learning methods (such as LSTM, CNN, Transformer, GNN, etc.), and various data sets such as BDD100K, nuScenes, Stanford Drone, ETH/UCY, INTERACTION, ApolloScape, etc. have also emerged. , which provides strong support for training and evaluating deep neural network models. Many SOTA models such as GroupNet, Trajectron, MultiPath, etc. have shown good performance.

The above models and data sets are concentrated in normal road driving scenarios, and make full use of infrastructure and features such as lane lines and traffic lights to assist in the prediction process; due to limitations of traffic regulations, The movement patterns of most vehicles are also relatively clear. However, in the "last mile" of autonomous driving - autonomous parking scenarios, we will face many new difficulties:

  • Traffic rules in the parking lot The requirements for lane lines and lane lines are not strict, and vehicles often drive at will and "take shortcuts"
  • In order to complete the parking task, vehicles need to complete more complex parking actions, including frequent reversing, Parking, steering, etc. When the driver is inexperienced, parking may become a long process
  • There are many obstacles and clutter in the parking lot, and the distance between vehicles is close. If you are not careful, you may cause Collisions and scratches
  • Pedestrians often walk through the parking lot at will, and vehicles need more avoidance actions
    In such a scenario, simply apply the existing It is difficult for the trajectory prediction model to achieve ideal results, and the retraining model lacks the support of corresponding data. Current parking scene-based data sets such as CNRPark EXT and CARPK are only designed for free parking space detection. The pictures come from the first-person perspective of surveillance cameras, have low sampling rates, and have many occlusions, making them unable to be used for trajectory prediction.

In the just concluded 25th IEEE International Conference on Intelligent Transportation Systems (IEEE ITSC 2022) in October 2022, from University of California, Berkeley Researchers released the first high-definition video & trajectory data set for parking scenes, and based on this data set, used CNN and Transformer architecture to propose a trajectory prediction model called "ParkPredict" .

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

  • Paper link: https://arxiv.org/abs/2204.10777
  • Dataset home page, trial and download application: https://sites.google.com/berkeley.edu/dlp-dataset (If you cannot access, you can Try the alternative page //m.sbmmt.com/link/966eaa9527eb956f0dc8788132986707 )
  • ##Dataset Python API: https://github.com/MPC- Berkeley/dlp-dataset
Dataset information

The data set was collected by drone, with a total duration of 3.5 hours, video resolution For 4K, the sampling rate is 25Hz. The view covers a car park area of ​​approximately 140m x 80m, with a total of approximately 400 parking spaces. The dataset is accurately annotated, and a total of 1216 motor vehicles, 3904 bicycles, and 3904 pedestrian trajectories were collected.

After reprocessing, the trajectory data can be read in the form of JSON and loaded into the data structure of the connection graph (Graph):

  • Individual (Agent): Each individual (Agent) is an object moving in the current scene (Scene). It has attributes such as geometric shape and type. Its movement trajectory is stored as a file containing Linked List of Instances
  • Instance: Each instance is the state of an individual (Agent) in a frame (Frame). Contains its position, angle, speed and acceleration. Each instance contains a pointer to the instance of the individual in the previous frame and the next frame
  • Frame (Frame): Each frame (Frame) is a sampling point, and its Contains all visible instances (Instance) at the current time, and pointers to the previous frame and the next frame
  • Obstacle (Obstacle): The obstacle is in this record Objects that do not move at all, including the position, corner and geometric size of each object
  • Scene (Scene): Each scene (Scene) corresponds to a recorded video file, which contains pointers , pointing to the first and last frames of the recording, all individuals (Agents) and all obstacles (Obstacles)

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

provided by the data set Two download formats:

JSON only (recommended) : JSON file contains the types, shapes of all individuals , trajectories and other information can be directly read, previewed, and generated semantic images (Semantic Images) through the open source Python API. If the research goal is only trajectory and behavior prediction, the JSON format can meet all needs.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

##Original video and annotation: If the research is based on the original camera For topics in the field of machine vision such as target detection, separation, and tracking of raw images, you may need to download the original video and annotation. If this is required, the research needs need to be clearly described in the dataset application. In addition, the annotation file needs to be parsed by itself.

Behavior and trajectory prediction model: ParkPredict

As an application example, in the paper "ParkPredict: Multimodal Intent and Motion Prediction for Vehicles in Parking Lots with CNN" at IEEE ITSC 2022 and Transformer", the research team used this data set to predict the vehicle's intent (Intent) and trajectory (Trajectory) in the parking lot scene based on the CNN and Transformer architecture.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

The team used the CNN model to predict the distribution probability of vehicle intent (Intent) by constructing semantic images. This model only needs to construct local environmental information of the vehicle, and can continuously change the number of available intentions according to the current environment.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

The team improved the Transformer model and provided the intent (Intent) prediction results, the vehicle's movement history, and the semantic map of the surrounding environment as inputs to achieve Multi-modal intention and behavior prediction.

Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.

Summary

  • As the first high-precision data set for parking scenarios, the Dragon Lake Parking (DLP) data set can achieve large-scale target recognition and tracking, idle Parking space detection, vehicle and pedestrian behavior and trajectory prediction, imitation learning and other research provide data and API support
  • By using CNN and Transformer architecture, the ParkPredict model’s behavior and performance in parking scenarios In addition to showing good capabilities in trajectory prediction
  • Dragon Lake Parking (DLP) data set is open for trial and application. You can visit the data set homepage https://sites.google.com/ berkeley.edu/dlp-dataset for more information (if you cannot access, you can try the alternative page //m.sbmmt.com/link/966eaa9527eb956f0dc8788132986707

The above is the detailed content of Berkeley open sourced the first high-definition data set and prediction model in parking scenarios, supporting target recognition and trajectory prediction.. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete