
DI驱动

Image Source : opendilab:dirve
Project Description
-
这是我在商汤科技实习期间参与的一个项目;
-
该项目属于 OpenDILab 下的一个子项目;
-
该项目致力于开发自动驾驶模拟器的决策任务平台;
-
更多的信息可以参考 github link.
My Contribution
-
Implemented macro-level functionalities within the Metadrive simulator' highway environment, employing the DQN algorithm to formulate lane-changing strategies.
-
Implemented Model Predictive Control (MPC) into the CARLA simulator, seamlessly combining trajectory tracking and path following modes.
-
Replicated the Disturbance-based Reward Extrapolation (D-REX) algorithm in the Drive environment, enabling the vehicle to acquire effective strateties and surpass provided demonstrations, even in the presence of suboptimal inputs.
Implement Details
Macro-Level Decision in Metadrive Simulator

-
Enable Metadrive to behave macro decision making strategies, such as Lane Change Left, Lane Change Right, Accelerating, Deccelerating and Maintain the Current State, like in highway-env.
-
Adapting Metadrive Simulator to the multi-process training in DI-engine in OpenDILab.
-
Adapting reinforcement learning algorithms for discrete actions, such as PPO and DQN, to this simulation environment.
Adapting MPC tracking algorithm to Carla SImulator

-
In the official Carla platform, there is only a simple PID control moudle available, which limit our ability to perform spatio-temporal planning tasks. For instance, we are unable to spedcify an entire trajectory and demand reaching a designated destination within a predefined time frame.
-
We designed two modes:
-
For Trajectory Planning: Tracking an entire trajectory with specified arrival times at specific locations.
-
For Lane-following tasks: Focus on tracking a path without time allocation.
-
-
Implementation can be found in this link.