Skip to content

Latest commit

 

History

History
22 lines (16 loc) · 1.97 KB

File metadata and controls

22 lines (16 loc) · 1.97 KB

Human-Like Autonomous Driving W/ Deep Reinforcement Learning

The objective is to develop human-like driving behaviors in autonomous vehicles using a hierarchical deep reinforcement learning (DRL) framework. Visual feature extraction and dimensionality reduction from RGB camera images are conducted using an autoencoder, which allows the agent to analyze the driving environment more effectively. Soft Actor-Critic (SAC) algorithm is selected as the DRL method for decision-making and control in autonomous driving; additionally, action noise is incorporated to facilitate a more efficient learning process. Action filtering techniques, such as Exponential Moving Average (EMA) and Weighted Moving Average (WMA), are employed to smooth the movements of the steering and throttle, thereby enhancing driving fluidity.

PS: Working on obstacle avoidence, currently.

The experiments carried out in this study aim to complete the course with a smooth human-like driving experience by making decisions and providing control with the proposed system, staying within the lane and at the targeted speed. The simulation environment required for the experiments was provided using the simulator of the Donkeycar project, an open-source autonomous vehicle driving platform.

Some detailed information about the study is given below. The study is in the development process and will be edited in the near future.

Video to observe performances: Video


System Architecture For Training
Screenshot from 2024-10-07 12-44-58
Model Type Description
SAC Basic SAC (Reference)
NSAC Action Noise Added SAC
NSAC-EMA Noised + Action filtered SAC w/ Exponential moving average (EMA)
NSAC-WMA Noised + Action filtered SAC w/ Weighted moving average (WMA)