scholarly journals Learn to Make Decision with Small Data for Autonomous Driving: Deep Gaussian Process and Feedback Control

2020 ◽  
Vol 2020 ◽  
pp. 1-11
Author(s):  
Wenqi Fang ◽  
Shitian Zhang ◽  
Hui Huang ◽  
Shaobo Dang ◽  
Zhejun Huang ◽  
...  

Autonomous driving is a popular and promising field in artificial intelligence. Rapid decision of the next action according to the latest few actions and status, such as acceleration, brake, and steering angle, is a major concern for autonomous driving. There are some learning methods, such as reinforcement learning which automatically learns the decision. However, it usually requires large volume of samples. In this paper, to reduce the sample size, we exploit the deep Gaussian process, where a regression model is trained on small sample datasets and captures the most significant features correctly. Besides, to realize the real-time and close-loop control, we combine the feedback control into the process. Experimental results on the Torcs simulation engine illustrate smooth driving on virtual road which can be achieved. Compared with the amount of training data in deep reinforcement learning, our method uses only 0.34% of its size and obtains similar simulation results. It may be useful for real road tests in the future.

2021 ◽  
Vol 70 ◽  
pp. 1517-1555
Author(s):  
Anirban Santara ◽  
Sohan Rudra ◽  
Sree Aditya Buridi ◽  
Meha Kaushik ◽  
Abhishek Naik ◽  
...  

Autonomous driving has emerged as one of the most active areas of research as it has the promise of making transportation safer and more efficient than ever before. Most real-world autonomous driving pipelines perform perception, motion planning and action in a loop. In this work we present MADRaS, an open-source multi-agent driving simulator for use in the design and evaluation of motion planning algorithms for autonomous driving. Given a start and a goal state, the task of motion planning is to solve for a sequence of position, orientation and speed values in order to navigate between the states while adhering to safety constraints. These constraints often involve the behaviors of other agents in the environment. MADRaS provides a platform for constructing a wide variety of highway and track driving scenarios where multiple driving agents can be trained for motion planning tasks using reinforcement learning and other machine learning algorithms. MADRaS is built on TORCS, an open-source car-racing simulator. TORCS offers a variety of cars with different dynamic properties and driving tracks with different geometries and surface.  MADRaS inherits these functionalities from TORCS and introduces support for multi-agent training, inter-vehicular communication, noisy observations, stochastic actions, and custom traffic cars whose behaviors can be programmed to simulate challenging traffic conditions encountered in the real world. MADRaS can be used to create driving tasks whose complexities can be tuned along eight axes in well-defined steps. This makes it particularly suited for curriculum and continual learning. MADRaS is lightweight and it provides a convenient OpenAI Gym interface for independent control of each car. Apart from the primitive steering-acceleration-brake control mode of TORCS, MADRaS offers a hierarchical track-position – speed control mode that can potentially be used to achieve better generalization. MADRaS uses a UDP based client server model where the simulation engine is the server and each client is a driving agent. MADRaS uses multiprocessing to run each agent as a parallel process for efficiency and integrates well with popular reinforcement learning libraries like RLLib. We show experiments on single and multi-agent reinforcement learning with and without curriculum


2021 ◽  
Author(s):  
Kazutaka Kanno ◽  
Atsushi Uchida

Abstract Reinforcement learning has been intensively investigated and developed in artificial intelligence in the absence of training data, such as autonomous driving vehicles, robot control, and internet advertising. However, the computational cost of reinforcement learning with deep neural networks is extremely high, and reducing the learning cost is a challenging issue. We propose a photonic on-line implementation of reinforcement learning using optoelectronic delay-based reservoir computing, both experimentally and numerically. In the proposed scheme, we accelerate reinforcement learning at a rate of several megahertz because there is no required learning process for the internal connection weights in reservoir computing. We perform two benchmark tasks, CartPole-v0 and MountanCar-v0 tasks, to evaluate the proposed scheme. Our results represent the first hardware implementation of reinforcement learning based on photonic reservoir computing and paves the way for fast and efficient reinforcement learning as a novel photonic accelerator.


2017 ◽  
Vol 7 (4) ◽  
pp. 243-255 ◽  
Author(s):  
Gennaro Notomista ◽  
Michael Botsch

AbstractA classification system for the segmentation of driving maneuvers and its validation in autonomous parking using a small-scale vehicle are presented in this work. The classifiers are designed to detect points that are crucial for the path-planning task, thus enabling the implementation of efficient autonomous parking maneuvers. The training data set is generated by simulations using appropriate vehicle-dynamics models and the resulting classifiers are validated with the small-scale autonomous vehicle. To achieve both a high classification performance and a classification system that can be implemented on a microcontroller with limited computational resources, a two-stage design process is applied. In a first step an ensemble classifier, the Random Forest (RF) algorithm, is constructed and based on the RF-kernel a General Radial Basis Function (GRBF) classifier is generated. The GRBF-classifier is integrated into the small-scale autonomous vehicle leading to excellent performance in parallel-, cross- and oblique-parking maneuvers. The work shows that segmentation using classifies and open-loop control are an efficient approach in autonomous driving for the implementation of driving maneuvers.


2016 ◽  
Vol 4 (2) ◽  
pp. 1-16
Author(s):  
Ahmed S. Khusheef

 A quadrotor is a four-rotor aircraft capable of vertical take-off and landing, hovering, forward flight, and having great maneuverability. Its platform can be made in a small size make it convenient for indoor applications as well as for outdoor uses. In model there are four input forces that are essentially the thrust provided by each propeller attached to each motor with a fixed angle. The quadrotor is basically considered an unstable system because of the aerodynamic effects; consequently, a close-loop control system is required to achieve stability and autonomy. Such system must enable the quadrotor to reach the desired attitude as fast as possible without any steady state error. In this paper, an optimal controller is designed based on a Proportional Integral Derivative (PID) control method to obtain stability in flying the quadrotor. The dynamic model of this vehicle will be also explained by using Euler-Newton method. The mechanical design was performed along with the design of the controlling algorithm. Matlab Simulink was used to test and analyze the performance of the proposed control strategy. The experimental results on the quadrotor demonstrated the effectiveness of the methodology used.


2019 ◽  
Vol 12 (2) ◽  
pp. 120-127 ◽  
Author(s):  
Wael Farag

Background: In this paper, a Convolutional Neural Network (CNN) to learn safe driving behavior and smooth steering manoeuvring, is proposed as an empowerment of autonomous driving technologies. The training data is collected from a front-facing camera and the steering commands issued by an experienced driver driving in traffic as well as urban roads. Methods: This data is then used to train the proposed CNN to facilitate what it is called “Behavioral Cloning”. The proposed Behavior Cloning CNN is named as “BCNet”, and its deep seventeen-layer architecture has been selected after extensive trials. The BCNet got trained using Adam’s optimization algorithm as a variant of the Stochastic Gradient Descent (SGD) technique. Results: The paper goes through the development and training process in details and shows the image processing pipeline harnessed in the development. Conclusion: The proposed approach proved successful in cloning the driving behavior embedded in the training data set after extensive simulations.


Sensors ◽  
2020 ◽  
Vol 20 (20) ◽  
pp. 5966
Author(s):  
Ke Wang ◽  
Gong Zhang

The challenge of small data has emerged in synthetic aperture radar automatic target recognition (SAR-ATR) problems. Most SAR-ATR methods are data-driven and require a lot of training data that are expensive to collect. To address this challenge, we propose a recognition model that incorporates meta-learning and amortized variational inference (AVI). Specifically, the model consists of global parameters and task-specific parameters. The global parameters, trained by meta-learning, construct a common feature extractor shared between all recognition tasks. The task-specific parameters, modeled by probability distributions, can adapt to new tasks with a small amount of training data. To reduce the computation and storage cost, the task-specific parameters are inferred by AVI implemented with set-to-set functions. Extensive experiments were conducted on a real SAR dataset to evaluate the effectiveness of the model. The results of the proposed approach compared with those of the latest SAR-ATR methods show the superior performance of our model, especially on recognition tasks with limited data.


2018 ◽  
Vol 51 (18) ◽  
pp. 31-36 ◽  
Author(s):  
Yuan Wang ◽  
Kirubakaran Velswamy ◽  
Biao Huang

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Ryoya Shiode ◽  
Mototaka Kabashima ◽  
Yuta Hiasa ◽  
Kunihiro Oka ◽  
Tsuyoshi Murase ◽  
...  

AbstractThe purpose of the study was to develop a deep learning network for estimating and constructing highly accurate 3D bone models directly from actual X-ray images and to verify its accuracy. The data used were 173 computed tomography (CT) images and 105 actual X-ray images of a healthy wrist joint. To compensate for the small size of the dataset, digitally reconstructed radiography (DRR) images generated from CT were used as training data instead of actual X-ray images. The DRR-like images were generated from actual X-ray images in the test and adapted to the network, and high-accuracy estimation of a 3D bone model from a small data set was possible. The 3D shape of the radius and ulna were estimated from actual X-ray images with accuracies of 1.05 ± 0.36 and 1.45 ± 0.41 mm, respectively.


Sign in / Sign up

Export Citation Format

Share Document