Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Member:sungbeanJo_paper [2021/03/04 19:50]
sungbean
Member:sungbeanJo_paper [2021/04/21 22:08] (current)
sungbean
Line 1: Line 1:
-Deep networks trained on demonstrations of human +get_config_param active timestamp_mode  
-driving have learned to follow roads and avoid obstacles. + TIME_FROM_INTERNAL_OSC 
-However, driving policies trained via imitation learning cannot +get_config_param active multipurpose_io_mode 
-be controlled at test time. A vehicle trained end-to-end to imitate + OUTPUT_OFF ​ 
-an expert cannot be guided to take a specific turn at an upcoming +get_config_param active sync_pulse_in_polarity 
-intersection. This limits the utility of such systems. We + ACTIVE_LOW 
-propose to condition imitation learning on high-level command +get_config_param active nmea_in_polarity 
-input. At test time, the learned driving policy functions as a + ACTIVE_HIGH 
-chauffeur that handles sensorimotor coordination but continues +get_config_param active nmea_baud_rate 
-to respond to navigational commands. We evaluate different + BAUD_9600
-architectures for conditional imitation learning in vision-based +
-driving. We conduct experiments in realistic three-dimensional +
-simulations of urban driving and on a 1/5 scale robotic truck +
-that is trained to drive in a residential area. Both systems +
-drive based on visual input yet remain responsive to high-level +
-navigational commands.+
  
-Imitation learning is receiving renewed interest as a 
-promising approach to training autonomous driving systems. 
-Demonstrations of human driving are easy to collect 
-at scale. Given such demonstrations,​ imitation learning can 
-be used to train a model that maps perceptual inputs to 
-control commands; for example, mapping camera images to 
-steering and acceleration. This approach has been applied to 
-lane following [27], [4] and off-road obstacle avoidance [22]. However, these systems have characteristic limitations. For 
-example, the network trained by Bojarski et al. [4] was given 
-control over lane and road following only. When a lane 
-change or a turn from one road to another were required, 
-the human driver had to take control 
- 
-Why has imitation learning not scaled up to fully autonomous 
-urban driving? One limitation is in the assumption 
-that the optimal action can be inferred from the perceptual 
-input alone. This assumption often does not hold in practice: 
-for instance, when a car approaches an intersection,​ the 
-camera input is not sufficient to predict whether the car 
-should turn left, right, or go straight. Mathematically,​ the 
-mapping from the image to the control command is no longer 
-a function. Fitting a function approximator is thus bound to 
-run into difficulties. This had already been observed in the 
-work of Pomerleau: “Currently upon reaching a fork, the 
-network may output two widely discrepant travel directions, 
-one for each choice. The result is often an oscillation in 
-the dictated travel direction” [27]. Even if the network can 
-resolve the ambiguity in favor of some course of action, it 
-may not be the one desired by the passenger, who lacks a 
-communication channel for controlling the network itself. 
-In this paper, we address this challenge with conditional 
-imitation learning. At training time, the model is given 
-not only the perceptual input and the control signal, but 
-also a representation of the expert’s intention. At test time, 
-the network can be given corresponding commands, which 
-resolve the ambiguity in the perceptuomotor mapping and 
-allow the trained model to be controlled by a passenger 
-or a topological planner, just as mapping applications and 
-passengers provide turn-by-turn directions to human drivers. 
-The trained network is thus freed from the task of planning 
-and can devote its representational capacity to driving. This 
-enables scaling imitation learning to vision-based driving in 
-complex urban environments. 
-We evaluate the presented approach in realistic simulations 
-of urban driving and on a 1/5 scale robotic truck. Both 
-systems are shown in Figure 1. Simulation allows us to 
-thoroughly analyze the importance of different modeling 
-decisions, carefully compare the approach to relevant baselines, 
-and conduct detailed ablation studies. Experiments 
-with the physical system demonstrate that the approach can 
-be successfully deployed in the physical world. Recordings 
-of both systems are provided in the supplementary video. 
Navigation