deep learning for self driving cars
play

Deep Learning for Self Driving Cars Link: deeplearning.mit.edu Lex - PowerPoint PPT Presentation

Deep Learning for Self Driving Cars Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu Course: Deep Learning for Self-Driving Cars http:// deeplearning.mit.edu Starts January 9, 2017 All lecture materials will be made publicly


  1. Deep Learning for Self Driving Cars Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  2. Course: Deep Learning for Self-Driving Cars http:// deeplearning.mit.edu • Starts January 9, 2017 • All lecture materials will be made publicly available • All in TensorFlow • Two video game competitions open to the public: • Rat Race: Winning the Morning Commute • 1000 Mile Race in American Truck Simulator Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  3. Semi-Autonomous Vehicle Components External 1. Radar 2. Visible-light camera 3. LIDAR 4. Infrared camera 5. Stereo vision 6. GPS/IMU 7. CAN 8. Audio Internal 1. Visible-light camera 2. Infrared camera 3. Audio Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  4. Self-Driving Car Tasks • Localization and Mapping: Where am I? • Scene Understanding: Where is everyone else? • Movement Planning: How do I get from A to B? • Driver State: What’s the driver up to? Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  5. Stairway to Automation Google Self-Driving Car Tesla Model S Data-driven approaches can help at every step Ford F150 not just at the top. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  6. Robotics at MIT: 31 Groups Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  7. Shared Autonomy: Self-Driving Car with Human-in-the-Loop Teslas instrumented: 17 Hours of data: 5,000+ hours Distance traveled: 70,000+ miles Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  8. Shared Autonomy: Self-Driving Car with Human-in-the-Loop Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  9. Driving: The Numbers (in United States, in 2014) Miles Fatalities • All drivers: 10,658 miles • Fatal crashes: 29,989 (29.2 miles per day) • All fatalities: 32,675 • Rural drivers: 12,264 miles • Car occupants: 12,507 • Urban drivers: 9,709 miles • SUV occupants: 8,320 • Pedestrians: 4,884 • Motorcycle: 4,295 • Bicyclists: 720 • Large trucks: 587 Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  10. Cars We Drive Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  11. Human at the Center of Automation: The Way to Full Autonomy Includes the Human Fully Fully Human Machine Controlled Controlled Ford F150 Tesla Model S Google Self-Driving Car Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  12. Human at the Center of Automation: The Way to Full Autonomy Includes the Human • Emergency • Automatic emergency breaking (AEB) • Warnings • Lane departure warning (LDW) • Forward collision warning (FCW) • Blind spot detection • Longitudinal • Adaptive cruise control (ACC) • Lateral • Lane keep assist (LKA) • Automatic steering • Control and Planning • Automatic lane change • Automatic parking Tesla Autopilot Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  13. Distracted Humans • Injuries and fatalities: What is distracted driving? 3,179 people were killed and 431,000 were • Texting injured in motor vehicle crashes involving • Using a smartphone distracted drivers (in 2014) • Eating and drinking • Texts: • Talking to passengers 169.3 billion text messages were sent in the • Grooming US every month. (as of December 2014) • Reading, including maps • Eye off road: • Using a navigation system 5 seconds is the average time your eyes are • Watching a video off the road while texting. When traveling at 55mph, that's enough time to cover the • Adjusting a radio length of a football field blindfolded. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  14. 4 D’s of Being Human: Drunk, Drugged, Distracted, Drowsy • Drunk Driving: In 2014, 31 percent of traffic fatalities involved a drunk driver. • Drugged Driving: 23% of night-time drivers tested positive for illegal, prescription or over-the-counter medications. • Distracted Driving: In 2014, 3,179 people (10 percent of overall traffic fatalities) were killed in crashes involving distracted drivers. • Drowsy Driving: In 2014, nearly three percent of all traffic fatalities involved a drowsy driver, and at least 846 people were killed in crashes involving a drowsy driver. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  15. In Context: Traffic Fatalities Total miles driven in U.S. in 2014: 3,000,000,000,000 (3 million million) Fatalities: 32,675 (1 in 90 million) Tesla Autopilot miles driven since October 2015: 200,000,000 (200 million) Fatalities: 1 Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  16. In Context: Traffic Fatalities Total miles driven in U.S. in 2014: 3,000,000,000,000 (3 million million) Fatalities: 32,675 (1 in 90 million) Tesla Autopilot miles driven since October 2015: 200,000,000 (200 million) Fatalities: 1 Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  17. In Context: Traffic Fatalities Total miles driven in U.S. in 2014: 3,000,000,000,000 (3 million million) Fatalities: 32,675 We (increasingly) understand this (1 in 90 million) Tesla Autopilot miles driven since October 2015: 130,000,000 (130 million) Fatalities: 1 We do not understand this (yet) We need A LOT of real-world semi-autonomous driving data! Computer Vision + Machine Learning + Big Data = Understanding Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  18. Vision-Based Shared Autonomy Face Video Instrument Cluster Supporting Sensors • Gaze • Autopilot state • Audio • Emotion • Vehicle state • CAN • Drowsiness • GPS • IMU Forward Video • SLAM Dash Video • Hands on wheel • Pedestrians, Vehicles • Activity • Lanes • Center stack interaction • Traffic signs and lights • Weather conditions Raw Video Behavior Shared Deep Nets Data Analysis Autonomy Semi-Supervised Learning Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  19. Camera and Lens Selection Logitech C920: Fisheye: Capture full range of head, body On-board H264 Compression movement inside vehicle. Case for C-Mount Lens: 2.8-12mm Focal Length: “Zoom” on the face Flexibility in lens selection without obstructing the driver’s view. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  20. Semi-Automated Annotation for Hard Vision Problems Hard = high accuracy requirements + many classes + highly variable conditions Trade-Off: Human Labor vs Accuracy Real Example: Gaze Classification 100% Accuracy 90% 0 a b Human Labor Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  21. Big Data is Easy… Big “Compute” Is Hard • Initial funding: $95M • 4,000 CPUs • 800 GPUs • 10 gigabit link to MIT Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  22. Self-Driving Car Tasks • Localization and Mapping: Where am I? • Scene Understanding: Where is everyone else? • Movement Planning: How do I get from A to B? • Driver State: What’s the driver up to? Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  23. Visual Odometry • 6-DOF: freed of movement • Changes in position: • Forward/backward: surge • Left/right: sway • Up/down: heave • Orientation: • Pitch, Yaw, Roll • Source: • Monocular: I moved 1 unit • Stereo: I moved 1 meter • Mono = Stereo for far away objects • PS: For tiny robots everything is “far away” relative to inter - camera distance Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  24. SLAM: Simultaneous Localization and Mapping What works: SIFT and optical flow Source: ORB-SLAM2 on KITTI dataset Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  25. Visual Odometry in Parts • (Stereo) Undistortion, Rectification • (Stereo) Disparity Map Computation • Feature Detection (e.g., SIFT, FAST) • Feature Tracking (e.g., KLT: Kanade-Lucas-Tomasi) • Trajectory Estimation • Use rigid parts of the scene (requires outlier/inlier detection) • For mono, need more info* like camera orientation and height of off the ground * Kitt, Bernd Manfred, et al. "Monocular visual odometry using a planar road model to solve scale ambiguity." (2011). Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  26. End-to-End Visual Odometry Konda, Kishore, and Roland Memisevic. "Learning visual odometry with a convolutional network." International Conference on Computer Vision Theory and Applications . 2015. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  27. Semi-Autonomous Tasks for a Self-Driving Car Brain • Localization and Mapping: Where am I? • Scene Understanding: Where is everyone else? • Movement Planning: How do I get from A to B? • Driver State: What’s the driver up to? Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  28. Object Detection • Past approaches: cascades classifiers (Haar-like features) • Where deep learning can help: recognition, classification, detection • TensorFlow : Convolutional Neural Networks Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  29. Computer Vision: Object Recognition / Classification Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  30. Computer Vision: Segmentation Original Ground Truth FCN-8 Source: Long et al. Fully Convolutional Networks for Semantic Segmentation. CVPR 2015. Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  31. Computer Vision: Object Detection Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

  32. Full Driving Scene Segmentation Fully Convolutional Network implementation: https://github.com/tkuanlun350/Tensorflow-SegNet Link: deeplearning.mit.edu Lex Fridman: fridman@mit.edu

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend