Mobileye Sensing Status and Road Map Dr. Gaby Hayon, EVP R&D 1 - - PowerPoint PPT Presentation

mobileye sensing status and road map
SMART_READER_LITE
LIVE PREVIEW

Mobileye Sensing Status and Road Map Dr. Gaby Hayon, EVP R&D 1 - - PowerPoint PPT Presentation

November 2019 Mobileye Sensing Status and Road Map Dr. Gaby Hayon, EVP R&D 1 Confidential The Challenge of Sensing for the automotive market ME sensing has three demanding customers Sensing state for ME policy under Smart agent for


slide-1
SLIDE 1

Confidential

1

  • Dr. Gaby Hayon, EVP R&D

Mobileye Sensing Status and Road Map

November 2019

slide-2
SLIDE 2

Smart agent for harvesting, localization and dynamic information for REM based map ADAS products working everywhere and at all conditions on millions of vehicles Sensing state for ME policy under the strict role of independency and redundancy.

The Challenge of Sensing for the automotive market

ME sensing has three demanding customers

slide-3
SLIDE 3

True redundancy

Surround computer vision Radar/Lidar sub-system

slide-4
SLIDE 4

ME’s AD Perception

surround computer vision

comprehensive env. model

ME’s AD Perception

slide-5
SLIDE 5

Comprehensive CV Environmental Model

Full and unified surround coverage of all decision-relevant environment elements. These are generally grouped into 4 categories:

Road Geometry (RG)

All driving paths, explicitly/partially/implicitly indicated, their surface profile and surface type.

Road Boundaries (RB)

Any delimiter of the drivable area, it’s 3D structure and semantics. Both laterally delimiting elements(FS) and longitudinally (general

  • bjects/debris).

Road Users (RU)

360 degrees detection and inter-camera tracking of any movable road-user, and actionable semantic-cues these users convey. (light indicators, gestures).

Road Semantics (RS)

Road-side directives (TFL/TSR) , on-road directives (text, arrows, stop-line , crosswalk) and their DP association.

slide-6
SLIDE 6

Object detection DNNs Texture engine , example Structure engine, example

Robust CV Environmental Model

Multiple independent visual-processing engines overlap in their coverage of the 4 categories (RG, RB, RU, RS)

To satisfy extremely-low nominal failure frequencies of the CV-Sub-system

Lanes detection DNN Single view Parallax-net elevation map Semantic Segmentation engine Multi-view Depth network Generalized-HPP (VF) Wheels DNN Road Semantic Networks RG RB, RU ,RS RB, RU RB, RU RG RU RU

slide-7
SLIDE 7

▪ Longitudinal and Lateral Driving plans / decisions

  • Overtake : Is the vehicle an obstacle?
  • Lane change: “Give-way“ /“take-way” labeling of objects
  • Assessment of objects likely trajectories by the scene.

▪ VRU related drive planning ▪ Environmental limitations ▪ Safe-stop possibility ▪ Emergency/Enforcement response

Support of different driving decisions & planning requires extraction of additional, essential set of contextual cues:

Actionable CV Environmental Model

Ped trajectory, intentions (head/body pose), relevance, vulnerability & host-path-access. visibility range , blockage, occlusions/view-range, road friction. Emergency vehicles / personnel detection, Gesture recognition. Is the road shoulder drivable? Is it safe to stop?

Cc cc

slide-8
SLIDE 8

Confidential

8

Visual perception

Environment Model Elements

slide-9
SLIDE 9

Road Users

slide-10
SLIDE 10

Road Users

360 degrees detection and inter-camera tracking of any movable road-user, and actionable semantic-cues these users convey (light indicators, gestures)

On top of the standalone Object detection networks running on all cameras, 2 Dedicated 360-stitching engines have been developed to assure completeness and coherency of the unified objects map:

  • Vehicle signature
  • Very close (part-of) vehicle in FOV : face & limits

“Full Image Detection”- raw signal “Full Image Detection” output- short range precise detection

slide-11
SLIDE 11

Road Users

slide-12
SLIDE 12

Road Users

slide-13
SLIDE 13

Road Users

Temporal tracker

Dimension net output

Metric Physical Dimensions estimation

dramatically improving measurements quality using novelty methods

slide-14
SLIDE 14

Road Users

Wheels- RU-part (relatively regular in shape) which we deliberately detect to affirm vehicle detections, 3D position, and tracking for high-function customers.

slide-15
SLIDE 15

Road Users

▪ The semantic segmentation is evident of all Road users, redundant to the dedicated networks ▪ It is also evident of extremely-small visible fragments of road users; These may potentially be used as scene-level contextual cues.

slide-16
SLIDE 16

Road Users – open door

Open car door is uniquely classified , as it is both extremely common, critical and of no ground intersection

slide-17
SLIDE 17

Road Users - VRU

Baby strollers and wheel chairs are detection through a dedicated engine on top

  • f the highly matured pedestrians detection system
slide-18
SLIDE 18

Road Users - VRU

Baby strollers and wheel chairs are detection through a dedicated engine on top of the highly matured pedestrians detection system

slide-19
SLIDE 19

Surround-view stitched SR FS

Road Boundaries

Occupancy Grid:

▪ Fusion of free space signal from 4 parking cameras, and front camera ▪ Main usages: a very accurate signal for handling crowded scenes, and a redundancy layer for

  • bjects detection, specifically general objects as containers, cones, carts, etc.

▪ Comparing the known scene (road edges and detected objects) with the occupancy

  • grid. The

differences are marked and reported as unknown objects.

slide-20
SLIDE 20
slide-21
SLIDE 21

Road Users

Emergency vehicle , light indicators Pedestrian understanding

Road users semantics

▪ Head/pose orientation Pedestrians posture/gesture. ▪ Vehicle light indicators Emergency vehicle/Personnel classification.

slide-22
SLIDE 22

Road Users

Pedestrian Gesture Understanding Come closer Stop! On the phone You can pass

slide-23
SLIDE 23

Confidential

Road Users

  • Redundant to the appearance-based engines
  • Reinforce detection and measurements to support higher level of end-functions
  • E.g.- dealing with “rear protruding” objects – which hover above the objects ground intersection.

Dense Structure-based Object detection

slide-24
SLIDE 24

Road Users

100° 100° 100°

  • Infers depth in "center" view using input from "center" and
  • verlapping "surround" cameras
  • Flexibility in camera placement and orientation compared to

canonical stereo-baseline camera pair setups

  • Covering blind-regions using e.g. parking camera in the front

region

  • Learning based approach allows finding good object shape

priors, and prediction in texture-less regions

  • Angular resolution much higher than Lidar
  • Provides independent measurement and detection modality
  • Does not rely on manual labeling
  • Predicts per-pixel depth independent of Lidar

DNN based multi-view stereo

How do we do this?

slide-25
SLIDE 25

Confidential

Road Users

slide-26
SLIDE 26

Road Users

DNN based multi-view stereo

slide-27
SLIDE 27

Road Users

DNN based multi-view stereo

slide-28
SLIDE 28

Leveraging Lidar Processing Module for Stereo Camera Sensing – “Pseudo-Lidar”

Road Users

Dense depth image from stereo cameras High-res Pseudo-Lidar Object detection Upright obstacle ‘stick’ extraction

slide-29
SLIDE 29

Road Users

  • RSS safety envelope should not be violate even in areas with limited

visibility

  • To ensure that, we must determine whether the reason for not detecting

an object is because it doesn't exist or due to an occlusion

  • The solution- creating a 360 deg visibility envelope and measuring

visibility range in all angles

  • Computation of information gathered from all cameras and the following

features:

  • Free space and road edges
  • Vehicles and pedestrians detection
  • REM map and road elevation

View Range knowing that you don’t know

slide-30
SLIDE 30

Road Users

Policy-level applications

placing "fake targets" in occluded areas that intersect with ego's planned path, assuming plausible speed and trajectory

Z axis view range

copping with occlusions deriving from road elevation Visible range Occluded

Ghost target

Visible range Occluded

slide-31
SLIDE 31

View range origin legend

Main Front Narrow Front Front Right Front Left Rear Right Rear Left Rear

slide-32
SLIDE 32

Road Boundaries

▪ Road ▪ Elevated ▪ Cars ▪ Bike, Bicycle ▪ Ped ▪ CA obj ▪ Guardrail ▪ Concrete ▪ Curbs ▪ Flat ▪ Snow ▪ Parking in ▪ Parking out Full Surface Segmentation Road/nRoad Detection of Any delimiter of the road surface- 3D structure and semantics. Both laterally delimiting elements(FS) and longitudinally (GO/debris) The Semantic segmentation engine provides a rich, high resolution pixel-level labeling; The SSN vocabulary is especially enriched to classify road delimiter types:

slide-33
SLIDE 33

Road Edge Car Bike Ped General object GuardRail Concret Curb Flat Snow

slide-34
SLIDE 34

Road Edge Car Bike Ped General object GuardRail Concret Curb Flat Snow

slide-35
SLIDE 35

Surround Road/nRoad classification

slide-36
SLIDE 36

Road Boundaries

Detection of Any delimiter of the road surface, it’s 3D structure and semantics. Both laterally delimiting elements(FS) and longitudinally (GO/debris)

The Parallax Net engine provides an accurate understanding of structure by assessing residual elevation (flow) from the locally governing road surface (homography). It is therefore evident of extremely small objects and low-elevation lateral boundaries.

slide-37
SLIDE 37

Debris detection identifies structural deviations from road surface. Structure from Motion approach: geometry-based & appearance-invariant.

detects any type of hazard.

Debris Detection

slide-38
SLIDE 38

Road Geometry - Road3 in production

https://www.youtube.com/watch?v=s7HCI33KVHA

Advanced lane applications (VW) Volkswagen Passat Travel Assist 2.0 with Mobileye camera

slide-39
SLIDE 39

Road Geometry

Road4 Technology provides deep lanes understanding rather than “simple” lane-marks detection

▪ Severely occluded lane-marks - Endures gaps of over 20m within marker ▪ Semi/partly/unmarked lane marker ▪ Multi-geometry lane structures – merge, split, HWE, junctions ▪ Stable DP map also pass-through Junctions and construction areas

Bots dots and occluded lane marks Lane detection on wet roads at night Merge and splits and passing through junctions

slide-40
SLIDE 40

Road Geometry

Parallax-Net

provides a dense understanding of all driving surface elevation model , and local detailed ‘longitudinal profile’ characteristics such as road bumps and ditches

slide-41
SLIDE 41

Road Geometry

▪ Host Driving Path : Geometry and Center ▪ Any-object (point) driving path ▪ Any-object (point) lane assignment ▪ Road-elevation - accounted-for by inference

The Generalized HPP technology (VF) provides

Does not involve explicit detection and modeling of lane-boundary evidence, but rather leverages top down contextual understanding.

slide-42
SLIDE 42

Road Semantics

▪ Road-side directives (TFL/TSR) ▪

  • n-road directives (text, arrows,

stop-line , crosswalk) ▪ Lane type- HOV, bicycle lane ▪ The DP association ▪ Road Friction ▪ Boundary type ▪ OCR

slide-43
SLIDE 43

Road Semantics

slide-44
SLIDE 44

Road Semantics

K

slide-45
SLIDE 45

Confidential

Lidar/Radar Sensing Subsystem

slide-46
SLIDE 46

Confidential

Lidar/Radar-only Subsystem Setup

slide-47
SLIDE 47

Environment Modeling– Road Users& Free- Space Detection

Free-Space detection via 3D Occupancy Engine

Model-based approach

Road User detection & tracking

Model-based approach

slide-48
SLIDE 48

Lidar Semantics - Shape Classification

Data-driven classification approach

Key use-case static object near crosswalk - distinguish between:

Dedicated Deep Neural Net fed with Lidar reflections to resolve semantic ambiguities. Pedestrians – give way Traffic signs – drive through

slide-49
SLIDE 49

Lidar-Localization in Camera-Generated Map

Localization in sparse semantic map is enabled by extracting rich Lidar features

Vehicle trajectory Semantic map information & Lidar reflections projected onto front camera Bird’s view display + map semantics