NATURAL, INTERACTIVE TRAINING OF SERVICE ROBOTS TO DETECT NOVEL - - PowerPoint PPT Presentation

natural interactive training of service
SMART_READER_LITE
LIVE PREVIEW

NATURAL, INTERACTIVE TRAINING OF SERVICE ROBOTS TO DETECT NOVEL - - PowerPoint PPT Presentation

MUNICH 10-12 OCT 2017 NATURAL, INTERACTIVE TRAINING OF SERVICE ROBOTS TO DETECT NOVEL OBJECTS Elisa Maiettini and Dr. Giulia Pasquale Joint work with: Prof. Lorenzo Natale, Prof. Lorenzo Rosasco iCub R1 What do we need? Target scenario


slide-1
SLIDE 1

Elisa Maiettini and Dr. Giulia Pasquale

Joint work with:

  • Prof. Lorenzo Natale, Prof. Lorenzo Rosasco

MUNICH 10-12 OCT 2017

NATURAL, INTERACTIVE TRAINING OF SERVICE ROBOTS TO DETECT NOVEL OBJECTS

slide-2
SLIDE 2

R1 iCub

slide-3
SLIDE 3

Target scenario What do we need?

WINDOW To close at night SOFA Elisa’s favorite sofa LIBRARY Contains books To dust every week PLANT To water every friday PLANT To water every monday

slide-4
SLIDE 4

Step 0: Object Recognition Step 1: Object Detection

Plant Window Sofa Sofa Library Are we done with Object Recognition? The R1 perspective. Giulia Pasquale, GTC 2017, San Jose, CA [http://on-demand.gputechconf.com/gtc/2017/video/s7295-giulia-pasquale-are-

we-done-with-object-recognition-the-r1-robot-perspective.PNG.mp4k ]

Sofa Plant

slide-5
SLIDE 5

Grid-based approaches[1][2]:

What is Where?

Region-based approaches[3][4][5]:

[3] Girshick R et al., 2014 [5] Shaoqing R et al., 2015 [4] Girshick R et al., 2015 [6] Uijlings J. R. R., 2013 [1] Redmond J. et al, 2016 [2] Liu W. et al, 2016

  • 1. Partition the image with a grid
  • 2. Run a classifier for each grid’s cell
  • 1. Identify Regions of Interest (RoI)
  • 2. Run a classifier for each RoI

Approaches to Object Detection

It’s a sofa!

No object … No object No object

We cannot run a classifier for each possible window!

Sliding window approach:

  • 1. Slide a window on the image
  • 2. Run a classifier for each window
slide-6
SLIDE 6

Region-based approach: Faster R-CNN[5]

RPN CNN

Feature Map

ROI pooling Layer For each ROI fc6 fc7

Region proposals Classification scores Predicted Bounding boxes Classifier Bounding box regressor

Where to look? Where? What?

[5] Shaoqing R et al., NIPS, 2015

 Modularity = Flexibility!!  RPN is faster and more efficient than external methods

slide-7
SLIDE 7

Object Detection task: Robotic setting

Robotics brings new challenges…

  • Open-set problem
  • Automatic self-supervision

…but also more information!

  • Time coherence
  • Contextual information

from sensors (e.g. depth)

?

slide-8
SLIDE 8

Object Detection for Robotics: our solution

  • 1. Data acquisition[8] and

model training

  • 2. Deployment on R1

[8] Pasquale et al., Frontiers 2016

Bounding boxes Labels

Detection system deployed on R1 thanks to the on board Jetson Tx2

slide-9
SLIDE 9

Object Detection for Robotics: our solution

iCubWorld Taransformations dataset[9]

[9] Pasquale et al. IROS 2016 [https://robotology.github.io/iCubWorld]

1 2 3 4 5 6 7 8 9 10

iCubWorld

slide-10
SLIDE 10

Object Detection for Robotics: evaluating models

Scenario 1: same HRI setting

1. Predictions compared with automatically acquired Ground Truth (Mean Average Precision = 0.71)

  • 2. Validate results: predictions compared with manual

Ground Truth (Mean Average Precision = 0.75)

Even better!

Scenario 2: different scenes

New sequences acquisition and manual annotation Promising results: mAPfloor=0.55, mAPtable=0.66 mAPshelf=0,53

slide-11
SLIDE 11

Deployment on R1

slide-12
SLIDE 12

Object Detection on R1 Train details

Performed offline on: GPU: NVIDIA Tesla P100 using: CNN: Zeiler and Fergus network[9] DATASET: iCubWorld Transformations with: Num images: ~27k RPN Train: Iterations: 81k Time: ~40 minutes Detector Train: Iterations: 54k Time: ~53 minutes Total train Time: ~3 hours

2 2

[9] Zeiler M. D. and Fergus R., CoRR, 2013

slide-13
SLIDE 13

Deployment details Object Detection on R1

Thanks to NVIDIA Jetson Tx2:  fast & easy  fully autonomous platform  easy systems integration CAFFE & YARP & Python Evaluation of Tensor RT framework

Regions per Frame Frame per second

100 ~4 300 ~3 1000 ~2

slide-14
SLIDE 14

Future steps:

  • 1. Further exploit contextual

information to improve precision (e.g. time coherence)

  • 2. Extend the system to open sets

towards scene understanding task

Contribution:

  • 1. Pipeline to overcome lack of manual

annotation for robotic platforms

  • 2. Deployment of detection system on

NVIDIA Jetson Tx2 on board

  • f R1
slide-15
SLIDE 15

Carlo Ciliberto Research Associate, UCL IRIS Vadim Tikhanoff Technologist Researcher, iCub Facility Raffaello Camoriano Postdoc, LCSL Ugo Pattacini Technologist Researcher, iCub Facility Marco Randazzo Senior Technician, iCub Facility Alberto Cardellino Junior Technician, iCub Facility Lorenzo Rosasco Team Leader, LCSL Lorenzo Natale Principal Investigator, iCub Facility Giorgio Metta Research Director, iCub Facility Alessandro Rudi Postdoc, LCSL Tanis Mar Postdoc, iCub Facility

  • …And all the team of the iCub Facility and the Laboratory for Computational and Statistical

Learning

Francesca Odone Assistant Professor,

  • Univ. of Genoa
slide-16
SLIDE 16

R1 is looking forward to meet you! Please come to see it at Booth number

slide-17
SLIDE 17

Thank you!

Any question?

Elisa Maiettini elisa.maiettini@iit.it Giulia Pasquale giulia.pasquale@iit.it