Image Processing-Based Object Recognition and Manipulation with a - - PowerPoint PPT Presentation

image processing based object recognition and
SMART_READER_LITE
LIVE PREVIEW

Image Processing-Based Object Recognition and Manipulation with a - - PowerPoint PPT Presentation

Image Processing-Based Object Recognition and Manipulation with a 5-DOF Smart Robotic Arm through a Smartphone Interface Using Human Intent Sensing Haiming Gang hg1169@nyu.edu Background Some of manipulator work followed by fixed built-in


slide-1
SLIDE 1

Image Processing-Based Object Recognition and Manipulation with a 5-DOF Smart Robotic Arm through a Smartphone Interface Using Human Intent Sensing

Haiming Gang hg1169@nyu.edu

slide-2
SLIDE 2

Background

  • Some of manipulator work followed by fixed built-in command
  • Some of manipulator work followed by people command
  • Need lots of sensors or marks
  • Low intelligence
  • Difficult to participate the daily life of human
slide-3
SLIDE 3

Solution

  • Adjustable manipulator

Adjust itself according to various requirements and environment

  • Easy using manipulator

Less sensors and marks

  • Smart manipulator
  • Human partner

Help people finish some tasks according to daily activity data of people

slide-4
SLIDE 4

System diagram

slide-5
SLIDE 5

Structure of arm robot

Bas e Shoulder pan joint Shoulder lift joint Elbow joint Wrist joint Finge r

slide-6
SLIDE 6

How it work?

  • 1. Mobile phone or tablet receives image data from camera

according to target object

  • 2. Mobile phone or tablet deals with image frame by frame
  • 3. Mobile phone or tablet sends position information of target
  • bject and robot
  • 4. Raspberry pi solves inverse kinematic equation to get joint

angular and sends these data to Arbotix-M microcontroller

  • 5. Arbotix-M covert joint angular to electrical signal and operate

robot to pick up target object

  • 6. Place target object to a fixed and preset position
slide-7
SLIDE 7

Mobile phone or tablet receives image data from camera according to target object

  • Camera connect with raspberry pi by usb

cable and normally open

  • Raspberry pi sends image frame to

mobile phone via mjpg-streamer (Wireless Transmission)

  • Choose object by clicking button in the

interface or mobile phone chooses preset

  • bject automatically according to pace of

running (Step count algorithm for iphone 5 and motion coprocessor for newer version)

slide-8
SLIDE 8

Mobile phone or tablet deals with image frame by frame

  • Object recognition by haar feature classifier (10 FPS)
  • Obtain the position information between target object and

manipulator (1 FPS)

  • The position information of mark can be saved in phone, so the

phone does not need to obtain this data every time when using

Box Beer Toothpast e Pump Cup

slide-9
SLIDE 9

Haar cascade

Initially, the algorithm needs a lot of positive images (images of faces) and negative images (images without faces) to train the

  • classifier. Then we need to extract features from it. For this,

haar features shown in below image are used. They are just like convolutional kernel. Each feature is a single value obtained by subtracting sum of pixels under white rectangle from sum of pixels under black rectangle.

slide-10
SLIDE 10

For example, consider the image below. Top row shows two good

  • features. The first feature selected seems to focus on the property

that the region of the eyes is often darker than the region of the nose and cheeks. The second feature selected relies on the property that the eyes are darker than the bridge of the nose. The haar feature reflects the change of the gray scalar of the image.

slide-11
SLIDE 11

Haar Cascade Classifier

Positive image Negative image Box 700 3500 Beer 550 2750 Toothpaste 700 3500 Pump 700 3500 Cup 450 2250

slide-12
SLIDE 12

Distinguish Condition

(a) (b) (c) (d) (e)

Box: B>G>R Beer: G>B, G>R, G<40 and B>30 Toothpaste box R>G>B Pump R>G, R>B and R<27 Cup R>B>G

slide-13
SLIDE 13

Obtain position information

1. Features2D + Homography to find mark 2. Houghline function to find the lines’ equations and use the equations to calculate the position of intersection points 3. Affine Transformation to get the birds’-eye-view and calculate the pixel distance between mark and object

slide-14
SLIDE 14

Affine transformation

  • Perspective Transform
  • Distance measurement
  • Fixed height of object

Robot Mark Bottom of Object Global Coordinate y x

slide-15
SLIDE 15

Publish the position information to the robot by the ios devices

slide-16
SLIDE 16

Improvement

  • Faster image process speed
  • More accurate recognition algorithm
  • More functions combined with mobile phone

a) Set an alarm to pick up preset object b) Utilize habits of people to determine different target objects in different time

slide-17
SLIDE 17

Demo

slide-18
SLIDE 18

Thank you! &Question?