Image Processing-Based Object Recognition and Manipulation with a 5-DOF Smart Robotic Arm through a Smartphone Interface Using Human Intent Sensing
Haiming Gang hg1169@nyu.edu
Image Processing-Based Object Recognition and Manipulation with a - - PowerPoint PPT Presentation
Image Processing-Based Object Recognition and Manipulation with a 5-DOF Smart Robotic Arm through a Smartphone Interface Using Human Intent Sensing Haiming Gang hg1169@nyu.edu Background Some of manipulator work followed by fixed built-in
Haiming Gang hg1169@nyu.edu
Adjust itself according to various requirements and environment
Less sensors and marks
Help people finish some tasks according to daily activity data of people
Bas e Shoulder pan joint Shoulder lift joint Elbow joint Wrist joint Finge r
according to target object
angular and sends these data to Arbotix-M microcontroller
robot to pick up target object
cable and normally open
mobile phone via mjpg-streamer (Wireless Transmission)
interface or mobile phone chooses preset
running (Step count algorithm for iphone 5 and motion coprocessor for newer version)
manipulator (1 FPS)
phone does not need to obtain this data every time when using
Box Beer Toothpast e Pump Cup
Initially, the algorithm needs a lot of positive images (images of faces) and negative images (images without faces) to train the
haar features shown in below image are used. They are just like convolutional kernel. Each feature is a single value obtained by subtracting sum of pixels under white rectangle from sum of pixels under black rectangle.
For example, consider the image below. Top row shows two good
that the region of the eyes is often darker than the region of the nose and cheeks. The second feature selected relies on the property that the eyes are darker than the bridge of the nose. The haar feature reflects the change of the gray scalar of the image.
Haar Cascade Classifier
Positive image Negative image Box 700 3500 Beer 550 2750 Toothpaste 700 3500 Pump 700 3500 Cup 450 2250
(a) (b) (c) (d) (e)
Box: B>G>R Beer: G>B, G>R, G<40 and B>30 Toothpaste box R>G>B Pump R>G, R>B and R<27 Cup R>B>G
Obtain position information
1. Features2D + Homography to find mark 2. Houghline function to find the lines’ equations and use the equations to calculate the position of intersection points 3. Affine Transformation to get the birds’-eye-view and calculate the pixel distance between mark and object
Robot Mark Bottom of Object Global Coordinate y x
a) Set an alarm to pick up preset object b) Utilize habits of people to determine different target objects in different time