Real-Time 3D volumetric human body reconstruction from single view - - PowerPoint PPT Presentation

real time 3d volumetric human body reconstruction from
SMART_READER_LITE
LIVE PREVIEW

Real-Time 3D volumetric human body reconstruction from single view - - PowerPoint PPT Presentation

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device n . N Rafael Diniz and Mylne C.Q. Farias University of Braslia, Brazil http://www.ene.unb.br/mylene EI-3DMP, January 16, 2019, California Real-Time 3D


slide-1
SLIDE 1

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

n . N Rafael Diniz and Mylène C.Q. Farias University of Brasília, Brazil

http://www.ene.unb.br/mylene

EI-3DMP, January 16, 2019, California

slide-2
SLIDE 2

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Summary

Motivation and Goals Review of current work Proposed work Preliminary Results Conclusions and Future Work

Mylène C.Q. Farias | 16-Jan-2019 2 / 25

slide-3
SLIDE 3

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Volumetric video

High popularity of immersive experiences; Point Clouds:

Set of coordinates indicating the location of each point, along with one

  • r more atributes such as color associated with each point;

Alternative 3D content representation that allows visualization of scenes in a more immersive way; Viable solution to represent visual stimuli because of the efficiency and simplicity for capturing, storing and rendering of 3D objects;

Mylène C.Q. Farias | 16-Jan-2019 3 / 25

slide-4
SLIDE 4

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Challenges

3D point clouds are captured using multiple cameras and depth sensors in various setups; This results in thousands up to billions of points in order to represent realistically reconstructed objects or scenes;

Mylène C.Q. Farias | 16-Jan-2019 4 / 25

slide-5
SLIDE 5

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Challenges

Each point may have a 3D position information (x, y, z), a color information (R, G, B), and possibly atributes such like transparency, time of acquisition, reflectance of laser, etc. Efficient representation of point clouds are needed to store or transmit these information; Compression is much more difficult because each point is basically not related each other, e.g., no orders and no local topology exists.

Mylène C.Q. Farias | 16-Jan-2019 5 / 25

slide-6
SLIDE 6

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Motivation

Applications

It is believed that a wide range of applications and use cases can benefit from this type of data representation;

Mylène C.Q. Farias | 16-Jan-2019 6 / 25

slide-7
SLIDE 7

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Motivation

Applications

It is believed that a wide range of applications and use cases can benefit from this type of data representation;

Mylène C.Q. Farias | 16-Jan-2019 7 / 25

slide-8
SLIDE 8

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Motivation

Applications

Video communications is certainly one exciting application, but to be popular the equipment must be affordable to the final regular user.

Mylène C.Q. Farias | 16-Jan-2019 8 / 25

slide-9
SLIDE 9

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Motivation

Mylène C.Q. Farias | 16-Jan-2019 9 / 25

slide-10
SLIDE 10

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Capturing

Kinect-2 capture device

Time-of-Flight depth ranging technology Depth resolution of 512x424, distances from 0.5 to 4.5m, FoV of 70.6◦ by 60◦ (HxV), millimeter accuracy RGB in HD, downconverted to match depth resolution

Mylène C.Q. Farias | 16-Jan-2019 10 / 25

slide-11
SLIDE 11

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Typical volumetric video capture setup

Example with 4 Kinect 2 devices.

Mylène C.Q. Farias | 16-Jan-2019 11 / 25

slide-12
SLIDE 12

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Proposed Work

Goals:

Design a simple and affordable human body reconstruction system, which reconstructs a 3d person representation from a live input from a single RGB-D camera and a pre-captured 3D model; Use 1 Kinetic-2 device and a simple multicore CPU, making this the system affordable for regular users; Main target applications are video telepresence communication.

Mylène C.Q. Farias | 16-Jan-2019 12 / 25

slide-13
SLIDE 13

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Experimental Setup

Code writen in C and C++; Libraries used: Open3D and Libfreenect; Kinect-2 as main capture device; Intel Xeon E5-2620, with 80GB of RAM hardware; Lenovo T430 for mobile capture device.

Mylène C.Q. Farias | 16-Jan-2019 13 / 25

slide-14
SLIDE 14

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Constraints:

Assumes that the back of the head of the person is rigid; The speaker is looking ahead during most of the time; Self-occlusions do not occur ofen; This way, higher dynamics of the object (mouth, nose, eyes) can be fully present in the reconstructed 3D volumetric stream; The method can be extended to other types of objects.

Mylène C.Q. Farias | 16-Jan-2019 14 / 25

slide-15
SLIDE 15

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 1

Create a volumetric representation of each person joining the volumetric video session: Use only 1 kinetic-2 device to capture the model by moving the capture device around the the person; Use Truncated Signed Distance Function and Kinect Fusion to assemble the volumetric 3D object representation. Captured model example

Mylène C.Q. Farias | 16-Jan-2019 15 / 25

slide-16
SLIDE 16

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 1

Captured model examples

Mylène C.Q. Farias | 16-Jan-2019 16 / 25

slide-17
SLIDE 17

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 2

Then model is segmented and stored: The segmentation method uses RGB and depth; The nose information (maximum or minimum depth, depending on coordinate system) and its neighboring region, including the eyes, are segmented separately; The back of the head is segmented separately, using similar approach; This optimizes the efficiency of the registration between the model and the input live volumetric stream, by using smaller point-cloud inputs. Segmented models

Mylène C.Q. Farias | 16-Jan-2019 17 / 25

slide-18
SLIDE 18

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 2

Segmented model examples

Mylène C.Q. Farias | 16-Jan-2019 18 / 25

slide-19
SLIDE 19

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 3

Afer the model is captured, the live capture system can start, and a pre-processing step is done for each captured RGB and Depth frame pair: For each RGB and Depth frame pair captured, an alignment is necessary because the timestamps of the color and depth frames differ between 10ms to 20ms, which although less than a frame period (∼33ms at 30fps), is important specially for high speed movements; The RGB and Depth frames are converted to point-cloud, with camera coordinates converted to world coordinates by the use of Kinect’s intrinsics parameters; Point-cloud created from a live RGB-D input

Mylène C.Q. Farias | 16-Jan-2019 19 / 25

slide-20
SLIDE 20

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 3

Point-cloud from live single view capture examples

Mylène C.Q. Farias | 16-Jan-2019 20 / 25

slide-21
SLIDE 21

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Step 4

We perform proposed volumetric object reconstruction from a single RGB-D sensor: In a similar approach of Step 2, the point-cloud obtained from the live feed has its nose and adjacency area segmented; Using a fast global registration method between the segmented face from the model in Step 2 and the segmented input point-cloud, a transformation matrix is obtained; The back of the head segmented from the model in Step 2 is transformed using the transformation matrix; Finally, the transformed 3D model and live captured point-cloud are merged and the live reconstructed volumetric video frame is created. Reconstructed frame

Mylène C.Q. Farias | 16-Jan-2019 21 / 25

slide-22
SLIDE 22

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Methodology

Step 4

Reconstructed Point-cloud

Mylène C.Q. Farias | 16-Jan-2019 22 / 25

slide-23
SLIDE 23

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Results

Realtime CPU execution - under 33ms at 30fps; Beter experience when compared to incomplete objects; Allows volumetric video use cases using single RGB-D capture device; Room lef for GPU offloading optimization; Can be extended to different type of objects;

Mylène C.Q. Farias | 16-Jan-2019 23 / 25

slide-24
SLIDE 24

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

Conclusions

The pre-processing steps made into the point-clouds before the registration step are important for both quality and real-time execution target; A fast global registration is needed, at the cost of having sometimes an imperfect transformation matrix; Work shows Mixed Reality Volumetric Video use cases using single RGB-D camera running in an affordable CPU is possible; Proposed framework can be further extended to different type of

  • bjects, especially the ones with large rigid areas.

Issues to be addressed

Solve the color difference problems between model and input frame; Fusion / Merge of Point Cloud issues; Improve segmentation using Machine Learning

Mylène C.Q. Farias | 16-Jan-2019 24 / 25

slide-25
SLIDE 25

Real-Time 3D volumetric human body reconstruction from single view RGB-D capture device

qatlho’ Danke 谢谢 Grazie Спасибо ขอบคุณ

9C4#5$Ì

اﺮﻜﺷ Merçi Gracias நன்றி Obrigado Ευχαριστώ

감사합니다

धयवाद Terima kasih Thank you ありがとう Tapadh leibh ཐཱུགས་རྗཻ་ཆཻ་། Go raibh maith agaibh Xin cảm ơn

Qestions?

mylene@ene.unb.br, http://www.ene.unb.br/mylene http://www.ene.unb.br/mylene/databases.html

Mylène C.Q. Farias | 16-Jan-2019 25 / 25