bayesian regression with input noise for high dimensional
play

Bayesian Regression with Input Noise for High Dimensional Data - PowerPoint PPT Presentation

Bayesian Regression with Input Noise for High Dimensional Data Jo-Anne Ting 1 , Aaron DSouza 2 , Stefan Schaal 1 1 University of Southern California, 2 Google, Inc. June 26, 2006 Agenda Relevance of high dimensional regression with input


  1. Bayesian Regression with Input Noise for High Dimensional Data Jo-Anne Ting 1 , Aaron D’Souza 2 , Stefan Schaal 1 1 University of Southern California, 2 Google, Inc. June 26, 2006

  2. Agenda Relevance of high dimensional regression with input noise Introduction to Bayesian parameter estimation – EM-based Joint Factor Analysis – Automatic feature detection – Making predictions with noiseless query points Evaluation on a 100-dimensional synthetic dataset Application to Rigid Body Dynamics parameter identification – What are RBD parameters? – Formulate it as a linear regression problem – How to ensure physically consistent parameters? ICML 2006 2

  3. We are interested in parameter estimation… Traditional regression techniques ignore noise in input data, for example, for linear regression*: Unbiased regression solution Biased regression solution Noiseless inputs Noisy inputs * Solutions to linear problems can be easily extended to nonlinear ICML 2006 3 systems via locally weighted methods (e.g. Atkeson et al. 1997)

  4. …and Prediction With Noiseless Query Points For physical systems such as humanoid robots: – Noisy input data, large number of input dimensions -- of which not all is relevant We want to control these robots using model-based controllers: Training Phase t is the desired Testing Phase (noiseless) target ICML 2006 4

  5. Current Methods Are Unsuitable Ignores input noise Accounts for input noise • Total LS/Orthogonal LS (e.g. Golub & VanLoan 1998, Hollerbach & Wampler 1996) • OLS with robust Unsuitable for matrix inversion (e.g. high dimensional Belsley et al. 1980): • Joint Factor Analysis (JFA) data O(d 2 ) at best (Massey 1965): computationally prohibitive in high dimensions • LASSO & Stepwise ??? Suitable for high regression (Tibshirani dimensional data 1996, Draper & Smith 1981) ICML 2006 5

  6. Agenda Relevance of high dimensional regression with input noise Introduction to Bayesian parameter estimation – EM-based Joint Factor Analysis – Automatic feature detection – Making predictions with noiseless query points Evaluation on a 100-dimensional synthetic dataset Application to Rigid Body Dynamics parameter identification – What are RBD parameters? – Formulate it as a linear regression problem – How to ensure physically consistent parameters? ICML 2006 6

  7. Computationally Prohibitive? Not Any More! Introduce hidden variables, z im (D’Souza et al. 2004) � d y i = w zm t im + � y � d y i = z im + � y m = 1 m = 1 � d x i = w xm t im + � x � d z im = w zm t i m + � m m = 1 m = 1 EM-based JFA: All EM update equations are O(d) ICML 2006 7

  8. …but Remember the Important Parameters � d y i � � y = (1) w zm t im m = 1 � d x i � � x = w xm t im (2) m = 1 Divide (1) by (2) to get: This is the solution to the regression � y i � � y d JFA w zm t im problem y=b T x -- m = 1 = or... x i � � x � d which is what we w xm t im m = 1 need for prediction w zm ( ) � d x i � � x + � y y i = m = 1 w xm ICML 2006 8

  9. Next, We Add Automatic Feature Detection Priors: ( ) = Gamma a m , b m ( ) p � m ( ) = Normal 0, 1 � m � � p w zm � m � � ( ) = Normal 0, 1 � m � � p w xm � m � � Coupled regularization of regression parameters Still O(d) per EM iteration ICML 2006 9

  10. Making Predictions with Noiseless Query Points ˆ b noise = ? For a noisy test input x q and its unknown output y q , Given: We can infer: ( ) = ( ) d p y q x q � � p y q , Z , T x q Z d T b noise = � y 1 T B � 1 T � x � 1 W z A � 1 W x � 1 ˆ � y � 1 T B 1 � z y q x q = ˆ T x q b noise ˆ For a noiseless test input t q and its unknown output y q , b true = ? � y 1 T C � 1 � 1 � 1 W z ˆ ˆ b true = � y � 1 T C � 1 1 � z lim b noise W x � x � 0 � � ...where C = 11 T � 1 � y + � z � � � � ICML 2006 10

  11. Agenda Relevance of high dimensional regression with input noise Introduction to Bayesian parameter estimation – EM-based Joint Factor Analysis – Automatic feature detection – Making predictions using noiseless query points Evaluation on a 100-dimensional synthetic dataset Application to Rigid Body Dynamics parameter identification – What are RBD parameters? – Formulate it as a linear regression problem – How to ensure physically consistent parameters? ICML 2006 11

  12. Construction of 100-dimensional dataset Constructed data with – 10 relevant dimensions – 90 redundant and/or irrelevant dimensions Explored different combinations of redundant (r) and irrelevant (u) dimensions – r = 90, u = 0: 90 redundant dimensions – r = 0, u = 90: 90 irrelevant dimensions – r = 30, u = 60 – r = 60, u = 30 Tested on strongly noisy (SNR=2) and less noisy (SNR=5) data Predicted outputs with noiseless test inputs ICML 2006 12

  13. 10-70% Improvement for Strongly Noisy Data (SNR=2) Bayesian parameter estimation generalizes 10-70% better for strongly noisy data ICML 2006 13

  14. 7-50% Improvement on Less Noisy Data (SNR=5) …and 7-50% better for less noisy data ICML 2006 14

  15. Agenda Relevance of high dimensional regression with input noise Introduction to Bayesian parameter estimation – EM-based Joint Factor Analysis – Automatic feature detection – Making predictions with noiseless query points Evaluation on a 100-dimensional synthetic dataset Application to Rigid Body Dynamics parameter identification – What are RBD parameters? – Formulate it as a linear regression problem – How to ensure physically consistent parameters? ICML 2006 15

  16. What are Rigid Body Dynamics (RBD) Parameters? Using Newton-Euler equations for a rigid body, we get the RBD equation (where q are joint angles): ( ) + G q ( ) && ( ) � = M q q + C & q , q Mass matrix Vector of gravity terms Centripetal & Coriolis terms M , C and G are functions of mass, centre of mass and moments of inertia -- all which are unknown; q’ s and τ are known We can re-express the above linearly as: ( ) � � = Y q , & q , && q ICML 2006 16

  17. Formulate RBD Parameter Identification As A Linear Regression Problem (e.g. An et al. 1988) ( ) � � = Y q , & q , && q where the RBD parameters are… � = [ m , mc x , mc y , mc z , I 11 , I 12 , I 13 , I 22 , I 23 , I 33 ] T RBD parameters: – Must satisfy physical constraints (positive mass, positive definite inertia matrix) – But.. not all parameters are identifiable due to insufficiently rich data & constraints of the physical system (i.e. data is ill-conditioned) ICML 2006 17

  18. Specifically, a High Dimensional Noisy Linear Regression Problem � To enforce physical constraints on , introduce virtual � ˆ parameters : � 1 = ˆ 2 , � 2 = ˆ � 2 ˆ 2 , � 3 = ˆ � 3 ˆ � 1 � 1 � 1 2 ( ) ˆ 2 + ˆ 2 + ˆ � 4 = ˆ � 4 ˆ 2 , � 5 = ˆ � 1 � 5 � 4 � 3 � 1 2 2 • 11 features per DOF � 6 = ˆ � 5 ˆ � 6 � ˆ � 2 ˆ � 3 ˆ 2 , � 7 = ˆ � 5 ˆ � 7 � ˆ � 2 ˆ � 4 ˆ � 1 � 1 2 • For a system with s DOF, ( ) ˆ 2 + ˆ 2 + ˆ 2 + ˆ there are 11 s features � 8 = ˆ � 6 � 8 � 2 � 4 � 1 2 2 � 9 = ˆ � 6 ˆ � 7 + ˆ � 8 ˆ � 9 � ˆ � 3 ˆ � 4 ˆ � 1 2 ( ) ˆ 2 + ˆ 2 + ˆ 2 + ˆ 2 + ˆ � 10 = ˆ 2 , � 11 = ˆ � 7 � 9 � 10 � 2 � 3 � 1 � 11 2 2 Consequently, for real world systems, we have a noisy, high dimensional, ill-conditioned linear regression problem ICML 2006 18

  19. How to Ensure Our Robust Parameter Estimates are Physically Consistent? Find physically consistent robust parameter estimates that are as ˆ close to as possible b true ˆ � optimal Do a constraint optimization step to find : ( ) � � ˆ � w ˆ b true � f ˆ � optimal = arg min � � � ˆ where w m = 0 if dimension m is not relevant and w m = 1 otherwise ˆ Finally, ensure redundant/irrelevant dimensions in remain so b true � optimal in ICML 2006 19

  20. 10-20% Improvement on Robotic Oculomotor Vision Head 7 DOFs: 3 in neck, 2 in each eye 11 features per DOF; total of 77 features RBD parameter estimates from ALL algorithms satisfy physical constraints Bayesian de-noising does ~10-20% better Root Mean Squared Errors Algorithm Position(rad) Velocity(rad/s) Feedback (Nm) Ridge regression 0.0291 0.2465 0.3969 Bayesian de-noising 0.0243 0.2189 0.3292 LASSO regression 0.0308 0.2517 0.4274 Stepwise regression FAILURE FAILURE FAILURE ICML 2006 20

  21. 5-17% Improvement on Robotic Anthropomorphic Arm 10 DOFs: 3 in shoulder, 1 in elbow, 3 in wrist, 3 in fingers 11 features per DOF; total of 110 features Bayesian de-noising does ~5-17% better Root Mean Squared Errors Algorithm Position(rad) Velocity(rad/s) Feedback (Nm) Ridge regression 0.0210 0.1119 0.5839 Bayesian de-noising 0.0201 0.0930 0.5297 LASSO regression FAILURE FAILURE FAILURE Stepwise regression FAILURE FAILURE FAILURE ICML 2006 21

  22. Summary Bayesian treatment of Joint Factor Analysis that performs parameter estimation with noisy input data O(d) complexity per EM iteration Automatic feature detection through joint regularization of both regression branches Significant improvement on synthetic data and real-world systems ICML 2006 22

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend