integrated sensor systems
play

Integrated Sensor Systems Dept. of Electrical Engineering and - PowerPoint PPT Presentation

Institute of Integrated Sensor Systems Dept. of Electrical Engineering and Information Technology Comparison of PSO-Based Optimized Feature Computation for Automated Configuration of Multi-Sensor Systems Kuncup Iswandy and Andreas Koenig


  1. Institute of Integrated Sensor Systems Dept. of Electrical Engineering and Information Technology Comparison of PSO-Based Optimized Feature Computation for Automated Configuration of Multi-Sensor Systems Kuncup Iswandy and Andreas Koenig Kuncup Iswandy and Andreas Koenig

  2. Overview • Introduction • Feature Computation Methods • Optimization Methods • Experiments and Results • Conclusions Kuncup Iswandy and Andreas Koenig

  3. Introduction Motivation � Intelligent sensor system find more widespread application, e.g., the fields of ambient intelligence and sensor networks. � The growing spectrum of available sensor principles and implementations require: � a large variety sensor electronics � sensor signal processing techniques � The design process goes through: � the principal steps of sensor selection and scene optimization, � choice of signal and feature processing, � dimensionality reduction, and � classification Kuncup Iswandy and Andreas Koenig

  4. Introduction Motivation � Currently, optimizing design of intelligent sensor systems is - still manually elaborated by human designer - tedious - time and labor consuming process - potentially subpotimal outcome Kuncup Iswandy and Andreas Koenig

  5. Introduction Goals Signal Sensor Processing Dimension & & Classification Reduction Scene Feature Computation Parameter Assessment Optimization General architecture of intelligent sensor system � Goals: � contribute to the design automation activities for intelligent (multi-) sensor systems � focus on optimization of feature computation � comparison between genetic algorithms and particle swarm optimization Kuncup Iswandy and Andreas Koenig

  6. Feature Computation Methods � The roles of feature computation techniques: • extract the meaningful information of raw data of sensor response • reduce the dimension size of variable vector of a pattern • increase the computation speed and classification accuracy 4 x 10 2.5 cycle 1 cycle 2 � In particular to the application of 2 gas sensor systems, two feature 500°C 500°C 500°C conductance [a.u.] 1.5 H2 : 7 ppm. computations have been proposed: CH4 : 1000 ppm. Ethanol : 0.8 ppm. CO : 40 ppm � Multi-level threshold 1 290°C 290°C Gaussian windowing � 0.5 (kernel technique) 900°C 90°C 23°C 23°C 0 0 50 100 150 200 250 300 350 400 time [ms] Sensor response patterns during two temperature cycles Kuncup Iswandy and Andreas Koenig

  7. Feature Computation Methods Multi-Level Threshold (MLT) � MLT computes the features with similar to histogram and amplitude distribution. � Two methods of MLT, i.e., differential (DM) and cumulative (CM) modes. � The features of MLT can be computed as N r ∑ = δ z y T T ( , , ) • i s p q = s 1 ≤ ≤ ⎧ T y T 1 • δ = 0 p s q y T T ⎨ ( , , ) s p q otherwise ⎩ y s : magnitude value of sensor signal with s = 1, 2, ..., N r : total samples of a pattern N r i : a number of features ( i = T - 1) T : a number of thresholds used Tp and Tq : level-values with q = (2, 3, ... T ) and p = q – 1 for DM, and Feature computation of MLT for a gas stimulus with q = T and p = (1, 2, 3, ... T -1) for CM presentation of first derivative of conductance Kuncup Iswandy and Andreas Koenig

  8. Feature Computation Methods Gaussian Windowing Function � Extract features directly from conductance curves (transient responses) � Each feature is represented by a kernel-base, i.e., a Gaussian exponential function � Parameters : mean μ and standard deviation σ � The features of Gaussian Windowing can be computed as N = ∑ r ⋅ μ σ z y G s • ( , , ) i s i i = s 1 2 ⎛ − ⎞ μ s 1 ⎜ ⎟ − i ⎜ ⎟ μ σ = • G s exp σ ⎝ ⎠ ( , , ) 2 i i i Feature computaion of Gaussian windowing (window time slicing) for a normalized conductance curve Kuncup Iswandy and Andreas Koenig

  9. Optimization Methods Genetic Algorithms for a real-valued case � GA procedure: 1. Generate randomly a real-valued population 2. Selection for mating processess using Roulette Wheel Selection 3. Discrete recombination with rate = 0.8, example: X = (x 1 , x 2 , x 3 , x 4 , x 5 ) ------> X´ = (x 1 , x 2 , y 3 , x 4 , x 5 ) Y = (y 1 , y 2 , y 3 , y 4 , y 5 ) ------> Y´ = (y 1 , y 2 , x 3 , y 4 , y 5 ) 4. Mutation with Gaussian perturbation and rate = 0.01, example: X´ = (x 1 , x 2 , ..., x s , ..., x i ) , 1 ≤ s ≤ i x snew = x s + α s N (0, σ s ) 5. Replacement: elitism 10 % of parent population and offspring individuals 6. Repeate to step 2; otherwise stop. Kuncup Iswandy and Andreas Koenig

  10. Optimization Methods Particle Swarm Optimization for a real-valued case � One of the evolutionary computation techniques. � Population-based search algorithm � Population of random solution is called particles. � Related to the bird flocking and swarm theory. � PSO algorithm: ( ) ( ) + = + − + − • v t wv t C rand p x C Rand p x ( 1 ) ( ) () () id id id id gd id 1 2 p i + = + + x t x t v t • ( 1 ) ( ) ( 1 ) id id id d -space w = 1 : inertia weight x i_new C 1 = C 2 = 2 : positive constants p g Rand() and rand() : random functions, [0,1] v i v i_new : best previous position of the i- th particle p i p g : best particle among all particles x i Kuncup Iswandy and Andreas Koenig

  11. Optimization Methods Feature Assessment - Nonparametric Overlap Measure � Inspired by probability estimation in Edited-Nearest Neighbor method � Used as an optimization criterion (fitness function) � NPOM is computed as: + ∑ k ∑ k q n N L 1 1 c ∑ ∑ = = i NN i i 1 1 = q ji o ∑ k L N n 2 = = c j = i c i 1 1 1 ω = ω ⎧ n d j i = i NN q ⎨ = 1 − with n and ji ω ≠ ω NN − n i ji d ⎩ j i i NN jk � q o is well normalized in [0,1]; „1“ indicates no overlap of class regions Kuncup Iswandy and Andreas Koenig

  12. Optimization Methods Approaches of Feature Computation / Feature Selection � Wrapper approach: Assessment and param. R Modification Classification Dimension. Raw feature Classifier Sensor result computation Train/Test Reduction � Filter approach: Assessment and Assessment and Modification Modification param. param. q o q o Classification Raw feature Dimension. Classifier Sensor result computation Reduction Train/Test Kuncup Iswandy and Andreas Koenig

  13. Optimization Methods Genetic Algorithms for a binary case • Representation: binary (switch variables) • One point Crossover; rate = 0.8 • Mutation; rate = 0.01 • Reproduction: best 10 % parents and offsprings Crossover Mutation Random Random Random P1 P P2 Ofs Ofs1 Ofs2 : 1 : 0 Kuncup Iswandy and Andreas Koenig

  14. Optimization Methods Particle Swarm Optimization for a binary case � Original PSO is designed for real-value problems � For binary, particles are assigned a binary value, e.g., 0 or 1 � Velocity are restricted to an interval value of [-4,4] � BPSO algorithm: ( ) ( ) + = + − + − • v t wv t C rand p x C Rand p x ( 1 ) ( ) () () id id id id gd id 1 2 ( ) < ⎧ if U sigmoid v 1 , ( 0 , 1 ) = id • x ⎨ id otherwise ⎩ 0 , ( ) 1 = • sigmoid v id − + v e 1 id � Parameters w , C 1 , and C 2 are same with original PSO Kuncup Iswandy and Andreas Koenig

  15. Experiments and Results Data Description and Parameter Setting � Applying a benchmark data of a gas sensor system � Types of gases, i.e., H 2 , CH 4 , ethanol and CO � The data set consists of 810 measure values and 264 patterns � Separated into training (144 patterns) and testing (120 patterns) sets � Each experiment is repeated using 10 runs � Each run is limited up to 100 iterations � Population size is 20 individuals for both GA and PSO. � The number of nearest neighbors is set to five for the overlap (NPOM) measurement and the k NN voting classifier. � The classification accuracy is estimated using holdout method and leave-one-out cross-validation approach. Kuncup Iswandy and Andreas Koenig

  16. Experiments and Results Comparison between GA and PSO MLT differential mode Method overlap Recocognition accuracy ( k NN) q o train(%) test(%) test-LOO(%) Mean / Std Mean / Std Mean / Std Mean / Std 0.9950 / 0.0035 99.44 / 0.55 99.67 / 0.58 99.17 / 0.79 GA 1.00 / 0 100 / 0 100 / 0 99.83 / 0.35 PSO MLT cumulative mode Method overlap Recocognition accuracy ( k NN) q o train(%) test(%) test-LOO(%) Mean / Std Mean / Std Mean / Std Mean / Std 0.9878 / 0.0044 98.89 / 0.36 99.50 / 6.36 98.67 / 1.48 GA 0.9953 / 0.0024 99.10 / 0.34 99.92 / 0.89 99.83 / 0.35 PSO Kuncup Iswandy and Andreas Koenig

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend