6 02 fall 2012 lecture 10
play

6.02 Fall 2012 Lecture #10 Linear time-invariant (LTI) models - PowerPoint PPT Presentation

6.02 Fall 2012 Lecture #10 Linear time-invariant (LTI) models Convolution 6.02 Fall 2012 Lecture 10, Slide #1 Modeling Channel Behavior codeword bits in x[n] generate 1001110101 DAC digitized modulate symbols NOISY &


  1. 6.02 Fall 2012 Lecture #10 • Linear time-invariant (LTI) models • Convolution 6.02 Fall 2012 Lecture 10, Slide #1

  2. Modeling Channel Behavior codeword bits in x[n] generate 1001110101 DAC digitized modulate symbols NOISY & DISTORTING ANALOG CHANNEL demodulate sample & 1001110101 ADC & filter threshold y[n] codeword bits out 6.02 Fall 2012 Lecture 10, Slide #2

  3. The Baseband** Channel input response x[n] y[n] S A discrete-time signal such as x[n] or y[n] is described by an infinite sequence of values, i.e., the time index n takes values in � ∞ to +∞. The above picture is a snapshot at a particular time n. In the diagram above, the sequence of output values y[.] is the response of system S to the input sequence x[.] The system is causal if y[k] depends only on x[j] for j≤k **From before the modulator till after the demodulator & filter 6.02 Fall 2012 Lecture 10, Slide #3

  4. Time Invariant Systems Let y[n] be the response of S to input x[n]. If for all possible sequences x[n] and integers N x[n-N] y[n-N] S then system S is said to be time invariant (TI). A time shift in the input sequence to S results in an identical time shift of the output sequence. In particular, for a TI system, a shifted unit sample δ [ n − N ] function at the input generates an identically h [ n − N ] shifted unit sample response at the output. 6.02 Fall 2012 Lecture 10, Slide #4

  5. Linear Systems Let y 1 [n] be the response of S to an arbitrary input x 1 [n] and y 2 [n] be the response to an arbitrary x 2 [n]. If, for arbitrary scalar coefficients a and b, we have: ay 1 [ n ] + by 2 [ n ] ax 1 [ n ] + bx 2 [ n ] S then system S is said to be linear . If the input is the weighted sum of several signals, the response is the superposition (i.e., same weighted sum) of the response to those signals. One key consequence: If the input is identically 0 for a linear system, the output must also be identically 0. 6.02 Fall 2012 Lecture 10, Slide #5

  6. Unit Sample and Unit Step Responses Unit sample Unit sample response δ [n] h[n] S The unit sample response of a system S is the response of the system to the unit sample input. We will always denote the unit sample response as h[n]. Similarly, the unit step response s[n]: Unit step Unit step response u[n] s[n] S 6.02 Fall 2012 Lecture 10, Slide #6

  7. Relating h[n] and s[n] of an LTI System Unit sample signal Unit sample response δ [n] h[n] S Unit step signal Unit step response u[n] s[n] S δ [ n ] = u [ n ] − u [ n − 1] h [ n ] = s [ n ] − s [ n − 1] n ∑ s [ n ] = h [ k ] from which it follows that k =−∞ s [ −∞ ] = 0 (assuming , e.g., a causal LTI system; more generally, a “right-sided” unit sample response) 6.02 Fall 2012 Lecture 10, Slide #7

  8. h[n] s[n] 6.02 Fall 2012 Lecture 10, Slide #8

  9. h[n] s[n] 6.02 Fall 2012 Lecture 10, Slide #9

  10. h[n] s[n] 6.02 Fall 2012 Lecture 10, Slide #10

  11. Unit Step Decomposition “ “Rectangular-wave” digital signaling waveforms, of the sort s we have been considering, are w easily decomposed into time- e shifted, scaled unit steps --- each s t transition corresponds to another shifted, scaled unit step. s e e.g., if x[n] is the transmission of 1001110 using 4 samples/bit: 1 x [ n ] = u [ n ] − u [ n − 4] + u [ n − 12] − u [ n − 24] 6.02 Fall 2012 6.02 Fall 2012 Lecture 10, Slide #11

  12. … so the corresponding response is y [ n ] x [ n ] = s [ n ] = u [ n ] − s [ n − 4] − u [ n − 4] + s [ n − 12] + u [ n − 12] − s [ n − 24] − u [ n − 24] Note how we have invoked linearity and time invariance! 6.02 Fall 2012 Lecture 10, Slide #12

  13. Example 6.02 Fall 2012 Lecture 10, Slide #13

  14. Tr ansmission Over a Channel Ignore this notation for now, will explain shortly 6.02 Fall 2012 Lecture 10, Slide #14

  15. Receiv ing the Response Digitization threshold = 0.5V 6.02 Fall 2012 Lecture 10, Slide #15

  16. Faster Tr ansmission 6.02 Fall 2012 Fall 2012 Noise margin? 0.5 � y[28] Noise m Lecture 10, Slide #16 Lecture 10, Slid

  17. Unit Sample Decomposition A discrete-time signal can be decomposed into a sum of time-shifted, scaled unit samples. Example: in the figure, x[n] is the sum of x[-2] δ [n+2] + x[-1] δ [n+1] + … + x[2] δ [n-2]. In general: ∞ x [ n ] = ∑ x [ k ] δ [ n − k ] k =−∞ For any particular index, only l i one term of this sum is non-zero 6.02 Fall 6.02 Fall 2012 2012 Lecture 10, Slide #17

  18. Modeling LT I Systems If system S is both linear and time-invariant (LTI), then we can use the unit sample response to predict the response to any input waveform x[n]: Sum of shifted, scaled responses Sum of shifted, scaled unit samples ∞ ∞ ∑ ∑ x [ n ] = x [ k ] δ [ n − k ] y [ n ] = x [ k ] h [ n − k ] S k =−∞ k =−∞ CONVOLUTION SUM Indeed, the unit sample response h[n] completely characterizes the LTI system S, so you often see x[n] y[n] h[.] 6.02 Fall 2012 Lecture 10, Slide #18

  19. Convolution Evaluating the convolution sum ∞ y [ n ] = ∑ x [ k ] h [ n − k ] k =−∞ for all n defines the output signal y in terms of the input x and unit-sample response h. Some constraints are needed to ensure this infinite sum is well behaved, i.e., doesn’t “blow up” --- we’ll discuss this later. ∗ ∗ We use to denote convolution, and write y=x h. We can then write the value of y at time n, which is given by the above sum, y [ n ] = x ∗ h [ n ] y [ n ] = ( x ∗ h )[ n ] as . We could perh aps even write 6.02 Fall 2012 Lecture 10, Slide #19

  20. Convolution Evaluating the convolution sum ∞ y [ n ] = ∑ x [ k ] h [ n − k ] k =−∞ for all n defines the output signal y in terms of the input x and unit-sample response h. Some constraints are needed to ensure this infinite sum is well behaved, i.e., doesn’t “blow up” --- we’ll discuss this later. ∗ ∗ We use to denote convolution, and write y=x h. We can thus write the value of y at time n, which is given by the above sum, as y [ n ] = ( x ∗ h )[ n ] y [ n ] = x [ n ] ∗ h [ n ] Instead you’ll find people writing , where the poor index n is doing double or triple duty. This is awful notation, but a super-majority of engineering professors (including at MIT) will inflict it on their students. Don’t stand for it! 6.02 Fall 2012 Lecture 10, Slide #20

  21. Properties of Convolution ∞ ∞ ( x ∗ h )[ n ] ≡ ∑ x [ k ] h [ n − k ] = ∑ h [ m ] x [ n − m ] k =−∞ m =−∞ The second equality above establishes that convolution is commutative: x ∗ h = h ∗ x Convolution is associative: ( ) x ∗ ( h 1 ∗ h 2 ) = x ∗ h 1 ∗ h 2 Convolution is distributive: ( ) x ∗ h 1 + h 2 = ( x ∗ 1 ) + ( x ∗ h 2 ) h 6.02 Fall 2012 Lecture 10, Slide #21

  22. Series Interconnection of LT I Systems w[n] x[n] y[n] h 1 [.] h 2 [.] ( ) = h 2 ∗ h ( ) ∗ x y = h 2 ∗ w = h 2 ∗ h 1 ∗ x 1 x[n] y[n] (h 2 � h 1 )[.] x[n] y[n] (h 1 � h 2 )[.] x[n] y[n] h 2 [.] h 1 [.] 6.02 Fall 2012 Lecture 10, Slide #22

  23. Spot Quiz input response Unit step response: s[n] x[n] y[n] S 1 0.5 0 1 2 3 4 5 … n Find y[n]: x[n] 1. Write x[n] as a function of 1 unit steps 0.5 2. Write y[n] as a function of unit step responses 0 1 2 3 4 5 6 7 8 9 n 3. Draw y[n] 6.02 Fall 2012 Lecture 10, Slide #23

  24. MIT OpenCourseWare http://ocw.mit.edu 6.02 Introduction to EECS II: Digital Communication Systems Fall 201 2 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend