6.01 Course Notes
1) LTI Signals and Systems
Table of Contents
- 1) LTI Signals and Systems
- 1.1) Introduction
- 1.2) The Signals and Systems Abstraction
- 1.2.1) Modularity, primitives, and composition
- 1.2.2) Discrete-time signals and systems
- 1.2.3) Linear, Time-Invariant Systems
- 1.3) Discrete-time Signals
- 1.3.1) Unit sample signal
- 1.3.2) Signal Combinators
- 1.3.3) Algebraic Properties of Operations on Signals
- 1.4) Feedforward systems
- 1.5) Feedback Systems
- 1.5.1) Accumulator example
- 1.5.2) General form of LTI systems
- 1.5.3) System Functionals
- 1.5.4) Primitive Systems
- 1.5.5) Combining System Functionals
- 1.6) Predicting system behavior
- 1.6.1) First-order systems
- 1.6.2) Second-order systems
- 1.6.2.1) Additive decomposition
- 1.6.2.2) Complex poles
- 1.6.2.3) Polar representation of complex numbers
- 1.6.2.4) Complex modes
- 1.6.2.5) Additive decomposition with complex poles
- 1.6.2.6) Importance of magnitude and period
- 1.6.2.7) Poles and behavior: summary
- 1.6.3) Higher-order systems
- 1.6.4) Finding poles
- 1.6.5) Superposition
- 1.7) Summary of system behavior
1.1) Introduction
Imagine that you are asked to design a system to steer a car straight down the middle of a lane. It seems easy, right? You can figure out some way to sense the position of the car within its lane. Then, if the car is right of center, turn the steering wheel to the left. As the car moves so that it is less to the right, turn the steering wheel less to the left. If it is left of center, turn the steering wheel to the right. This sort of proportional controller works well for many applications, but not for steering, as can be seen below.
It is relatively easy to describe better algorithms in terms that humans would understand: e.g., Stop turning back and forth! It is not so easy to specify exactly what one might mean by that, in a way that it could be automated. In this chapter, we will develop a Signals and Systems framework to facilitate reasoning about the dynamic behaviors of systems. This framework will enable construction of simple mathematical models that are useful in both analysis and design of a wide range of systems, including the car-steering system.
1.2) The Signals and Systems Abstraction
To think about dynamic behaviors of systems, we need to think not only about how to describe the system but also about how to describe the signals that characterize the inputs and outputs of the system, as illustrated below.
This diagram represents a system with one input and one output. Both the input and output are signals. A signal is a mathematical function with an independent variable (most often it will be time for the problems that we will study) and a dependent variable (that depends on the independent variable). The system is described by the way that it transforms the input signal into the output signal. In the simplest case, we might imagine that the input signal is the time sequence of steering-wheel angles (assuming constant speed) and that the output signal is the time sequence of distances between the center of the car and the midline of the lane.
Representing a system with a single input signal and a single output signal seems too simplistic for any real application. For example, the car in the steering example (in the figure above) surely has more than one possible output signal.List at least four possible output signals for the car-steering problem.
Show/Hide
Possible output signals include
- its three-dimensional position (which could be represented by a 3D vector $\hat{p}(t)$ or by three scalar functions of time),
- its angular position,
- the rotational speeds of the wheels,
- the temperature of the tires, and many other possibilities.
The important point is that the first step in using the signals and systems representation is abstraction: we must choose the output(s) that are most relevant to the problem at hand and abstract away the rest. To understand the steering of a car, one vital output signal is the lateral position p_ o(t) within the lane, where p_ o(t) represents the distance (in meters) from the center of the lane. That signal alone tells us a great deal about how well we are steering. Consider a plot of p_ o(t) that corresponds to the figure above, as follows.
The oscillations in p_ o(t) as a function of time correspond to the oscillations of the car within its lane. Thus, this signal clearly represents an important failure mode of our car steering system. Is p_ o(t) the only important output signal from the car-steering system? The answer to this question depends on your goals. Analyzing a system with this single output is likely to give important insights into some systems (e.g., low-speed robotic steering) but not others (e.g., NASCAR). More complicated applications may require more complicated models. But all useful models focus on the most relevant signals and ignore those of lesser significance1. Throughout this chapter, we will focus on systems with one input signal and one output signal. When multiple output signals are important for understanding a problem, we will find that it is possible to generalize the methods and results developed here for single-input and single-output systems to systems with multiple inputs and outputs. The signals and systems approach has very broad applicability: it can be applied to mechanical systems (such as mass-spring systems), electrical systems (such as circuits and radio transmissions), financial systems (such as markets), and biological systems (such as insulin regulation or population dynamics). The fundamental notion of signals applies no matter what physical substrate supports them: it could be sound or electromagnetic waves or light or water or currency value or blood sugar levels.
1.2.1) Modularity, primitives, and composition
The car-steering system can be analyzed by thinking of it as the combination of car and steering sub-systems. The input to the car is the angle of the steering wheel. Let's call that angle \phi (t) . The output of the car is its position in the lane, p_ o(t), measured as the lateral distance to the center of the lane.
The steering controller turns the steering wheel to compensate for differences between our desired position in the lane, p_ i(t) (which is zero since we would like to be in the center of the lane), and our actual position in the lane p_ o(t). Let e(t)=p_ i(t)-p_ o(t). Thus, we can think about the steering controller as having an input e(t) and output \phi (t).
In the composite system (below), the steering controller determines \phi (t), which is the input to the car. The car generates p_ o(t), which is subtracted from p_ i(t) to get e(t), which is the input to the steering controller. The triangular component is called a gain or scale of -1: its output is equal to -1 times its input. More generally, we will use a triangle symbol to indicate that we are multiplying all the values of the signal by a numerical constant, which is shown inside the triangle.
The dashed-red box in the figure above illustrates modularity of the signals and systems abstraction. Three single-input, single-output sub-systems (steering controller, car, and inverter) and an adder (two inputs and 1 output) are combined to generate a new single-input (p_ i(t)), single-output (p_ o(t)) system. By abstraction, we could treat this new system as a primitive (represented by a single-input single-output box) and combine it with other subsystems to create a new, more complex, system. A principal goal of this chapter is to develop methods of analysis for the sub-systems that can be combined to analyze the overall system.
1.2.2) Discrete-time signals and systems
This chapter focuses on signals whose independent variables are discrete (e.g., take on only integer values). Some such signals are found in nature. For example, the primary structure of DNA is described by a sequence of base-pairs. However, we are primarily interested in discrete-time signals, not so much because they are found in nature, but because they are found in computers. Even though we focus on interactions with the real world, these interactions will primarily occur at discrete instants of time. For example, the difference between our desired position p_ i(t) and our actual position p_ o(t) is an error signal e(t), which is a function of continuous time t. If the controller only observes this signal at regular sampling intervals T, then its input could be regarded as a sequence of values x[n] that is indexed by the integer n. The relation between the discrete-time sequence x[n] (note square brackets) and the continuous signal x(t) (note round brackets) is given by x[n] = x(nT), where T is the length of one timestep, which we call the sampling relation. Sampling converts a signal of continuous domain to one of discrete domain. While our primary focus will be on time signals, sampling works equally well in other domains. For example, images are typically represented as arrays of pixels accessed by integer-valued rows and columns, rather than as continuous brightness fields, indexed by real-valued spatial coordinates. If the car-steering problem in the beginning were modeled in discrete time, we could describe the system with a diagram that is very similar to the continuous-time diagram above. However, only discrete time instants are considered and the output position is now only defined at discrete times, as shown below.
1.2.3) Linear, Time-Invariant Systems
We will be focusing our attention on a specific class of systems: discrete-time linear time-invariant (LTI) systems. In the first few labs, we will see how it is possible to simulate these kinds of systems; what's more, we will see that limiting our attention to this subclass of systems will allow us to predict the behavior of such systems without even simulating them, which will allow deeper forms of analysis. In an LTI system:
- The output at time n is a linear combination of previous and current inputs, and previous outputs.
- The behavior of the system does not depend on the absolute time at which it was started.
We are particularly interested in LTI systems because they can be analyzed mathematically, in a way that lets us characterize some properties of their output signal for any possible input signal. This is a much more powerful kind of insight than can be gained by trying a machine out with several different inputs. Another important property of LTI systems is that they are compositional: the cascade, parallel, and feedback combinations of LTI system are themselves LTI systems.
1.3) Discrete-time Signals
In this section, we will work through the PCAP system for discrete time signals, by introducing a primitive and three methods of composition, and the ability to abstract by treating composite signals as if they themselves were primitive. A signal is an infinite sequence of sample values at discrete time steps. We will use the following common notational conventions: A capital letter X stands for the whole input signal and x[n] stands for the value of signal X at time step n
- It is conventional, if there is a single system under discussion, to use X for the input signal to that system and Y for the output signal.
We will say that systems transduce input signals into output signals.
1.3.1) Unit sample signal
We will work with a single primitive, called the unit sample signal, \Delta . It is defined on all positive and negative integer indices as follows2:
That is, it has value 1 at index n=0 and 0 otherwise, as shown below:
1.3.2) Signal Combinators
1.3.2.1) Scaling
Our first operation will be scaling, or multiplication by a scalar. A scalar is any real number. The result of multiplying any signal X by a scalar c is a signal, so that,
if Y = c \cdot X, then y[n] = c \cdot x[n]
That is, the resulting signal has a value at every index n that is c times the value of the original signal at that location. Here are the signals 4 \Delta and -3.3 \Delta:
1.3.2.2) Delay
The next operation is the delay operation. The result of delaying a signal X is a new signal \mathcal{R} X such that:if Y = \mathcal{R} X then y[n] = x[n-1].
That is, the resulting signal has the same values as the original signal, but delayed by one step in time. You can also think of this, graphically, as shifting the signal one step to the Right. Here is the unit sample delayed by 1 and by 3 steps. We can describe the second signal as \mathcal{R} \mathcal{R} \mathcal{R} \Delta, or, using shorthand, as \mathcal{R} ^3\Delta.
1.3.2.3) Addition
Finally, we can add two signals together. Addition of signals is accomplished component-wise, so that:
if Y = X_1 + X_2 then y[n] = x_1[n] + x_2[n]
That is, the value of the composite signal at step n is the sum of the values of the component signals. Here are some new signals constructed by summing, scaling, and delaying the unit sample.
Note that, because each of our operations returns a signal, we can use their results again as elements in new combinations, showing that our system has true compositionality. In addition, we can abstract, by naming signals. So, for example, we might define Y = 3\Delta + 4\mathcal{R} \Delta - 2\mathcal{R} ^2\Delta, and then make a new signal Z = Y + 0.3\mathcal{R} Y, which would look like this:
Be sure you understand how the heights of the spikes are determined by the definition of Z .
Draw a picture of samples -1 through 4 of Y - \mathcal{R} Y.
It is important to remember that, because signals are infinite objects, these combination operations are abstract mathematical operations. You could never somehow 'make' a new signal by calculating its value at every index. It is possible, however, to calculate the value at any particular index, as it is required.
1.3.3) Algebraic Properties of Operations on Signals
Adding and scaling satisfy the familiar algebraic properties of addition and multiplication: addition is commutative and associative, scaling is commutative (in the sense that it doesn't matter whether we pre- or post-multiply) and scaling distributes over addition:
which can be verified by defining Y = c\cdot (X_1+X_2) and Z = c\cdot X_1 + c \cdot X_2 and checking that y[n] = z[n] for all n:
which clearly holds based on algebraic properties of arithmetic on real numbers. In addition, \mathcal{R} distributes over addition and scaling, so that:
Verify that \mathcal{R} distributes over addition and multiplication by checking that the appropriate relations hold at some arbitrary step n.
These algebraic relationships mean that we can take any finite expression involving \Delta, \mathcal{R} + and \cdot and convert it into the form
That is, we can express the entire signal as a polynomial in \mathcal{R}, applied to the unit sample. In our previous example, it means that we can rewrite 3\Delta + 4\mathcal{R} \Delta - 2\mathcal{R} ^2\Delta as (3 + 4\mathcal{R} - 2\mathcal{R} ^2)\Delta.
1.4) Feedforward systems
We will start by looking at a subclass of discrete-time LTI system, which are exactly those that can be described as performing some combination of scaling, delay, and addition operations on the input signal. We will develop several ways of representing such systems, and see how to combine them to get more complex systems in this same class.1.4.1) Representing systems
We can represent systems using operator equations, difference equations, block diagrams, and Python state machines. Each makes some things clearer and some operations easier. It is important to understand how to convert between the different representations.1.4.1.1) Operator equation
An operator equation is a description of how signals are related to one another, using the operations of scaling, delay, and addition on whole signals. Consider a system that has an input signal X, and whose output signal is X - \mathcal{R} X. We can describe that system using the operator equationFeedforward systems can always be described using an operator equation of the form Y = \Phi X, where \Phi is a polynomial in \mathcal{R}.
1.4.1.2) Difference Equation
An alternative representation of the relationship between signals is a difference equation. A difference equation describes a relationship that holds among samples (values at particular times) of signals. We use an index n in the difference equation to refer to a particular time index, but the specification of the corresponding system is that the difference equation hold for all values of n- The operator equation Y = X - \mathcal{R} Xcan be expressed as this equivalent difference equation:y[n] = x[n] - x[n-1]The operation of delaying a signal can be seen here as referring to a sample of that signal at time step n-1
- Difference equations are convenient for step-by-step analysis, letting us compute the value of an output signal at any time step, given the values of the input signal.
So, if the input signal X is the unit sample signal,
then, using a difference equation, we can compute individual values of the output signal Y:
1.4.1.3) Block Diagrams
Another way of describing a system is by drawing a block diagram, which is made up of components,connected by lines with arrows on them. The lines represent signals; all lines that are connected to one another (not going through a round, triangular, or circular component) represent the same signal.The components represent systems. There are three primitive components corresponding to our operations on signals:
- Delay components are drawn as rectangles, labeled Delay, with two lines connected to them, one with an arrow coming in and one going out. If X is the signal on the line coming into the delay, then the signal coming out is RX.
- Scale(or gain) components are drawn as triangles, labeled with a positive or negative number c, with two lines connected to them, one with an arrow coming in and one going out. If X is the signal on the line coming into the gain component, then the signal coming out is c \cdot X
- Adder components are drawn as circles, labeled with+, three lines connected to them, two with arrows coming in and one going out. If X_1 and X_2 are the signals on the lines point into the adder, then the signal coming out is X_1+X_2.
For example, the system Y = X-{\cal R}X can be represented with this block diagram:
1.4.2) Combinations of systems
To combine LTI systems, we will use the same cascade and parallel-add operations as we had for state machines.1.4.2.1) Cascade multiplication
When we make a cascade combination of two systems, we let the output of one system be the input of another. So, if the system M_1 has operator equation Y = \Phi _1X and system M_2 has operator equation Z = \Phi _2 W , and then we compose M_1 and M_2 in cascade by setting Y = W , then we have a new system, with input signal X , output signal Z , and operator equation Z = (\Phi _2 \cdot \Phi _1) X- The product of polynomials is another polynomial, so \Phi _2 \cdot \Phi _1 is a polynomial in \mathcal{R}
- Furthermore, because polynomial multiplication is commutative, cascade combination is commutative as well (as long as the systems are at rest, which means that their initial states are 0).
So, for example,
Cascade combination, because it results in multiplication, is also associative, which means that any grouping of cascade operations on systems has the same result.
1.4.2.2) Parallel addition
When we make a parallel addition combination of two systems, the output signal is the sum of the output signals that would have resulted from the individual systems. So, if the system M_1 is represented by the equation Y = \Phi _1X and system M_2 is represented by the equation Z = \Phi _2 X, and then we compose M_1 and M_2 with parallel addition by setting output W = Y + Z, then we have a new system, with input signal X, output signal W, and operator equation W = (\Phi _1 + \Phi _2) X. Because addition of polynomials is associative and commutative, then so is parallel addition of feed-forward linear systems.1.4.2.3) Combining cascade and parallel operations
Finally, the distributive law applies for cascade and parallel combination, for systems at rest, in the same way that it applies for multiplication and addition of polynomials, so that if we have three systems, with operator equations:and we form a cascade combination of the sum of the first two, with the third, then we have a system describable as:
So, for example,
Here is another example of two equivalent operator equations
Convince yourself that all of these systems are equivalent. One strategy is to convert them all to operator equation representation.
1.5) Feedback Systems
So far, all of our example systems have been _feedforward_: the dependencies have all flowed from the input through to the output, with no dependence of an output on previous output values. In this section, we will extend our representations and analysis to handle the general class of LTI systems in which the output can depend on any finite number of previous input or output values.1.5.1) Accumulator example
Consider this block diagram, of an accumulator:It's reasonably straightforward to look at this block diagram and see that the associated difference equation is
We immediately run up against a question: what is the value of y[n-1]? The answer clearly has a profound effect on the output of the system. In our treatment of feedback systems, we will generally assume that they start 'at rest', which means that all values of the inputs and outputs at steps less than 0 are 0. That assumption lets us fill in the following table:
Here are plots of the input signal X and the output signal Y:
This result may be somewhat surprising! In feedforward systems, we saw that the output was always a finite sum of scaled and delayed versions of the input signal; so that if the input signal was transient (had a finite number of non-zero samples) then the output signal would be transient as well. But, in this feedback system, we have a transient input with a persistent (infinitely many non-zero samples) output. We can also look at the operator equation for this system. Again, reading it off of the block diagram, it seems like it should be
It's a well-formed equation, but it isn't immediately clear how to use it to determine Y. Using what we already know about operator algebra, we can rewrite it as:
which defines Y to be the signal such that the difference between Y and \mathcal{R} Y is X. But how can we find that Y?
We will now show that we can think of the accumulator system as being equivalent to another system, in which the output is the sum of infinitely many feedforward paths, each of which delays the input by a different, fixed value. This system has an operator equation of the form
These systems are equivalent in the sense that if each is initially at rest, they will produce identical outputs from the same input. We can see this by taking the original definition and repeatedly substituting in the definition of Y in for its occurrence on the right hand side:
Now, we can informally derive a 'definition' of the reciprocal of 1 - \mathcal{R} (the mathematical details underlying this are subtle and not presented here),
In the following it will help to remind ourselves of the derivation of the formula for the sum of an infinite geometric series:
Subtracting the second equation from the first we get
And so, provided |x| \lt 1,
Similarly, we can consider the system O, where
And so,
Check this derivation by showing that
So, we can rewrite the operator equation for the accumulator as
1.5.2) General form of LTI systems
We can now treat the general case of LTI systems, including feedback. In general, LTI systems can be described by difference equations of the form:
1.5.3) System Functionals
Now, we are going to engage in a shift of perspective. We started by defining a new signal Y in terms of an old signal X, much as we might, in algebra, define y = x + 6. Sometimes, however, we want to speak of the relationship between x and y in the general case, without a specific x or y in mind. We do this by defining a function f: f(x) = x + 6
We can do the same thing with LTI systems, by defining system functionals. If we take the general form of an LTI system given in the previous section and write it as an operator equation, we have
We can rewrite this as
The system functional is most typically written in the form
1.5.4) Primitive Systems
Just as we had a PCAP system for signals, we have one for LTI system, in terms of system functionals, as well. We can specify system functionals for each of our system primitives. A gain element is governed by operator equation Y = kX, for constant k, so its system functional is1.5.5) Combining System Functionals
We have three basic composition operations: sum, cascade, and feedback. This PCAP system, as our previous ones have been, is _ compositional_, in the sense that whenever we make a new system functional out of existing ones, it is a system functional in its own right, which can be an element in further compositions.1.5.5.1) Addition
The system functional of the sum of two systems is the sum of their system functionals. So, given two systems with system functionals {\cal H}_1and {\cal H}_2, connected like this:
and letting
where {\cal H} = {\cal H}_1 + {\cal H}_2.
1.5.5.2) Cascade
The system functional of the cascade of two systems is the product of their system functionals. So, given two systems with system functionals {\cal H}_1 and {\cal H}_2, connected like this:
and letting
where {\cal H} = {\cal H}_2 {\cal H}_1. And note that, as was the case with purely feedforward systems, cascade combination is still commutative, so it doesn't matter whether {\cal H}_1 or {\cal H}_2 comes first in the cascade. This surprising fact holds because we are only considering LTI systems starting at rest; for more general classes of systems, such as the general class of state machines we have worked with before, the ordering of a cascade does matter.
1.5.5.3) Feedback
There are several ways of connecting systems in feedback. Here we study a particular case of negative feedback combination, which results in a classical formula called Black's formula. Consider two systems connected like this
and pay careful attention to the negative sign on the feedback input to the addition. It is really just shorthand; the negative sign could be replaced with a gain component with value -1. This negative feedback arrangement is frequently used to model a case in which X is a desired value for some signal and Y is its actual value; thus the input to {\cal H}_1 is the difference between the desired an actual values, often called an error signal. We can simply write down the operator equation governing this system and use standard algebraic operations to determine the system functional:
where \displaystyle{{\cal H} = \frac{{\cal H}_1}{1 + {\cal H}_1 {\cal H}_2}}
Armed with this set of primitives and composition methods, we can specify a large class of systems.
1.6) Predicting system behavior
We have seen how to construct complex discrete-time LTI systems; in this section we will see how we can use properties of the system functional to predict how the system will behave, in the long term, and for any input. We will start by analyzing simple systems and then move to more complex ones. We can provide a general characterization of the long-term behavior of the output, as increasing or decreasing, with constant or alternating sign, for any finite input to the system. We will begin by studying the unit-sample response of systems, and then generalize to more general input signals; similarly, we will begin by studying simple systems and generalize to more complex ones.
1.6.1) First-order systems
Systems that only have forward connections can only have a finite response; that means that if we put in a unit sample (or other signal with only a finite number of non-zero samples) then the output signal will only have a finite number of non-zero samples. Systems with feedback have a surprisingly different character. Finite inputs can result in persistent response; that is, in output signals with infinitely many non-zero samples. Furthermore, the qualitative long-term behavior of this output is generally independent of the particular input given to the system, for any finite input. In this section, we will consider the class of first-order systems, in which the denominator of the system functional is a first-order polynomial (that is, it only involves \mathcal{R}, but not \mathcal{R}^2 or other higher powers of \mathcal{R}.)
Let's consider this very simple system
for which we can write an operator equation
and derive a system functional
Recall the infinite series representation of this system functional:
We can make intuitive sense of this by considering how the signal flows through the system. On each step, the output of the system is being fed back into the input. Consider the simple case where the input is the unit sample (X = \Delta). Then, after step 0, when the input is 1, there is no further input, and the system continues to respond. In this table, we see that the whole output signal is a sum of scaled and delayed copies of the input signal; the bottom row of figures shows the first three terms in the infinite sum of signals, for the case where p_0 = 0.9.
\frac{Y}{X} = 1 + \ldots | \frac{Y}{X} = 1 + p_0\mathcal{R} + \ldots | \frac{Y}{X} = 1 + p_0\mathcal{R} + p_0^2\mathcal{R}^2 + \ldots |
\Delta |
0.9\mathcal{R}\Delta |
0.81\mathcal{R}^2\Delta |
The overall response of this system is shown below:
\Delta + 0.9\mathcal{R}\Delta + 0.81\mathcal{R}^2\Delta + 0.729\mathcal{R}^3\Delta + 0.6561\mathcal{R}^3\Delta + \ldots
For the first system, the unit sample response is y[n] = (0.5)^ n; for the second, it's y[n] = (1.2)^ n. These system responses can be characterized by a single number, called the pole, which is the base of the geometric sequence. The value of the pole, p_0, determines the nature and rate of growth.
- If p_0 \lt-1, the magnitude increases to infinity and the sign alternates.
- If -1\lt p_0\lt 0, the magnitude decreases and the sign alternates.
- If 0\lt p_0\lt 1, the magnitude decreases monotonically.
- If p_0>1, the magnitude increases monotonically to infinity.
Any system for which the magnitude of the unit sample response increases to infinity, whether monotonically or not, is called unstable. Systems whose unit sample response decreases or stays constant in magnitude are called stable.
Because we are dealing with linear, time-invariant systems, these definitions hold not only for the unit sample response, but for the response of the system to any transient input (i.e., any input with finitely-many non-zero values).
|
|
|
|
1.6.2) Second-order systems
We will call these persistent long-term behaviors of a signal (and, hence, of the system that generates such signals) modes. For a fixed p_0, the first-order system only exhibited one mode (but different values of p_0 resulted in very different modes). As we build more complex systems, they will have multiple modes, which manifest as more complex behavior. Second-order systems are characterized by a system functional whose denominator polynomial is second order; they will generally exhibit two modes. We will find that it is the mode whose pole has the largest magnitude that governs the long-term behavior of the system. We will call this pole the dominant pole. Consider this system:
We can describe it with the operator equation

We can try to understand its behavior by decomposing it in different ways. First, let's see if we can see it as a cascade of two systems. To do so, we need to find {\cal H}_1 and {\cal H}_2 such that {\cal H}_1 {\cal H}_2 = {\cal H}. We can do that by factoring {\cal H} to get
1.6.2.1) Additive decomposition
Another way to try to decompose the system is as the sum of two simpler systems. In this case, we seek {\cal H}_1 and {\cal H}_2 such that {\cal H}_1 + {\cal H}_2 = {\cal H}. We can do a partial fraction decomposition (don't worry if you don't remember the process for doing this...we won't need to solve problems like this in detail). We start by factoring, as above, and then figure out how to decompose into additive terms:To find values for A and B, we start with
Equating the terms that involve equal powers of \mathcal{R} (including constants as terms that involve \mathcal{R}^0), we have:
Solving, we find A = 4.5 and B = -3.5, so
Verify that {\cal H}_1 + {\cal H}_2 = {\cal H}.
Here is (yet another) equivalent block diagram for this system, highlighting its decomposition into a sum:
We can understand this by looking at the responses of {\cal H}_1 and of {\cal H}_2 to the unit sample, and then summing them to recover the response of H to the unit sample. In the next figure, the blue stem plot is the overall signal, which is the sum of the green and red signals, which correspond to the top and bottom parts of the diagram, respectively.

In this case, both of the poles (0.9 and 0.7) are less than 1, so the magnitude of the responses they generate decreases monotonically; their sum does not behave monotonically, but there is a time step at which the dominant pole completely dominates the other one, and the convergence is monotonic after that. If, instead, we had a system with system functional

Here is a plot of the output of the system described by

The green dots are generated by the component with pole 0.9, the red dots are generated by the component with pole -1.1 and the blue dots are generated by the sum.
In these examples, as in the general case, the long-term unit-sample response of the entire system is governed by the unit-sample response of the mode whose pole has the larger magnitude. We call this pole the dominant pole. In the long run, the rate of growth of the exponential with the largest exponent will always dominate.
1.6.2.2) Complex poles
Consider a system described by the operator equation:
But now, if we attempt to perform an additive decomposition on it, we find that the denominator cannot be factored to find real poles. Instead, we find that
So, the poles are approximately -1 + 1.732 j and -1 - 1.732 j. Note that we are using j to signal the imaginary part of a complex number3. What does that mean about the behavior of our system? Are the outputs real or complex? Do the output values grow or shrink with time? Difference equations that represent physical systems have real-valued coefficients. For instance, a bank account with interest \rho might be described with difference equation
The position of a robot moving toward a wall might be described with difference equation:
1.6.2.3) Polar representation of complex numbers
Sometimes it's easier to think about a complex number a + bj instead as r e^{j\Omega }, whereSo, if we think of (a, b) as a point in the complex plane, then (r, \Omega ) is its representation in polar coordinates.

which can be directly derived from series expansions of e^ z, \sin z and \cos z. To see that this is reasonable, let's take our number, represent it as a complex exponential, and then apply Euler's equation:
Why should we bother with this change of representation? There are some operations on complex numbers that are much more straightforwardly done in the exponential representation. In particular, let's consider raising a complex number to a power. In the Cartesian representation, we get complex trigonometric polynomials. In the exponential representation, we get, in the quadratic case,
More generally, we have that
1.6.2.4) Complex modes
Now, we're equipped to understand how complex poles of a system generate behavior: they produce complex-valued modes. Remember that we can characterize the behavior of a mode asWhat happens as n tends to infinity when p is complex? Think of p^ n as a point in the complex plane with coordinates (r^ n, \Omega n). The radius, r^ n, will grow or shrink depending on the mode's magnitude, r. And the angle, \Omega n, will simply rotate, but will not affect the magnitude of the resulting value. Note that each new point in the sequence p^ n will be rotated by \Omega from the previous one. We will say that the period of the output signal is 2\pi / \Omega ; that is the number of samples required to move through 2 \pi radians (although this need not actually be an integer). The sequence spirals around the origin of the complex plane. Here is a case for r = 0.85, \Omega = \pi /4, that spirals inward, rotating \pi /4 radians on each step:

p = 0.85 e^{j \pi / 4} \approx 0.601 + 0.601 j
Here, for example, are plots of two other complex modes, with different magnitudes and rates of oscillation. In all cases, we are plotting the unit-sample response, so the first element in the series is the real value 1. The first signal clearly converges in magnitude to 0; the second one diverges (look carefully at the scales).

p = 0.85 e^{j \pi / 8} \approx 0.785 + 0.325 j

p = 1.1 e^{j \pi / 8} \approx 1.016 + 0.421 j
If complex modes occurred all by themselves, then our signals would have complex numbers in them. But isolated complex pole can result only from a difference equation with complex-valued coefficients. For example, to end up with this system functional
where the only difference is in the sign of the angular part. If we look at a second-order system with complex-conjugate poles, the resulting polynomial has real-valued coefficients. To see this, consider a system with poles re^{j\Omega } and re^{-j\Omega }, so that
Using the definition of the complex exponential e^{jx} = \cos x + j \sin x,
Using trigonometric identities \sin -x = -\sin x and \cos -x =\cos x,
So,
1.6.2.5) Additive decomposition with complex poles
You can skip this section if you want to. To really understand these complex modes and how they combine to generate the output signal, we need to do an additive decomposition. That means doing a partial fraction expansion ofwhich is entirely real. The figure below shows the modes for a system with poles 0.85 e^{j\pi / 8} and 0.85 e^{-j \pi / 8} : the blue series starts at \frac{1}{2}(1 - j\cot (\pi /8)) and traces out the unit sample response of the first component; the second red series starts at \frac{1}{2}(1 + j \cot (\pi /8)) and traces out the unit sample response of the second component.


1.6.2.6) Importance of magnitude and period
Both r and \Omega tells us something very useful about the way in which the system behaves. In the previous section, we derived an expression for the samples of the unit sample response for a system with a pair of complex poles. It has the form
The value of r governs the rate of exponential decrease. The value of \Omega governs the rate of oscillation of the curve. It will take 2 \pi / \Omega (the period of the oscillation) samples to go from peak to peak of the oscillation4. In our example, \Omega = \pi / 8 so the period is 16; you should be able to count 16 samples from peak to peak.
1.6.2.7) Poles and behavior: summary
In a second-order system, if we let p_0 be the pole with the largest magnitude, then there is a time step at which the behavior of the dominant pole begins to dominate; after that time step
- If p_0 is real and
- p_0\lt -1, the magnitude increases to infinity and the sign alternates.
- -1\lt p_0\lt 0, the magnitude decreases and the sign alternates.
- 0\lt p_0\lt 1, the magnitude decreases monotonically.
- p_0\gt 1, the magnitude increases monotonically to infinity.
- If p_0 is complex and \Omega is its angle, then the signal will be periodic, with period 2 \pi / \Omega .
- If p_0 is complex
- and |p_0| \lt 1, the amplitude of the response decays to 0.
- and |p_0| \gt 1, the amplitude of the response increases to infinity.
Remember that any system for which the output magnitude increases to infinity, whether monotonically or not, is called unstable. Systems whose output magnitude decreases or stays constant are called stable. As we have seen in our examples, when we add multiple modes, it is the mode with the largest pole, that is, the dominant pole that governs the long-term behavior.
1.6.3) Higher-order systems
Recall that we can describe any system in terms of a system functional that is the ratio of two polynomials in \mathcal{R} (and assuming a_0 = 1):Returning to the general polynomial ratio
1.6.4) Finding poles
In general, we will find that if the denominator of the system functional {\cal H} is a k th order polynomial, then it can be factored into the form (1 - p_0\mathcal{R})(1 - p_1\mathcal{R})\ldots (1 - p_{k-1}\mathcal{R}). We will call the p_ i values the poles of the system. The entire persistent output of the system can be expressed as a scaled sum of the signals arising from each of these individual poles. We're doing something interesting here! We are using the PCAP system backwards for analysis. We have a complex thing that is hard to understand monolithically, so we are taking it apart into simpler pieces that we do understand. It might seem like factoring polynomials in this way is tricky, but there is a straightforward way to find the poles of a system given its denominator polynomial in \mathcal{R}. We'll start with an example. Assume the denominator is 12\mathcal{R}^2 - 7 \mathcal{R} + 1. If we play a quick trick, and introduce a new variable z = 1 / \mathcal{R}, then our denominator becomes
1.6.4.1) Pole-Zero cancellation
For some systems, we may find that it is possible to cancel matching factors from the numerator and denominator of the system functional. If the factors represent poles that are not at zero, then although it may be theoretically supportable to cancel them, it is unlikely that they would match exactly in a real system. If we were to cancel them in that case, we might end up with a system model that was a particularly poor match for reality5. So, we will only consider canceling poles and zeros at zero. In this example:1.6.4.2) Repeated roots
In some cases, the equation in z will have repeated roots. For example, z^2 - z + 0.25, which has two roots at 0.5. In this case, the system has a repeated pole; it is still possible to perform an additive decomposition, but it is somewhat trickier. Ultimately, however, it is still the magnitude of the largest root that governs the long-term convergence properties of the system.1.6.5) Superposition
The principle of superposition states that the response of a LTI system to a sum of input signals is the sum of the responses of that system to the components of the input. So, given a system with system functional {\cal H}, and input X = X_1 + X_2,We can express U as an infinite sum of increasingly delayed unit-sample signals:
The response of a system to U will therefore be an infinite sum of unit-sample responses. Let Z = {\cal H} \Delta be the unit-sample response of {\cal H}.
Then:
Let's consider the case where {\cal H} is a first-order system with a pole at p. Then,
It's clear that, if |p| \gt = 1 then y[n] will grow without bound; but if 0 \lt p \lt 1 then, as n goes to infinity, y[n] will converge to 1 / (1 - p). We won't study this in any further detail, but it's useful to understand that the basis of our analysis of systems applies very broadly across LTI systems and inputs.
1.7) Summary of system behavior
Here is some terminology that will help us characterize the long-term behavior of systems.- A signal is transient if it has finitely many non-zero samples. Otherwise, it is persistent.
- A signal is bounded if there is exist upper and lower bound values such that the samples of the signal never exceed those bounds; item otherwise it is unbounded.
Now, using those terms, here is what we can say about system behavior:
- A transient input to an acyclic (feed-forward) system results in a transient output.
- A transient input to a cyclic (feed-back) system results in a persistent output.
- The poles of a system are the roots of the denominator polynomial of the system functional written in terms of z = 1/\mathcal{R}.
- The dominant pole is the pole with the largest magnitude.
- If the dominant pole has magnitude \gt 1, then in response to a bounded input, the output signal will be unbounded.
- If the dominant pole has magnitude \lt 1, then in response to a bounded input, the output signal will be bounded; in response to a transient input, the output signal will converge to 0.
- If the dominant pole has magnitude 1, then in response to a bounded input, the output signal will be bounded; in response to a transient input, it will converge to some constant value.
- If the dominant pole is real and positive, then in response to a transient input, the signal will, after finitely many steps, begin to increase or decrease monotonically.
- If the dominant pole is real and negative, then in response to a transient input, the signal will, after finitely many steps, begin to alternate signs (or, to say it another way, the signal will start to oscillate with a period of 2 timesteps).
- If the dominant pole is complex, then in response to a transient input, the signal will, after finitely many steps, begin to be periodic, with a period of 2 \pi / \Omega , where \Omega is the 'angle' of the pole.
Footnotes
1There are always unimportant outputs, and which outputs we care about depend on the system we are desigining or analyzing. Think about the number of moving parts in a car; they are not all important for steering!
2Note that \delta is the lowercase version of \Delta, both of which are the Greek letter "delta."
3We use j instead of i because, to an electrical engineer, i stands for current, and cannot be re-used!
4We are being informal here, in two ways. First, the signal does not technically have a period, because unless r=1, it doesn't return to the same point. Second, unless r=1, then distance from peak to peak is not exactly 2\pi / \Omega; however, for most signals, it will give a good basis for estimating \Omega
5Don't worry too much if this doesn't make sense to you. Take 6.003 and/or 6.302 to learn more!