# Signals and Systems/Time Domain Analysis

There are many tools available to analyze a system in the time domain, although many of these tools are very complicated and involved. Nonetheless, these tools are invaluable for use in the study of linear signals and systems, so they will be covered here.

## Contents

## Linear Time-Invariant (LTI) Systems[edit]

This page will contain the definition of a LTI system and this will be used to motivate the definition of convolution as the output of a LTI system in the next section. To begin with a system has to be defined and the LTI properties have to be listed. Then, for a given input it can be shown (in this section or the following) that the output of a LTI system is a convolution of the input and the system's impulse response, thus motivating the definition of convolution.

Consider a system for which an input of *x _{i}(t)* results in an output of

*y*respectively for

_{i}(t)*i = 1, 2*.

### Linearity[edit]

There are 2 requirements for *linearity*. A function must satisfy all 2 to be called "linear".

**Additivity**: An input of results in an output of .**Homogeneity**: An input of results in an output of

Being linear is also known in the literature as "satisfying the principle of superposition". **Superposition** is a fancy term for saying that the system is additive and homogeneous. The terms linearity and superposition can be used interchangably, but in this book we will prefer to use the term linearity exclusively.

We can combine the two requirements into a single equation: In a linear system, an input of results in an output of .

### Additivity[edit]

A system is said to be additive if a sum of inputs results in a sum of outputs. To test for additivity, we need to create two arbitrary inputs, *x _{1}(t)* and

*x*. We then use these inputs to produce two respective outputs:

_{2}(t)Now, we need to take a sum of inputs, and prove that the system output is a sum of the previous outputs:

If this final relationship is not satisfied for *all possible inputs*, then the system is not additive.

### Homogeneity[edit]

Similar to additivity, a system is homogeneous if a scaled input (multiplied by a constant) results in a scaled output. If we have two inputs to a system:

Where

Where *c* is an arbitrary constant. If this is the case then the system is homogeneous if

for any arbitrary *c*.

### Time Invariance[edit]

If the input signal *x(t)* produces an output *y(t)* then any time shifted input, *x(t + δ)*, results in a time-shifted output *y(t + δ)*.

This property can be satisfied if the transfer function of the system is not a function of time except expressed by the input and output.

### Example: Simple Time Invariance[edit]

To demonstrate how to determine if a system is time-invariant, consider the two systems:

- System A:
- System B:

Since system A explicitly depends on *t* outside of *x(t)* and *y(t)*, it is **time-variant**. System B, however, does not depend explicitly on *t* so it is time-invariant (given *x(t)* is time-invariant).

### Example: Formal Proof[edit]

A more formal proof of why systems A & B from above are respectively time varying and time-invariant is now presented. To perform this proof, the second definition of time invariance will be used.

- System A
- Start with a time shift of the input
- Now describe the time shifted output by δ
- Clearly , therefore the system is not time-invariant.

- System B
- Start with a time shift of the input
- Now time shift the output by δ
- Clearly , therefore the system is time-invariant.

## Linear Time Invariant (LTI) Systems[edit]

The system is linear time-invariant (LTI) if it satisfies both the property of linearity and time-invariance. This book will study LTI systems almost exclusively, because they are the easiest systems to work with, and they are ideal to analyze and design.

## Other Function Properties[edit]

Besides being linear, or time-invariant, there are a number of other properties that we can identify in a function:

### Memory[edit]

A system is said to have memory if the output from the system is dependent on past inputs (or future inputs) to the system. A system is called *memoryless* if the output is only dependent on the current input. Memoryless systems are easier to work with, but systems with memory are more common in digital signal processing applications. A memory system is also called a dynamic system whereas a memoryless system is called a static system.

### Causality[edit]

Causality is a property that is very similar to memory. A system is called **causal** if it is only dependent on past or current inputs. A system is called **non-causal** if the output of the system is dependent on future inputs. Most of the practical systems are causal.

### Stability[edit]

Stability is a very important concept in systems, but it is also one of the hardest function properties to prove. There are several different criteria for system stability, but the most common requirement is that the system must produce a finite output when subjected to a finite input. For instance, if we apply 5 volts to the input terminals of a given circuit, we would like it if the circuit output didn't approach infinity, and the circuit itself didn't melt or explode. This type of stability is often known as "Bounded Input, Bounded Output" stability, or BIBO.

Studying BIBO stability is a relatively complicated course of study, and later books on the Electrical Engineering bookshelf will attempt to cover the topic.

## Linear Operators[edit]

Mathematical operators that satisfy the property of linearity are known as **linear operators**. Here are some common linear operators:

- Derivative
- Integral
- Fourier Transform

### Example: Linear Functions[edit]

Determine if the following two functions are linear or not:

## Impulse Response[edit]

**Impulse response** tells us how a system **reacts** when we hit it with an **impulse signal**(also called as **dirac delta function**).This Impulse response is very important term in analyzing the behaviour of systems.

### Zero-Input Response[edit]

### Zero-State Response[edit]

zero state response means steady state or forced response. This is the system response y(t) to an input f(t) when the system is in zero state; that is, when all initial conditions are zero.

### Second-Order Solution[edit]

- Example. Finding the total response of a driven RLC circuit.

## Convolution[edit]

Convolution (folding together) is a complicated operation involving integrating, multiplying, adding, and time-shifting two signals together. Convolution is a key component to the rest of the material in this book.

The convolution *a * b* of two functions *a* and *b* is defined as the function:

The greek letter τ (tau) is used as the integration variable, because the letter *t* is already in use. τ is used as a "dummy variable" because we use it merely to calculate the integral.

In the convolution integral, all references to *t* are replaced with τ, except for the *-t* in the argument to the function *b*. Function *b* is *time inverted* by changing τ to -τ. Graphically, this process moves everything from the right-side of the *y* axis to the left side and vice-versa. Time inversion turns the function into a mirror image of itself.

Next, function *b* is *time-shifted* by the variable *t*. Remember, once we replace everything with τ, we are now computing in the *tau domain*, and not in the time domain like we were previously. Because of this, *t* can be used as a shift parameter.

We multiply the two functions together, time shifting along the way, and we take the area under the resulting curve at each point. Two functions overlap in increasing amounts until some "watershed" after which the two functions overlap less and less. Where the two functions overlap in the *t* domain, there is a value for the convolution. If one (or both) of the functions do not exist over any given range, the value of the convolution operation at that range will be zero.

After the integration, the definite integral plugs the variable *t* back in for remaining references of the variable τ, and we have a function of *t* again. It is important to remember that the resulting function will be a combination of the two input functions, and will share some properties of both.

### Properties of Convolution[edit]

The convolution function satisfies certain conditions:

- Commutativity

- Associativity

- Distributivity

- Associativity With Scalar Multiplication

for any real (or complex) number *a*.

- Differentiation Rule

### Example 1[edit]

Find the convolution, *z(t)*, of the following two signals, *x(t)* and *y(t)*, by using (a) the integral representation of the convolution equation and (b) muliplication in the Laplace domain.

The signal *y(t)* is simply the Heaviside step, *u(t)*.

The signal *x(t)* is given by the following infinite sinusoid, *x _{0}(t)*, and windowing function,

*x*:

_{w}(t)Thus, the convolution we wish to perform is therefore:

From the distributive law:

## Correlation[edit]

Akin to Convolution is a technique called "Correlation" that combines two functions in the time domain into a single resultant function in the time domain. Correlation is not as important to our study as convolution is, but it has a number of properties that will be useful nonetheless.

The correlation of two functions, *g(t)* and *h(t)* is defined as such:

Where the capital *R* is the **Correlation Operator**, and the subscripts to *R* are the arguments to the correlation operation.

We notice immediately that correlation is similar to convolution, except that we don't time-invert the second argument before we shift and integrate. Because of this, we can define correlation in terms of convolution, as such:

### Uses of Correlation[edit]

Correlation is used in many places because it demonstrates one important fact: Correlation determines how much similarity there is between the two argument functions. The more the area under the correlation curve, the more is the similarity between the two signals.

### Autocorrelation[edit]

The term "autocorrelation" is the name of the operation when a function is correlated with itself. The autocorrelation is denoted when both of the subscripts to the Correlation operator are the same:

While it might seem ridiculous to correlate a function with itself, there are a number of uses for autocorrelation that will be discussed later. Autocorrelation satisfies several important properties:

- The maximum value of the autocorrelation always occurs at
*t = 0*. The function always decreases, stays constant, or fluctuates (if the signal is periodic) as*t*approaches infinity. - Autocorrelation is symmetric about the
*x*axis.

### Crosscorrelation[edit]

Cross correlation is every instance of correlation that is not considered "autocorrelation". In general, crosscorrelation occurs when the function arguments to the correlation are not equal. Crosscorrelation is used to find the similarity between two signals.

### Example: RADAR[edit]

RADAR is a system that uses pulses of electromagnetic waves to determine the position of a distant object. RADAR operates by sending out a signal, and then listening for echos. If there is an object in range, the signal will bounce off that object and return to the RADAR station. The RADAR will then take the cross correlation of two signals, the sent signal and the received signal. A spike in the cross correlation signal indicates that an object is present, and the location of the spike indicates how much time has passed (and therefore how far away the object is).