Difference between revisions of "ECE 280/Concept List/S24"

From PrattWiki
Jump to navigation Jump to search
Line 75: Line 75:
 
* Elevators are arguably more complicated than toasters...
 
* Elevators are arguably more complicated than toasters...
  
<!--
+
== Lecture 5 - 9/11 - System Properties ==
  
  
== Lecture 5 - 9/11 - System Interconnections and System Properties ==
 
* Systems can be connected together in cascade, parallel, and feedback configurations
 
** For LTI systems, systems in cascade will have their impulse responses convolved together and their transfer functions multiplied by each other
 
** For LTI systems, systems in parallel will have their impulse responses added and their transfer functions multiplied by each other
 
** For LTI systems, systems with negative feedback are a little more complicated
 
 
* System properties - see [[System_Properties]] for some ways to check some of the properties
 
* System properties - see [[System_Properties]] for some ways to check some of the properties
 
** Linearity (linear versus nonlinear)
 
** Linearity (linear versus nonlinear)
Line 95: Line 90:
 
*** Systems whose responses depend only on current and previous values of the independent variable are causal; if they depend at all on future values, they are non-causal.
 
*** Systems whose responses depend only on current and previous values of the independent variable are causal; if they depend at all on future values, they are non-causal.
 
*** Real systems with time $$t$$ as the independent variable are causal; systems with location as the independent value may be non-causal
 
*** Real systems with time $$t$$ as the independent variable are causal; systems with location as the independent value may be non-causal
** Invertibility (invertible versus non-invertible)
 
*** Invertible systems are systems where you can uniquely determine what the input was based on what the output is.
 
*** Determining invertibility is...complicated...
 
  
 +
 +
<!--
 
== Lecture 6 - 9/15 - LTI Systems and Convolution ==
 
== Lecture 6 - 9/15 - LTI Systems and Convolution ==
 
* Introduction to LTI system analysis:
 
* Introduction to LTI system analysis:
Line 221: Line 215:
 
** Time scaling of $$a$$ changes the energy or power by a factor of $$\frac{1}{a}$$
 
** Time scaling of $$a$$ changes the energy or power by a factor of $$\frac{1}{a}$$
 
** Neither time shift nor reversal impact energy or power, so you can shift and flip signal components to more mathematically convenient locations.
 
** Neither time shift nor reversal impact energy or power, so you can shift and flip signal components to more mathematically convenient locations.
 +
-->
 +
 +
<!--
 +
* Systems can be connected together in cascade, parallel, and feedback configurations
 +
** For LTI systems, systems in cascade will have their impulse responses convolved together and their transfer functions multiplied by each other
 +
** For LTI systems, systems in parallel will have their impulse responses added and their transfer functions multiplied by each other
 +
** For LTI systems, systems with negative feedback are a little more complicated
 +
** Invertibility (invertible versus non-invertible)
 +
*** Invertible systems are systems where you can uniquely determine what the input was based on what the output is.
 +
*** Determining invertibility is...complicated...
 
-->
 
-->

Revision as of 04:40, 26 January 2024

Lecture 1 - 8/28 - Class introduction; basic signal classifications

  • Class logistics and various resources on Canvas
  • Definition of signals and systems from OW
  • Systems will often be represented with block diagrams. System operations for linear, time-invariant (more on that later) systems may be characterized in the frequency domain using transfer functions.
  • Signal classifications
    • Dimensionality ($$x(t)$$, $$g(x, y)$$, etc)
    • Continuous versus discrete
  • Analog versus digital and/or quantized
    • Periodic versus aperiodic
      • Generally $$x(t)=x(t+kT)$$ for all integers k (i.e. $$x(t)=x(t+kT), k\in \mathbb{Z}$$). The period $$T$$ (sometimes called the fundamental period $$T_0$$) is the smallest value for which this relation is true
      • A periodic signal can be defined as an infinite sum of shifted versions of one period of the signal: $$x(t)=\sum_{n=-\infty}^{\infty}g(t\pm nT)$$ where $$g(t)$$ is only possibly nonzero within one particular period of the signal and 0 outside of that period.
  • Energy, power, or neither
    • Energy signals have a finite amount of energy: $$E_{\infty}=\int_{-\infty}^{\infty}|x(\tau)|^2\,d\tau<\infty$$
      • Examples: Bounded finite duration signals; exponential decay
    • Power signals have an infinite amount of energy but a finite average power over all time: $$P_{\infty}=\lim_{T\rightarrow\infty}\frac{1}{T}\int_{-T/2}^{T/2}|x(\tau)|^2\,d\tau=\lim_{T\rightarrow\infty}\frac{1}{2T}\int_{-T}^{T}|x(\tau)|^2\,d\tau<\infty$$ and $$E_{\infty}=\infty$$
      • Examples: Bounded infinite duration signals, including periodic signals
      • For periodic signals, only need one period (that is, remove the limit and use whatever period definition you want): $$P_{\infty}=\frac{1}{T}\int_{T}|x(\tau)|^2\,d\tau$$
    • If both the energy and the overall average power are infinite, the signal is neither an energy signal nor a power signal.
      • Examples: Certain unbounded signals such as $$x(t)=e^t$$
  • Useful math shortcut
    • For a trapezoidal pulse
      $$x(t)=\begin{cases}mt+b, &0<t\leq\Delta t\\0,&\mathrm{otherwise}\end{cases}$$
      where
      $$x(0)=b=H_1,~x(\Delta t)=b+m\,\Delta t=H_2$$
      the energy is:
      $$E=\frac{(b+m\,\Delta t)^3-b^3}{3m}=\frac{H_1^2+H_1H_2+H_2^2}{3}\Delta t$$
    • For a rectangular pulse where $$H_1=H_2=A$$, this yields:
      $$E=A^2\,\Delta t$$
    • For a triangle pulse where $$H_1=0$$ and $$H_2=A$$, this yields:
      $$E=\frac{1}{3}A^2\,\Delta t$$
  • If a signal has components that have finite total energy and other components that have finite average power, the finite total energy parts do not contribute to the average power.


Lecture 2 - 9/1 - Periodicity, even and odd, basic transformations, steps and ramps

  • More on periodic signals
    • The sum or difference of two periodic signals will be periodic if their periods are commensurable (i.e. if their periods form a rational fraction) or if any aperiodic components are removed through addition or subtraction.
    • The period of a sum or difference of periodic signals, if periodic, will be at most the least common multiple of the component signal periods; the actual period could be less than this period depending on interference
    • The product of two signals with periodic components will have elements at frequencies equal to the sums and differences of the frequencies in the first signal and the second signal. If the periods represented by those components are commensurable, the signal will be periodic, and again the upper bound on the period will be the least common multiple of the component periods (the sums and differences, not the original).
    • Best bet is to combine the signals, determine the angular frequencies of each component, and determine if all pairs of frequencies are commensurable; if they are, find the largest number that can be multiplied by integers to get all the component frequencies - that number is the fundamental frequency $$\omega_0$$.
  • Evan and Odd
    • Purely even signals: $$x(t)=x(-t)$$ (even powered polynomials, cos, $$|t|$$)
    • Purely odd: $$x(t)=x(-t)$$ (odd-powered polynomials, sin)
    • Even component: $$\mathcal{Ev}\{x(t)\}=x_e(t)=\frac{x(t)+x(-t)}{2}$$
    • Odd component: $$\mathcal{Od}\{x(t)\}=x_o(t)=\frac{x(t)-x(-t)}{2}$$
    • $$x_e(t)+x_o(t)=x(t)$$
    • The even and odd components of $$x(t)=e^{at}$$ end up being $$\cosh(at)$$ and $$\sinh(at)$$
    • The even and odd components of $$x(t)=e^{j\omega t}$$ end up being $$\cos(\omega t)$$ and $$j\,\sin(\omega t)$$
  • Singularity functions - see Singularity_Functions and specifically Singularity_Functions#Accumulated_Differences
    • Unit step: $$u(t)=\begin{cases}1, t>0\\0, t<0\end{cases}$$
    • Unit ramp: $$r(t)=\int_{-\infty}^{t}u(\tau)\,d\tau=\begin{cases}t, t>0\\0, t<0\end{cases}=t\,u(t)$$
    • Unit quadratic: $$q(t)=\frac{1}{2}t^2\,u(t)$$
    • Unit cubic: $$c(t)=\frac{1}{6}t^3\,u(t)$$

Lecture 3 - 9/4 - Signal Transformations, power and energy with transformations, impulse functions

  • Signal transformations
    • $$z(t)=K\,x(a(t-t_0))+C$$ with
    • $$K$$: vertical scaling factor
    • $$\pm a$$: time scaling (with reversal if negative); $$|a|>1$$ speeds things up / compresses the signal while $$|a|<1$$ slows things down / expands the signal
    • $$t_0$$: time shift
    • $$C$$: vertical shift
    • Get into the form above first; for example, rewrite $$3\,x\left(\frac{t}{2}+4\right)$$ as $$3\,x\left(\frac{1}{2}(t+8)\right)$$ first, find the "new origin" (the old point (0, 0) will be at the new point $$(t_0, C)$$ and then go from there
  • Power and energy and transformations
    • Power of $$A\,\cos(\omega t+\phi)=\frac{A^2}{2}$$ - note that $$A$$ is a factor but neither $$\omega$$ nor $$\phi$$ are.
    • In previous lectures, we looked at power signals and their overall average power $$P_{\infty}$$ and energy signals and their total energy $$E_{\infty}$$
    • We've now looked at a transformation of variables $$z(t)=K\,x(\pm a(t-t_0))+C$$
    • If $$x(t)$$ is an energy signal with energy $$E_{\infty,x}$$ or a power signal with overall average power $$P_{\infty,x}$$, then the different transformations above have the following individual impacts:
      • A multiplicative factor of $$K$$ in the transformation means the energy or power of $$z(t)$$ will be the energy or power of $$x(t)$$ multiplied by $$K^2$$
      • A time scale of $$a$$ in the transformation means the energy of $$z(t)$$ will be the energy of $$x(t)$$ multiplied by $$\frac{1}{a}$$; time scaling has no influence on average power
      • Time reversal has no influence on energy or power
      • Time shifts have no influence on energy or power
      • Vertical shifts have a complicated influence on power
  • Definition of the impulse function: Area of 1 at time 0; 0 elsewhere
    • Sifting property - figure out when $$\delta$$ fires off, see if that argument happens or if there are restrictions based on integral limits
  • Integrals with unit steps - figure out when integrand might be non-zero and work from there
  • See Singularity_Functions and especially Singularity_Functions#General_Simplification_of_Integrals and Singularity_Functions#Convolution_Integral_Simplification_with_Step_Function_Product_as_Part_of_Integrand

Lecture 4 - 1/23 - System Representations and Interconnections

  • Systems may be defined generally with an input and an output or specifically for linear and time invariant (LTI) systems with a transfer function
  • The frequency representation of the output of an LTI system is the frequency representation of the input of the system multiplied by the transfer function of the circuit.
  • The transfer function of a feedback system can be found by tracing the signals around the system
  • Elevators are arguably more complicated than toasters...

Lecture 5 - 9/11 - System Properties

  • System properties - see System_Properties for some ways to check some of the properties
    • Linearity (linear versus nonlinear)
      • Common nonlinearities include additive constants, non-unity powers of signals
    • Time invariance (time invariant versus time-varying)
      • Common time-varying elements include $$t$$ outside of arguments of signals, time reversals, or time scales other than 1
    • Stability (stable versus unstable)
      • Common instabilities involve inverses, integrals, some trig functions, and derivatives if you are including discontinuities
    • Memoryless (memoryless versus having memory)
      • Memoryless signals can *only* depend on "right now"; some debate about derivatives
    • Causality (causal versus non-causal)
      • Systems whose responses depend only on current and previous values of the independent variable are causal; if they depend at all on future values, they are non-causal.
      • Real systems with time $$t$$ as the independent variable are causal; systems with location as the independent value may be non-causal