\h1{Miscellaneous} \h2{Functional Utilities & Notation Conveniences} \layout{ \note{ \h3{Right to Left Evaluation} \equation{ f \triangleleft x &= f(x) } \equation{ f \circ g \triangleleft x &= f(g(x)) \\ &= f \triangleleft g \triangleleft x } } \note{ \h3{Left to Right Evaluation} \equation{ x \triangleright f &= f(x) } } \note{ \h3{Derivative Shorthand} \equation{ \delta f(x) &= \frac{\mathrm{d}}{\mathrm{d}x} f(x) = f^\prime(x) } For this notation, the derivative with respect to a given variable, is implicit. } } \h2{Radians & Radian Conversion} \layout{ \note{ \h3{Constants} \equation{ \tau &= 2\pi = 360^{\circ} \\ \pi &= \frac{1}{2}\tau = 180^{\circ} } } \note{ \h3{Conversion} \equation{ \text{Given}\\ \textcolor{blue}{1}^{\circ} &= \frac{\textcolor{blue}{1}}{360} \tau \; {\displaystyle {\mathrm{rad}}} \\ \textcolor{blue}{1} \; \mathrm{rad} &= \frac{\textcolor{blue}{1}}{\tau} \cdot 360^{\circ} = \textcolor{blue}{1} \cdot \frac{360^{\circ}}{\tau} \\ \\\text{Degrees to Radians}\\ \textcolor{blue}{x^{\circ}} &= \frac{\textcolor{blue}{x}}{360} \tau \; {\displaystyle {\mathrm{rad}}} \\ \\\text{Radians to Degrees}\\ \textcolor{blue}{x} \; \mathrm{rad} &= \frac{\textcolor{blue}{x}}{\tau} \cdot 360^{\circ} \\ &= \textcolor{blue}{x} \cdot \frac{360^{\circ}}{\tau} \\ &= \textcolor{blue}{x} \cdot \frac{180^{\circ}}{\pi} \\ } } } \h1{Constants} \layout{ \note{ \h2{\{ℯ} (Euler's number)} \equation{ \mathrm{e} &= \sum_{n = 0}^\infty \frac{1}{n!} \\ &= \lim_{n\to\infty} \left(1 + \frac{1}{n}\right)^n \\ &= \lim_{t \to 0} \left(1 + t\right)^{\frac{1}{t}} } \hr \equation{ \mathrm{e}^x &= 1 + \frac{x}{1!} + \frac{x^2}{2!} + \frac{x^3}{3!} + \cdots \\ &= \sum_{n = 0}^\infty \frac{x^n}{n!} \\ &= \lim_{n\to\infty} \left(1 + \frac{x}{n}\right)^n } } } \h1{Algebra} \h2{Properties} \layout{ \equation{ a^m \cdot b^n &= a^{m+n}\\ \left(a^m\right)^n &= a^{m\cdot\,n} = \left(a^n\right)^m\\ \left(a\cdot\,b\right)^n &= a^n \cdot b^n\\ \left(\frac{a}{b}\right)^{-n} &= \left(\frac{b}{a}\right)^n\\ \frac{x^n}{y^n} &= \left(\frac{x}{y}\right)^n\\ x^{y^z} &= x^{\left(y ^ z\right)} \neq \left(x^y\right)^z } \equation{ |x| &= \sqrt{x^2} \neq x } \equation{ \log_{β}(α) &= \gamma\\ β^{\gamma} &= α\\ β^{\log_{β}(N)} &= N\;\text{for all $N > 0$}\\ \log_{β}(β^x) &= x } } \h1{Trigonometry} \h2{Trigonometric Identities} \layout{ \note{ \h3{Pythagorean Identities} \equation{\cos^2(\theta) + \sin^2(\theta) = 1} \layout{ \equation{ \sec^2(\theta) - \tan^2(\theta) &= 1 \\ \sec^2(\theta) &= 1 + \tan^2(\theta) } \equation{ \csc^2(\theta) - \cot^2(\theta) &= 1 \\ \csc^2(\theta) &= 1 + \cot^2(\theta) } } } \note{ \h3{Sum and Difference Identities} \mathblock{ \equation{ \split{ \cos(α - β) &= \cos(α) \cdot \cos(β) + \sin(α) \cdot \sin(β) \\ \cos(α + β) &= \cos(α) \cdot \cos(β) - \sin(α) \cdot \sin(β) \\ &\\ \sin(α - β) &= \sin(α) \cdot \cos(β) - \cos(α) \cdot \sin(β) \\ \sin(α + β) &= \sin(α) \cdot \cos(β) + \cos(α) \cdot \sin(β) \\ &\\ \tan(α + β) &= \frac{\tan(α) + \tan(β)}{1 - \tan(α) \cdot \tan(β)} \\ \tan(α - β) &= \frac{\tan(α) - \tan(β)}{1 + \tan(α) \cdot \tan(β)} } } } } \note{ \h3{Cofunction Identities} \layout{ \equation{\sin(\theta) = \cos(\frac{1}{4}\tau - \theta)} \equation{\cos(\theta) = \sin(\frac{1}{4}\tau - \theta)} \equation{\tan(\theta) = \cot(\frac{1}{4}\tau - \theta)} \equation{\cot(\theta) = \tan(\frac{1}{4}\tau - \theta)} \equation{\csc(\theta) = \sec(\frac{1}{4}\tau - \theta)} \equation{\sec(\theta) = \csc(\frac{1}{4}\tau - \theta)} } } \note{ \h3{Ratio Identities} \equation{ \tan(90^\circ - x) &= \frac{\sin(90^\circ - x)}{\cos(90^\circ - x)} = \frac{\cos(x)}{\sin(x)} = \cot(x) \\ \\ \cot(90^\circ - x) &= \frac{\cos(90^\circ - x)}{\sin(90^\circ - x)} = \frac{\sin(x)}{\cos(x)} = \tan(x) \\ } } \note{ \h3{Double-Angle Identities} \equation{ \sin(2α) &= 2\sin(α)\cos(α) \\ \cos(2α) &= \cos^2(α) - \sin^2(α) \\ &= 1 - 2\sin^2(α) \\ &= 2\cos^2(α) - 1 \\ &\\ \tan(2α) &= \frac{2\tan(α)}{1 - \tan^2(α)} } } \note{ \h3{Half-Angle Identities} \equation { \sin \frac{α}{2} &= \pm \sqrt{\frac{1 - \cos(α)}{2}} \\ \cos \frac{α}{2} &= \pm \sqrt{\frac{1 + \cos(α)}{2}} \\ \tan \frac{α}{2} &= \pm \sqrt{\frac{1 - \cos(α)}{1 + \cos(α)}} \\ &= \frac{sin(α)}{1 + \cos(α)} \\ &= \frac{1 - \cos(α)}{sin(α)} } } \note{ \h3{Power-Reducing Identities} \equation{ \sin^2(α) &= \frac{1 - \cos(2α)}{2} \\ &\\ \cos^2(α) &= \frac{1 + \cos(2α)}{2} \\ &\\ \tan^2(α) &= \frac{1 - \cos(2α)}{1 + \cos(2α)} } \hr \equation{ \sinα\cdot\cos α &= \frac{1}{2}\sin(2α) } } \note{ \h3{Product-to-Sum Identities} \mathblock{ \equation{ \split{ \sin(α) \cdot \cos(β) &= \frac{1}{2} \Big[ \sin(α + β) + \sin(α - β) \Big] \\ \cos(α) \cdot \sin(β) &= \frac{1}{2} \Big[ \sin(α + β) - \sin(α - β) \Big] \\ \cos(α) \cdot \cos(β) &= \frac{1}{2} \Big[ \cos(α + β) + \cos(α - β) \Big] \\ \sin(α) \cdot \sin(β) &= \frac{1}{2} \Big[ \cos(α - β) - \cos(α + β) \Big] } } } } \note{ \h3{Sum-to-Product-Identities} \mathblock{ \equation{ \split{ \sin(x) + \sin(y) &= 2 \cdot \sin\left( \frac{x + y}{2} \right) \cdot \cos\left( \frac{x - y}{2} \right) \\ \cos(x) + \cos(y) &= 2 \cdot \cos\left( \frac{x + y}{2} \right) \cdot \cos\left( \frac{x - y}{2} \right) \\ \sin(x) - \sin(y) &= 2 \cdot \cos\left( \frac{x + y}{2} \right) \cdot \sin\left( \frac{x - y}{2} \right) \\ \cos(x) - \cos(y) &= -2 \sin \cos\left( \frac{x + y}{2} \right) \cdot \sin\left( \frac{x - y}{2} \right) } } } } } \h2{Trigonometric Equations} \layout{ \note{ \h3{Euler's Formula} \equation{ e^{i x} &= \cos(x) + \mathrm{i}\sin(x) } \hr \equation{ \mathrm{e} &= \sum_{n = 0}^\infty \frac{1}{n!} \\ &= \lim_{n\to\infty} \left(1 + \frac{1}{n}\right)^n \\ &= \lim_{t \to 0} \left(1 + t\right)^{\frac{1}{t}} } \hr \equation{ \mathrm{e}^x &= 1 + \frac{x}{1!} + \frac{x^2}{2!} + \frac{x^3}{3!} + \cdots \\ &= \sum_{n = 0}^\infty \frac{x^n}{n!} \\ &= \lim_{n\to\infty} \left(1 + \frac{x}{n}\right)^n } } } \h1{Coordinate & Number Systems} \layout{ \note{ \equation{ x &= r \cdot \cos\,\theta \\ y &= r \cdot \sin\,\theta \\ } \hr \equation{ z &= x + \mathrm{i}y \\ &= r \left(\cos\, \theta + \mathrm{i}\sin\, \theta\right)\\ & = r\;\mathrm{cis}\; \theta \\ } } \note{ \equation{ e^{i x} &= \cos\, x + \mathrm{i}\sin\, x \\ e ^{i x} &= r \left(\cos\, \theta + \mathrm{i}\sin\, \theta\right) \\ \left(\cos\, x + \mathrm{i}\sin\, x\right)^n &= \cos(n x) + \mathrm{i}\sin(n x) \\ } } \note{ \equation{ r &= |z| = |x + \mathrm{i}y| = \sqrt{x^2 + y^2} \\ r^2 &= x^2 + y^2 \\ \tan \theta &= \frac{y}{x} \\ \theta &= \arctan\left(\frac{y}{x}\right) } } } \layout{ \note{ \h2{Polar Coordinate System} Given \equation{ \tan \theta &= \frac{y}{x} \\ r^2 &= x^2 + y^2 \\ } Then \equation{ x &= r \cdot \cos\,\theta \\ y &= r \cdot \sin\,\theta \\ } } \note{ \h2{Properties} Given \equation{ z_1 &= r_1\;\mathrm{cis}\;\theta_1 \\ z_2 &= r_1\;\mathrm{cis}\;\theta_1 } Then \equation{ z_1 \cdot z_2 &= r_1 \cdot r_2 \;\mathrm{cis}\;\left(\theta_1 + \theta_2\right) \\ \frac{z_1}{z_2} &= \frac{r_1}{r_2} \;\mathrm{cis}\;\left(\theta_1 - \theta_2\right) \\ } \hr De Moivre’s Theorem \equation{ z^n &= r^n \;\mathrm{cis}\;\;n\theta } De Moivre’s Theorem For Finding Roots \equation{ \underbrace{w_k = r^{\frac{1}{n}}\;\mathrm{cis}\;\frac{\theta + \tau\cdot{k}}{n}} _{\split{\forall\; k &= 0,1,2\cdots,n-1\\ n &\geq 1}} } } \note{ \h2{Trigonometric form of a complex number} \equation{ z &= x + \mathrm{i}y \\ &= r \left(\cos\, \theta + \mathrm{i}\sin\, \theta\right)\\ & = r\;\mathrm{cis}\;\theta \\ } } } \h1{Vectors} \note{ \h2{Quick Facts} \ul { \li{Two vectors are equal if they share the same magnitude and direction. Initial points don't matter.} \li{ \note{ You can define a vector with two points. \img[width=300px center src=../static/drawings/matrix/trinagle-def-of-a-vector.png] } } \li{Do not divide vectors!} \li{\{\vec{v} \cdot \vec{v} = ||\vec{v}||^2 \implies \text{constant}}} \li{\{\vec{v} \cdot \left(\vec{v}\right)^\prime = 0}} } } \h2{Vector Operations} \layout{ \note{ \h3{Dot Product} \equation{ \vec{a}\cdot\vec{b} &= \bmatrix{ a_1 \\ a_2 \\ a_3 } \cdot \bmatrix{ b_1 \\ b_2 \\ b_3 } = a_1b_1 + a_2b_2 + a_3b_3\\\\ \vec{a}\cdot\vec{b} &= \bmatrix{ a_1 \\ a_2 } \cdot \bmatrix{ b_1 \\ b_2 } = a_1b_1 + a_2b_2 } } \note{ \h3{Cross Product} \equation{ \vec{a}\times\vec{b} &= \bmatrix {a_1 & a_2 & a_3}\times\bmatrix{ b_1 & b_2 & b_3 }\\ &= \mathrm{det} \pmatrix{ \mathrm{\hat{i}} & \mathrm{\hat{j}} & \mathrm{\hat{k}}\\ a_1 & a_2 & a_3\\ b_1 & b_2 & b_3 }\\ &= \mathrm{\hat{i}} \vmatrix{ e & f\\ h & i } - \mathrm{\hat{j}} \vmatrix { d & f\\ g & i } + \mathrm{\hat{k}} \vmatrix{ d & e\\ g & h }\\ &= \mathrm{\hat{i}}\left(e i - f h\right) - \mathrm{\hat{j}}\left(d i - f g\right) + \mathrm{\hat{k}}\left(d h - e g\right)\\ &= \bmatrix { e i - f h & d i - f g & d h - e g }\\ &= \vec{c} } } \note{ \h3{Length of a Vector} \equation{ |a| &= \sqrt{(a_1)^2 + (a_2)^2} \\ |a| &= \sqrt{(a_1)^2 + (a_2)^2 + (a_3)^2} } } \note{ \h3{Definition of Vector Addition} If \{\vec{u}} and \{\vec{v}} are positioned so the initial point of \{\vec{v}} is at the terminal point of \{\vec{u}}, then the sum \{\vec{u} + \vec{v}} is the vector from the initial point of \{\vec{u}} to the terminal point of \{\vec{v}}. \img[width=300px center src=../static/drawings/matrix/Definition-of-Vector-Addition.png] \hr Given some vectors \{\vec{u}} and \{\vec{v}}, the vector \{\vec{u} - \vec{v}} is the vector that points from the head of \{\vec{v}} to the head of \{\vec{u}} \img[width=200px center src=../static/drawings/matrix/vector-u-v.png] } } \layout{ \note{ \h3{Standard Basis Vectors} \equation{ \mathrm{\hat{i}} &= \bmatrix{ 1 & 0 & 0 }\\ \mathrm{\hat{j}} &= \bmatrix{ 0 & 1 & 0 }\\ \mathrm{\hat{k}} &= \bmatrix{ 0 & 0 & 1 } } \hr \img[width=200px center src=../static/drawings/matrix/unit-vectors.png] } \note{ \h3{Orthogonal} Two vectors are orthogonal if and only if \equation{ \vec{a}\cdot\vec{b} = 0 } } \note{ \h3{The Unit Vector} \equation{ \hat{u} &= \frac{\vec{a}}{|\vec{a}|} } } \note{ \h3{If \{\theta} is the angle between the vectors \{\vec{a}} and \{\vec{b}}, then} \equation{ \vec{a} \cdot \vec{a} = |\vec{a}| \cdot |\vec{a}| \cos\theta } } \note{ \h3{If \{\theta} is the angle between the nonzero vectors \{\vec{a}} and \{\vec{b}}, then} \equation{ \cos\theta = \frac {\vec{a}\cdot\vec{b}} {|\vec{a}| |\vec{b}|} } } \note{ \h3{Two nonzero vectors \{\vec{a}} and \{\vec{b}} are parallel if and only if} \equation{ \vec{a}\times\vec{b} = 0 } } } \h3{Properties of the Dot Product} \layout{ \note{ \equation{ \vec{a} \cdot \vec{a} = |\vec{a}|^2 } } \note{ \equation{ \vec{a} \cdot \left(\vec{b} + \vec{c}\right) = \vec{a}\vec{b} + \vec{a}\vec{c} } } \note{ \equation{ \vec{a} \cdot \vec{b} = \vec{b} \cdot \vec{a} } } \note{ \equation{ \left(c \cdot \vec{a}\right) \cdot \vec{b} = c\left(\vec{a} \cdot \vec{b}\right) } } } \h2{Direction Cosines & Direction Angles of a Vector} \note{ \img[width=300px center src=../static/drawings/matrix/Direction-Cosines-Direction-Angles-of-a-Vector.png] Where \equation{ \vec{v} &= \bmatrix{ v_x & v_y & v_z } \;\;\;\; ||\vec{v}|| &= \sqrt{(v_x)^2 + (v_y)^2 + (v_z)^2} } } \layout{ \note{ \h3{Direction Cosines} \equation{ \cos α &= \frac{v_x}{||\vec{v}||}\\ \cos β &= \frac{v_y}{||\vec{v}||}\\ \cos\gamma &= \frac{v_z}{||\vec{v}||} } } \note{ \h3{Direction Angles} \equation{ α &= \arccos \left(\frac{v_x}{||\vec{v}||}\right)\\ β &= \arccos \left(\frac{v_y}{||\vec{v}||}\right)\\ \gamma &= \arccos \left(\frac{v_z}{||\vec{v}||}\right) } } \note{ \h3{Theorem} \equation{ \cos^2 α + \cos^2 β + \cos^2\gamma = 1 } \hr \h5{Proof} \equation{ \vec{v} &= \hat{i}\cos α + \hat{k}\cos β + \hat{k}\cos\gamma\\ ||\vec{v}|| &= || \hat{i}\cos α + \hat{k}\cos β + \hat{k}\cos\gamma|| = 1 } Given \equation{ \sqrt{\cos^2α + \cos^2β + \cos^2\gamma} &= 1\\ \sqrt{\cos^2α + \cos^2β + \cos^2\gamma}^2 &= 1^2\\ \cos^2α + \cos^2β + \cos^2\gamma &= 1 } Therefore \equation{ \cos^2α + \cos^2β + \cos^2\gamma = 1\\ } } } \h2{Vector Relations} \layout{ \note{ \h3{Parallel Vectors} \ul { \li{When two vectors are parallel; they never intersect (duh).} } Given some vectors \equation{ \vec{a} &= \bmatrix{ a_x & a_y & a_z }\\ \vec{b} &= \bmatrix{ b_x & b_y & b_z } } The vectors \{\vec{a}} and \{\vec{b}} are parallel if and only if they are scalar multiples of one another. \equation{ \vec{a} &= k\;\vec{b} \;\;\;\;\forall k \neq 0 } \h4{Alternatively} \equation{ \frac{a_x}{b_x} = \frac{a_y}{b_y} = \frac{a_z}{b_z} } } \note{ \h3{Orthogonal Vectors} \ul { \li{When two vectors are orthogonal; they meet at right angles.} } Given some vectors \equation{ \vec{a} &= \bmatrix{ a_x & a_y & a_z }\\ \vec{b} &= \bmatrix{ b_x & b_y & b_z } } Two vectors are orthogonal if and only if \equation{ \vec{a}\times\vec{b} = 0 } } } \h2{Reparameterization of the position vector \{\vec{v}(t)} in terms of length \{S(t)}} \ul { \li{We can parametrize a curve with respect to arc length; because arc length arises naturally from the shape of the curve and does not depend on any coordinate system.} } \layout{ \note{ \h3{The Arc Length Function} Given \equation{ \vec{v} &= \bmatrix{ x & y & z } = \bmatrix{ f(t) & g(t) & h(t) } } We can redefine \{\vec{v}} in terms of arc length between two endpoints \equation{ S(t) &= \Long\\ &= \AltLong\\ &= \short } \!where { {\Long} => { \int_a^t \sqrt{ \left(f^\prime(u)\right)^2 \left(g^\prime(u)\right)^2 \left(h^\prime(u)\right)^2 } } {\AltLong} => { \int_a^t \sqrt{ \left(\frac{\mathrm{d}x}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}y}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}z}{\mathrm{d}u}\right)^2 } } {\short} => { \int_a^t ||\left(v^\prime(u)\right)|| } } That is, \{S(t)} is the length of the curve (\{C}) between \{r(a)} and \{r(b)}. \hr Furthermore from the adjacent definition; we can simply the above to \equation{ S(t) = \int_a^t \frac{\mathrm{d}S}{\mathrm{d}t} } } \note{ \h3{The Arc Length Function} \equation{ \frac{\mathrm{d}S}{\mathrm{d}t} \equiv || v^\prime(t) || } That is \equation{ \frac{\mathrm{d}S}{\mathrm{d}t} &= \Long\\ &= \AltLong\\ &= \short } \!where { {\Long} => { \sqrt{ \left(f^\prime(u)\right)^2 \left(g^\prime(u)\right)^2 \left(h^\prime(u)\right)^2 } } {\AltLong} => { \sqrt{ \left(\frac{\mathrm{d}x}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}y}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}z}{\mathrm{d}u}\right)^2 } } {\short} => { ||\left(v^\prime(u)\right)|| } } } } \h2{Vectors Derived From Some Curve Defined by \{\vec{v}}} \note{ \img[width=300px center src=../static/drawings/matrix/curvature.png] } \note{ \h3{The Unit Vector} \equation{ \hat{U} \equiv \frac{\vec{v}}{||\vec{v}||} } } \note{ \h3{The Unit Tangent Vector} \equation{ \vec{T} &\equiv \frac {v^\prime(t)} {||v^\prime(t)||} &\equiv \frac {\mathrm{d}v} {\mathrm{d}S} } } \note{ \h3{The Unit Normal Vector} \equation{ \vec{N} \equiv \frac {T^\prime} {||T^\prime||} } } \note{ \h3{The Binormal Vector} \equation{ \vec{B} \equiv \vec{T}\times\vec{N} } \ul { \li{Therefore, the binormal vector is orthogonal to both the tangent vector and the normal vector.} \li{The plane determined by the normal and binormal vectors N and B at a point P on a curve C is called the normal plane of C at P.} \li{The plane determined by the vectors T and N is called the osculating plane of C at P. The name comes from the Latin osculum, meaning “kiss.” It is the plane that comes closest to containing the part of the curve near P. (For a plane curve, the osculating plane is simply the plane that contains the curve.)} } } \note{ \h3{Kappa - Curvature of a Vector} \equation{ \kappa &\equiv \left|\frac{\mathrm{d}T}{\mathrm{d}S}\right| &\equiv \frac {\left| T^\prime \right|} {\left| r^\prime \right|} &\equiv \frac {\left| r^\prime \times r^{\prime\prime} \right|} {\left| r^\prime \right|^3} } } \note{ \h3{Tangential & Normal Components of the Acceleration Vector of the Curve} When we study the motion of a particle, it is often useful to resolve the acceleration into two components, one in the direction of the tangent and the other in the direction of the normal. \equation{ a_{\vec{T}} &= \frac{\mathrm{d}}{\mathrm{d}t} \left|\vec{v}\right| = \frac{r^\prime \cdot r^{\prime\prime}}{|r^\prime|}\\ a_{\vec{N}} &= \kappa \left|\vec{v}\right|^2 = \frac{\left|r^\prime \times r^{\prime\prime}\right|}{|r^\prime|}\\ \vec{a} &= a_{\vec{T}} \vec{T} + a_{\vec{N}} \vec{N} } Specifically \equation{ \left.\aligned{ a_{\vec{T}}\\ a_{\vec{N}} }\right\} \text{Tangential & Normal Components of $\vec{a}$} } } \h2{Vector Calculus} \note{ \h3{The Position Vector \{\vec{r}(t)}} (Original Function) } \note{ \h3{The Velocity Vector \{\vec{v}(t)}} (First Derivative) \ul { \li{The velocity vector is also the tangent vector and points in the direction of the tangent line.} \li{The speed of the particle at time t is the magnitude of the velocity vector, that is, \equation{ \underbrace{|\vec{v}(t)| = |(\vec{r})^\prime(t)| = \frac{\mathrm{d}s}{\mathrm{d}t}} _{\text{rate of change of distance with respect to time}} } } } } \note{ \h3{The Acceleration Vector \{\vec{a}(t)}} (Second Derivative) } \h1{Matrices} \h2{Reference} \layout{ \note{ \h3{The Determinant of A Matrix} \equation{ |A| &= \vmatrix{a & b \\ c & d} = ad - bc\\\\ |A| &= \vmatrix{a & b & c \\ d & e & f \\ g & h & i}\\ &= a \vmatrix{e & f \\ h & i } - b \vmatrix{d & f\\ g & i} + c \vmatrix{d & e\\g & h} } \img[center width=400px src=../static/drawings/matrix/3x3-determinant.png] \note{ Only works for square matrices. } } \note{ \h3{The Cross Product} \equation{ \vec{a}\times\vec{b} &= \bmatrix{ a_1 & a_2 & a_3 }\times\bmatrix{ b_1 & b_2 & b_3 }\\ &= \mathrm{det} \pmatrix{ \mathrm{\hat{i}} & \mathrm{\hat{j}} & \mathrm{\hat{k}}\\a_1 & a_2 & a_3\\b_1 & b_2 & b_3 }\\ &= \mathrm{\hat{i}}\vmatrix{e & f\\h & i} - \mathrm{\hat{j}}\vmatrix{d & f\\g & i} + \mathrm{\hat{k}}\vmatrix{d & e\\g & h}\\ &= \mathrm{\hat{i}}\left(e i - f h\right) - \mathrm{\hat{j}}\left(d i - f g\right) + \mathrm{\hat{k}}\left(d h - e g\right)\\ &= \bmatrix{e i - f h & d i - f g & d h - e g}\\ &= \vec{c} } } } \h1{Geometry} \h2{Definition of a Line} \img[width=500px center src=../static/drawings/matrix/vector-equation-of-a-line.png] \layout{ \note{ \h3{Vector Equation of a Line} Given \equation{ \colorB{P_1} &= \bmatrix{ \colorB{x_1} & \colorB{y_1} & \colorB{z_1} }\\ \colorC{P_2} &= \bmatrix{ \colorC{x_2} & \colorC{y_2} & \colorC{z_2} }\\ } \equation{ \colorB{P_1} &= \bmatrix{ \colorB{x_1} & \colorB{y_1} & \colorB{z_1} }\\ \colorC{P_2} &= \bmatrix{ \colorC{x_2} & \colorC{y_2} & \colorC{z_2} }\\ } We can define a vector between \{\colorB{P_1}} and \{\colorC{P_2}} \equation{ \colorA{\overrightarrow{\Delta\mathsf{v}}} &= \bmatrix{ \colorC{x_2} \\ \colorC{y_2} \\ \colorC{z_2} } - \bmatrix{ \colorB{x_1} \\ \colorB{y_1} \\ \colorB{z_1} } = \bmatrix{ \colorC{x_2}-\colorB{x_1} \\ \colorC{y_2}-\colorB{y_1} \\ \colorC{z_2}-\colorB{z_1} } = \bmatrix{ \colorA{\Delta_x} \\ \colorA{\Delta_y} \\ \colorA{\Delta_z} } } \h4{Therefore} The equation of a line in 3D space or \{\mathbb{R}^3} can be defined VIA the following options \equation{ L &= \colorB{P_1} + t\cdot\colorA{\overrightarrow{\Delta\mathsf{v}}} \\ \bmatrix{ x \\ y \\ z } &= \bmatrix{ \colorB{x_1} \\ \colorB{y_1} \\ \colorB{z_1} } + t\bmatrix{ \colorA{\Delta_x} \\ \colorA{\Delta_y} \\ \colorA{\Delta_z} } = \bmatrix{ \colorB{x_1} \\ \colorB{y_1} \\ \colorB{z_1} } + \bmatrix{ t\;\colorA{\Delta_x} \\ t\;\colorA{\Delta_y} \\ t\;\colorA{\Delta_z} } \\ &= \bmatrix{ \colorB{x_1} + t\;\colorA{\Delta_x} \\ \colorB{y_1} + t\;\colorA{\Delta_y} \\ \colorB{z_1} + t\;\colorA{\Delta_z} }\\ &= \bmatrix{ \colorB{x_1} + t(\colorC{x_2}-\colorB{x_1}) \\ \colorC{x_2} + t(\colorC{x_2}-\colorB{x_1}) \\ x_3 + t(\colorC{x_2}-\colorB{x_1}) } } That is \equation{ L &= \colorB{P_1} + t\cdot\colorA{\overrightarrow{\Delta\mathsf{v}}}\\ \bmatrix{ x \\ y \\ z } &= \bmatrix{ \colorB{x_1} + t\;\colorA{\Delta_x} \\ \colorB{y_1} + t\;\colorA{\Delta_y} \\ \colorB{z_1} + t\;\colorA{\Delta_z} } = \bmatrix{ \colorB{x_1} + t(\colorC{x_2}-\colorB{x_1}) \\ \colorB{y_1} + t(\colorC{y_2}-\colorB{y_1}) \\ \colorB{z_1} + t(\colorC{z_2}-\colorB{z_1}) } } } \layout{ \note{ \h3{Parametric Equation of a Line} \equation{ \underbrace{\split{ x &= \colorB{x_1} + t(\colorC{x_2}-\colorB{x_1}) = \colorB{x_1} + t\;\colorA{\Delta_x}\\ y &= \colorB{x_1} + t(\colorC{y_2}-\colorB{y_1}) = \colorB{y_1} + t\;\colorA{\Delta_y}\\ z &= \colorB{x_1} + t(\colorC{z_2}-\colorB{z_1}) = \colorB{z_1} + t\;\colorA{\Delta_z} }}_{ r \;=\; r_0 \;+\; a \;=\; r_0 \;+\; t\,v } } } \note{ \h3{Essentially} \equation{ L &= \colorB{P_1} + t\cdot\colorA{\overrightarrow{\Delta\mathsf{v}}}\;\;\forall t\in\mathbb{R} } That is, \{t} is the scaling factor. In a way, it's like it's a function of \{t}, but also similar to the slope (\{m}) in \{y = mx + b}, except \{m} (i.e. \{t}) is parameterized. \hr Sometimes this will be (confusingly) denoted as \equation{ \vec{r} &= \vec{r_0} + \vec{a} = \vec{r_0} + t\vec{v}\\ } } } \note{ \h3{Symmetric Equation of a Line} \equation{ t &= \frac{x - \colorB{x_1}}{\colorC{x_2}-\colorB{x_1}} = \frac{x - \colorB{x_1}}{\colorA{\Delta_x}}\\ t &= \frac{y - \colorB{y_1}}{\colorC{y_2}-\colorB{y_1}} = \frac{y - \colorB{y_1}}{\colorA{\Delta_y}}\\ t &= \frac{z - \colorB{z_1}}{\colorC{z_2}-\colorB{z_1}} = \frac{z - \colorB{z_1}}{\colorA{\Delta_z}} } Therefore \equation{ \frac{x - \colorB{x_1}}{\colorA{\Delta_x}} &= \frac{y - \colorB{y_1}}{\colorA{\Delta_y}} = \frac{z - \colorB{z_1}}{\colorA{\Delta_z}} \\\\ \frac{x - \colorB{x_1}}{\colorC{x_2}-\colorB{x_1}} &= \frac{y - \colorB{y_1}}{\colorC{y_2}-\colorB{y_1}} = \frac{z - \colorB{z_1}}{\colorC{z_2}-\colorB{z_1}} } \hr \h4{Rationale} We rewrite \{r = r_0 + a = r_0 + t v} in terms of \{t}. That is \equation{ x &= \colorB{x_1} + t(\colorC{x_2}-\colorB{x_1}) = \colorB{x_1} + t\;\colorA{\Delta_x}\\ t\;\colorA{\Delta_x} &= x - \colorB{x_1} = t(\colorC{x_2}-\colorB{x_1})\\ t &= \frac{x - \colorB{x_1}}{\colorC{x_2}-\colorB{x_1}} = \frac{x - \colorB{x_1}}{\colorA{\Delta_x}} \\\\ y &= \colorB{y_1} + t(\colorC{y_2}-\colorB{y_1}) = \colorB{y_1} + t\;\colorA{\Delta_y}\\ t\;\colorA{\Delta_y} &= y - \colorB{y_1} = t(\colorC{y_2}-\colorB{y_1})\\ t &= \frac{y - \colorB{y_1}}{\colorC{y_2}-\colorB{y_1}} = \frac{y - \colorB{y_1}}{\colorA{\Delta_y}} \\\\ z &= \colorB{z_1} + t(\colorC{z_2}-\colorB{z_1}) = \colorB{z_1} + t\;\colorA{\Delta_z}\\ t\;\colorA{\Delta_z} &= z - \colorB{z_1} = t(\colorC{z_2}-\colorB{z_1}) \\ t &= \frac{z - \colorB{z_1}}{\colorC{z_2}-\colorB{z_1}} = \frac{z - \colorB{z_1}}{\colorA{\Delta_z}} } } } \h2{Parameterizations of a curve} \dt{Parametrized curve} \dd{A curve in the plane is said to be parameterized if the set of coordinates on the curve, (x,y), are represented as functions of a variable t.} \dd{A parametrized Curve is a path in the xy-plane traced out by the point \{\left(x(t), y(t)\right)} as the parameter \{t} ranges over an interval \{I}.} \dd{A parametrized Curve is a path in the xyz-plane traced out by the point \{\left(x(t), y(t), z(t)\right)} as the parameter \{t} ranges over an interval \{I}.} \h2{Curvature Properties} \layout{ \note{ \h3{Length of a Curve} \mathblock { \equation{ L &= \Long\\ &= \AltLong\\ &= \short\\ &\implies \text{some constant} } } \!where { {\Long} => { \int_a^b \sqrt{ \left(f^\prime(t)\right)^2 \left(g^\prime(t)\right)^2 \left(h^\prime(t)\right)^2 } } {\AltLong} => { \int_a^b \sqrt{ \left(\frac{\mathrm{d}x}{\mathrm{d}t}\right)^2 \left(\frac{\mathrm{d}y}{\mathrm{d}t}\right)^2 \left(\frac{\mathrm{d}z}{\mathrm{d}t}\right)^2 } } {\short} => { \int_a^b ||\left(r^\prime(t)\right)|| } } } \note{ \h3{The Arc Length Function} Suppose \ul { \li{Given some curve \{C} defined by some vector \{\vec{r}} in \{\mathbb{R}^3}} \li{where \{r^\prime} is continuous and \{C} is traversed exactly once as \{t} increases from \{a} to \{b}} } We can define it's arc length function VIA \mathblock{ \equation{ S(t) &= \Long\\ &= \AltLong\\ &= \short } } \!where { {\Long} => { \int_a^t \sqrt{ \left(f^\prime(u)\right)^2 \left(g^\prime(u)\right)^2 \left(h^\prime(u)\right)^2 } } {\AltLong} => { \int_a^t \sqrt{ \left(\frac{\mathrm{d}x}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}y}{\mathrm{d}u}\right)^2 \left(\frac{\mathrm{d}z}{\mathrm{d}u}\right)^2 } } {\short} => { \int_a^t ||\left(r^\prime(u)\right)|| } } } } \h1{Limits} \layout{ \note{ \h2{L’Hospital’s Rule} \equation{ \lim_{x\to{a}}\frac{f(x)}{g(x)} &= \lim_{x\to{a}}\frac{f^\prime(x)}{g^\prime(x)}\\ \text{if}\;\lim_{x\to{a}}\frac{f(x)}{g(x)} = \frac{0}{0} \;&\text{or}\;\lim_{x\to{a}}\frac{f(x)}{g(x)} = \frac{\infty}{\infty} } In other words, if L’Hospital’s Rule applies to indeterminate forms. } } \h2{Limit Laws} \layout{ \equation{\lim_{x \to a} \Big[ f(x) + g(x) \Big] &= \lim_{x \to a} f(x) + \lim_{x \to a} g(x)} \equation{\lim_{x \to a} \Big[ f(x) - g(x) \Big] &= \lim_{x \to a} f(x) - \lim_{x \to a} g(x)} \equation{\lim_{x \to a} \Big[ c \cdot f(x) \Big] &= c \cdot \lim_{x \to a} f(x)} \equation{\lim_{x \to a} \Big[ f(x) \cdot g(x) \Big] &= \lim_{x \to a} f(x) \cdot \lim_{x \to a} g(x)} \equation{\lim_{x \to a} \frac{f(x)}{g(x)} &= \frac{\lim_{x \to a} f(x)}{\lim_{x \to a}g(x)} \; \text{if} \lim_{x \to a}g(x) \neq 0} \equation{\lim_{x \to a} \Big[ f(x) \Big]^n = \Big[ \lim_{x \to a} f(x) \Big]^n\;\text{if}\;n\in\mathbb{Z}^+} \equation{\underbrace{\lim_{x \to a} \sqrt[n]{f(x)} = \sqrt[n]{\lim_{x \to a} f(x)}\;\text{if}\;n\in\mathbb{Z}^+}_{\text{if $n$ is even we assume $\lim_{x \to a} f(x) > 0$}}} \equation{\lim_{x\to{c}}\, \left(f \circ g\right)(x) &= f\left(\lim_{x\to{c}}\, g(x)\right)} } \h2{Limit Formulas} \layout{ \equation{\lim_{x\to\infty} \frac{1}{x^r} &= 0\;,\;\forall r > 0} \equation{\lim_{x\to\infty} \frac{x^n}{!n} &= 0\;,\;\forall n \in \mathbb{R}} } \h2{Growth Rates} \ol { \li{Factorial functions grow faster than exponential functions.} \li{Exponential functions grow faster than polynomials.} } \h1{Calculus} \note{ \h2{Derivative Tables} \layout \equation{ \delta\sin(x) &= \cos(x) \\ \delta\csc(x) &= -\cot(x) \cdot \csc(x) } \equation{ \delta\cos(x) &= -\sin(x) \\ \delta\sec(x) &= \tan(x) \cdot \sec(x) } \equation{ \delta\tan(x) &= \sec^2(x) \\ \delta\cot(x) &= -\csc^2(x) } } \layout \equation{ \delta\sin^{-1}(x) &= \frac{1}{\sqrt{1 - x^2}} } \equation{ \delta\cos^{-1}(x) &= -\frac{1}{\sqrt{1 - x^2}} } \equation{ \delta\tan^{-1}(x) &= \frac{1}{1 + x^2} } } } \note{ \h2{Integration Tables} \layout \equation{ \int\sin(x)\;\mathrm{d}x &= -\cos x \\ \int\csc^2(x)\;\mathrm{d}x &= -\cot x \\ \int\csc(x)\cdot\cot(x)\;\mathrm{d}x &= -\csc x \\ \int\csc(x)\;\mathrm{d}x &= \ln\left|\csc x - \cot x\right| \\ \int\sinh(x)\;\mathrm{d}x &= \cosh(x) \\ } \equation{ \int\cos(x)\;\mathrm{d}x &= \sin x \\ \int\sec^2(x)\;\mathrm{d}x &= \tan x \\ \int\sec(x)\cdot\tan(x)\;\mathrm{d}x &= \sec x \\ \int\sec(x)\;\mathrm{d}x &= \ln\left|\sec x + \tan x\right| \\ \int\cosh(x)\;\mathrm{d}x &= \sinh(x) \\ } \equation{ \int\tan(x)\;\mathrm{d}x &= \ln| \sec x | \\ \int\cot(x)\;\mathrm{d}x &= \ln| \sin x | \\ } } \layout \equation{ \int b^x\;\mathrm{d}x &= \frac{b^x}{\ln(b)} } \equation{ \int\frac{1}{x^2 + a^2}\;\mathrm{d}x &= \frac{1}{a}\tan^{-1}\left(\frac{x}{a}\right) } \equation{ \int\frac{1}{x^2 - a^2}\;\mathrm{d}x &= \frac{1}{2a}\ln\left|\frac{x-a}{x+a}\right| } \equation{ \int\frac{1}{\sqrt{a^2 - x^2}}\;\mathrm{d}x &= \sin^{-1}\left(\frac{x}{a}\right),\;a>0 } \equation{ \int\frac{1}{\sqrt{x^2 \pm a^2}}\;\mathrm{d}x &= \ln\left|x+\sqrt{x^2\pm a^2}\right| } } } \note{ \h2{Riemann Sums} Given \equation{ A &= \int_{a}^{b} f(x) = \lim_{n\to\infty}\sum_{i=1}^{n} \Delta{x} \cdot f(x) \;\text{where}\; \left\{\array{{ll} \Delta{x} = \frac{b - a}{n} }\right. } \layout{ \note{ \h3{Left Riemann Sum} \equation{ A &= \int_{a}^{b} f(x) \;\mathrm{d}x \approx L_n = \sum_{i = 0}^{n-1}\, \Delta{x}\cdot f\left(a + i\cdot\Delta{x}\right) } } \note{ \h3{Right Riemann Sum} \equation{ A &= \int_{a}^{b} f(x) \;\mathrm{d}x \approx R_n = \sum_{i = 1}^{n}\, \Delta{x}\cdot f\left(a + i\cdot\Delta{x}\right) } } } \note{ \h3{Midpoint Riemann Sum} \equation{ A = \int_{a}^{b} f(x) \;\mathrm{d}x &\approx \generalFormat\\ &\approx \SigmaExampleOne\\ \text{or alternatively}&\\ A &\approx \SigmaExampleTwo } \!where { {\generalFormat} => { \sum_{\small{\cdots}}^{\small{\cdots}}\, \Delta{x}\cdot f\left(a + \text{“avg. of $x_i$ and $x_{i-1}$”} \cdot \Delta{x}\right) } {\SigmaExampleOne} => { \sum_{i = 1}^{n}\, \Delta{x}\cdot f\left(a + \frac{x_{i - 1} + x_i}{2} \cdot \Delta{x}\right) } {\SigmaExampleTwo} => { \sum_{i = 0}^{n-1}\, \Delta{x}\cdot f\left(a + \frac{x_{i + 1} + x_i}{2} \cdot \Delta{x}\right) } } \hr We can also do away with the index notation and simplify things. \equation{ A = \int_{a}^{b} f(x) \;\mathrm{d}x &\approx \SigmaExampleOne\\ &\approx \SigmaExampleTwo } \!where { {\SigmaExampleOne} => { \sum_{i = 1}^{n}\, \Delta{x}\cdot f\left(a + \frac{(i - 1) + i}{2} \cdot \Delta{x}\right) = \sum_{i = 1}^{n}\, \Delta{x}\cdot f\left(a + \frac{2i - 1}{2} \cdot \Delta{x}\right) } {\SigmaExampleTwo} => { \sum_{i = 0}^{n-1}\, \Delta{x}\cdot f\left(a + \frac{(i + 1) + i}{2} \cdot \Delta{x}\right) = \sum_{i = 0}^{n-1}\, \Delta{x}\cdot f\left(a + \frac{2i + 1}{2} \cdot \Delta{x}\right) } } } \note{ \h3{Trapezoidal Riemann Sum} \mathblock{ \dots } } \note{ \h3{Simpson's Rule} \equation{ \dots } } } \layout{ \note{ \h2{Improper Integrals} \equation{ \text{Given}&\;L = \int_{a}^{\infty} f(x) \;\mathrm{d}x = \lim_{t\to\infty} \int_{a}^{\infty} f(x) \;\mathrm{d}x = \lim_{t\to\infty} F(x)\\ \text{If}&\;L\;\text{“exists”}\;\text{then $L$ is}\;\mathbf{\text{convergent}}\\ \text{If}&\;L\;\text{“does not exists”}\;\text{then $L$ is}\;\mathbf{\text{divergent}} } } } \h2{Infinite Sequences} \layout{ \note{ \h3{Infinite Sequence} \equation{ \text{Given}&\\ &S_n = \{a_n\}_{n=1}^{\infty}\\ \text{Tests}&\\ &\text{If}\;\lim_{n\to\infty}\,S_n \text{“exists”}\;\text{then}\;\mathbf{\text{$S_n$ is convergent}}\\ &\text{If}\;\lim_{n\to\infty}\,S_n \text{“does not exists”}\;\text{then}\;\mathbf{\text{$S_n$ is Divergent}}\\ } } \note{ \h3{Helpful Theorem} \equation{ \text{If}&\\ \lim_{x\to\infty}\,|a_n| &= 0\\ \text{Then}&\\ \lim_{x\to\infty}\,a_n &= 0\\ } } \note{ \h3{Example} \equation{ \text{Given}\\ S_n &= \{\cos\left(\frac{n\pi}{2}\right)\}\\ L &=\lim_{n\to\infty}\,\cos\left(\frac{n\pi}{2}\right)\\ &= \cos\left(\lim_{n\to\infty}\,\frac{n\pi}{2}\right) \\ &= \cos\left(\infty\right) \\ &= \text{undefined} \\ \text{Therefore}\\ \therefore\;&\text{$S_n$ is Divergent}& } } \note{ \h3{Example} \equation{ \text{Given}\\ S_n &= \{\sin\left(\frac{\pi}{n}\right)\}\\ L &= \lim_{n\to\infty}\,\sin\left(\frac{\pi}{n}\right)\\ &= \sin\left(\lim_{n\to\infty}\,\frac{\pi}{n}\right) \\ &= \sin\left(0\right) \\ &= 0\\ \text{Therefore}\\ \therefore\;&\text{$S_n$ is Convergent} } } } \h2{Infinite Series} \layout{ \note{ \h3{Infinite Series} \equation{ \text{Given}&\\ &S_n = \sum_{n = 1}^{\infty}\,a_n\\ \text{Tests}&\\ &\text{If}\;\lim_{n\to\infty}\,a_n = 0\;\text{then}\;\text{$S_n$ may be $\mathbf{\text{convergent}}$}\\ &\text{If}\;\lim_{n\to\infty}\,a_n \neq 0\;\text{then}\;\mathbf{\text{$S_n$ is divergent}}\\ &\text{If}\;\lim_{n\to\infty}\,a_n \text{“does not exists”}\;\text{then}\;\mathbf{\text{$S_n$ is divergent}}\\ } Note that the limit of every convergent series is equal to zero. But the inverse isn't always true. If the limit is equal to zero, it may not be convergent. For example, \{\sum_{n=1}^\infty \frac{1}{n}} does diverge; but it's limit is equal to zero. If the limit is equal to zero; the test is inconclusive. } \note{ \h3{Geometric Series} \equation{ \text{Given}&\\ &S_n = \sum_{n = 1}^{\infty}\,a_n = \sum_{n = 1}^{\infty}\,a\cdot r^{n - 1}\;\text{where}\; \left\{\array{{ll} \split{ a &= a_1\\ r &= \frac{S_2}{S_1} } }\right.\\ \text{Alternatively}&\\ &S_n = \sum_{n=0}^{\infty}\,a_n = \sum_{n=0}^{\infty}\,a\cdot r^{n}\\ \text{Tests}&\\ &\text{If}\;|r|\geq{1}\;\text{then}\;\mathbf{\text{$S_n$ is divergent}}\\ &\text{If}\;|r|<1\;\text{then}\;\mathbf{\text{$S_n$ is convergent}}\\ \text{Furthermore}&\\ &\sum_{n = 1}^{\infty}\,a_n = \sum_{n = 1}^{\infty}\,a\cdot r^{n - 1} = \frac{a}{1 - r}\;\text{for all}\;|r|<1 } } \note{ \h3{The Integral Test} \equation{ &\text{Given}\\ &\;\;\;\;a_n = f(n)\;\text{$\forall$n on}\;[1,n)\\ &\;\;\;\;S_n = \sum_{n = 1}^{\infty}\,a_n\\ &\;\;\;\;F(x) = \int_{1}^{\infty}f(x)\;\mathrm{d}x\\ &\text{Where}\\ &\;\;\;\;f(x) > 0, \forall\,x\in\,[1, \infty)\\ &\;\;\;\;f^\prime(x) < 0, \forall\,x\in\,[1, \infty)\\ &\text{Tests}\\ &\;\;\;\;\text{If $S_n$ convergent; then $F(x)$ is $\mathbf{convergent}$}\\ &\;\;\;\;\text{If $S_n$ divergent; then $F(x)$ is $\mathbf{divergent}$} } \note{ \h4{Constraints on \{[1,n)}} \ul { \li{Continuous} \li{Positive} \li{Decreasing (i.e. use derivative test)} } } } \note{ \h3{P-Series -or- Harmonic Series} \equation{ &\text{Given}\\ &\;\;\;\;S_n=\sum_{n=1}^{\infty}\frac{1}{n^p}\\ &\text{Tests}\\ &\;\;\;\;\text{If $p>1$ then $S_n$ is $\mathbf{convergent}$}\\ &\;\;\;\;\text{If $0 < p \leq{1}$ then $S_n$ is $\mathbf{divergent}$} } \note{ Note: the Harmonic series is the special case where \{p=1} } } \note{ \h3{Comparison Test} \equation{ &\text{Given}\\ &\;\;\;\;A_n = \sum_{n=1}^\infty\,a_n\\ &\;\;\;\;B_n = \sum_{n=1}^\infty\,b_n\\ &\text{Where}\\ &\;\;\;\;a_n, b_n \geq 0\\ &\;\;\;\;a_n \leq b_n\\ &\text{Tests}\\ &\;\;\;\;\text{If $B_n$ converges $\implies A_n$ converges}\\ &\;\;\;\;\text{If $A_n$ diverges $\implies B_n$ diverges}\\ } } \note{ \h3{Limit Comparison Test} \equation{ &\text{Given}\\ &\;\;\;\;A_n = \sum_{n=1}^\infty\,a_n\\ &\;\;\;\;B_n = \sum_{n=1}^\infty\,b_n\\ &\;\;\;\; L = \lim_{n\to\infty}\frac{a_n}{b_n}\\ &\text{Where}\\ &\;\;\;\;L > 0,\;L \neq \pm \infty\\ &\text{Therefore either both converge or diverge}\\ &\;\;\;\;\text{$A_n$ converges $\Longleftrightarrow B_n$ converges}\\ &\;\;\;\;\text{$A_n$ diverges $\Longleftrightarrow B_n$ diverges} } \note{ \h4{Warning} \ul { \li{If \{L > 0}, this only means that the limit comparison test can be used. You still need to determine if either \{A_n} or \{B_b} converges or diverges.} \li{Therefore, this does not apply to any arbitrary rational function.} } } \note{ \h4{Notes} \ul { \li{For many series, we find a suitable comparison, \{B_n}, by keeping only the highest powers in the numerator and denominator of \{A_n}.} } } } \note{ \h3{Estimating Infinite Series} \math[block]{ \cdots } } } \h2{Differential Equations} \layout{ \note{ \h3{Separable Differential Equations} \equation{ \text{Given}\\ \frac{\mathrm{d}y}{\mathrm{d}x} &= g(x) \cdot f(y)\\ &= \frac{g(x)}{\frac{1}{f(y)}}\\ &= \frac{g(x)}{h(y)}\;\text{where}\;h(x) = \frac{1}{f(y)}\\ \\\text{Therefore (restated)}\\ \frac{\mathrm{d}y}{\mathrm{d}x} &= \frac{g(x)}{h(y)}\\ \\\text{Multiply reciprocals}\\ h(y)\;\mathrm{d}y &= g(x)\;\mathrm{d}x\\ \\\text{Integrate}\\ \int h(y)\;\mathrm{d}y &= \int g(x)\;\mathrm{d}x\\ \\\text{Differentiate}\\ \frac{d}{dx} \left(\int h(y)\;\mathrm{d}y\right) &= \frac{d}{dx} \left(\int g(x)\;\mathrm{d}x\right)\\ \\\text{Given}\\ \frac{\mathrm{d}}{dx} &= \frac{\mathrm{d}}{dx} \cdot \frac{\mathrm{d}y}{\mathrm{d}y} = \textcolor{blue}{\frac{\mathrm{d}}{dy}} \cdot \textcolor{JungleGreen}{\frac{\mathrm{d}y}{dx}}\\ \\\text{Therefore the LHS is equal to}\\ \frac{\mathrm{d}}{dx} \left(\int h(y) \mathrm{d}y\right) &= \textcolor{blue}{\frac{\mathrm{d}}{dy}} \left(\int h(y) \mathrm{d}y\right) \textcolor{JungleGreen}{\frac{\mathrm{d}y}{dx}}\\ &= h(y) \textcolor{JungleGreen}{\frac{\mathrm{d}y}{dx}} \\\text{Therefore (in conclusion)}\\ \therefore \; h(y)\;\frac{\mathrm{d}y}{dx} &= g(x) } } \note{ \h3{Growth and Decay Models} \equation{ \text{Given}\\ \frac{\mathrm{d}y}{\mathrm{d}t} &= \textcolor{Periwinkle}{k} y\\ \text{Proof}\\ \frac{1}{y}\;\mathrm{d}y &= \textcolor{Periwinkle}{k}\;\mathrm{d}t \\ \int \frac{1}{y}\;\mathrm{d}y &= \int \textcolor{Periwinkle}{k}\;\mathrm{d}t \\ \ln(y) &= \textcolor{Periwinkle}{k} t + \textcolor{SeaGreen}{C} \\ e^{\ln(y)} &= e^ {\textcolor{Periwinkle}{k} t + \textcolor{SeaGreen}{C}} \\ y &= e^{\textcolor{Periwinkle}{k} t} \cdot e^{\textcolor{SeaGreen}{C}} \\ y &= e^{\textcolor{Periwinkle}{k} t} \cdot \textcolor{SeaGreen}{C} \\ \text{Therefore}\\ y &= \textcolor{SeaGreen}{C} e^{\textcolor{Periwinkle}{k} t} } The above states that all solutions for \{y^\prime = k y} are of the form \{y = C e^{k t}}. Where \equation{ \textcolor{SeaGreen}{C} &= \textcolor{SeaGreen}{\text{Initial value of $y$}}\\ \textcolor{Periwinkle}{k} &= \textcolor{Periwinkle}{\text{Proportionality constant}}\\ } Exponential growth occurs when \{\textcolor{Periwinkle}{k > 0}}, and exponential decay occurs when \{\textcolor{Periwinkle}{k < 0}}. \hr The Law of Natural Growth: \equation{ \frac{\mathrm{d}\textcolor{DarkOrchid}{P}}{\mathrm{d}t} = \textcolor{Periwinkle}{k} \textcolor{DarkOrchid}P } The Logistic Model of Population Growth: \equation{ \frac{\mathrm{d}\textcolor{DarkOrchid}P}{\mathrm{d}t} = \textcolor{Periwinkle}{k} \textcolor{DarkOrchid}P \left(1 - \frac{\textcolor{DarkOrchid}{P}}{\textcolor{Aquamarine}{L}}\right) } Where \equation{ \textcolor{Periwinkle}{k} &= \textcolor{Periwinkle}{\text{Constant of proportionality}}\\ \textcolor{DarkOrchid}{P} &= \textcolor{DarkOrchid}{\text{Population at time $t$}}\\ \textcolor{Aquamarine}{L} &= \textcolor{Aquamarine}{\text{Max size of population}}\\ } } \note{ \h3{Solving the Logistic Equation} \equation{ \D{P}{t} &= kP \parens{1 - \frac{P}{L}} \\ \s{d}P &= kP \parens{1 - \frac{P}{L}} \s{d}t \\ \reciprocal{P \parens{1 - \frac{P}{L}}} \s{d}P &= k \s{d}t } Via partial fraction decomposition \equation{ \reciprocal{P \parens{1 - \frac{P}{L}}} &= \frac{A}{P} + \frac{B}{\parens{1 - \frac{P}{L}}}\\ 0P + 1 &= A\parens{1 - \frac{P}{L}} + BP\\ 0P + 1 &= A - \frac{P A}{L} + BP\\ 0P + 1 &= P\parens{B - \frac{A}{L}} + A \\ \text{Where}\\ \left.\array{{ll} B - \frac{A}{L} &= 0\\ A &= 1 }\right\} \;\array{{ll} A &= 1\\ B &= \reciprocal{L} }\\ \text{Therefore}\\ \reciprocal{P \parens{1 - \frac{P}{L}}} &= \frac{1}{P} + \frac{\reciprocal{L}}{\parens{1 - \frac{P}{L}}} } Rewriting the differential equation \equation{ \reciprocal{P \parens{1 - \frac{P}{L}}} \s{d}P &= \frac{1}{P} + \frac{\reciprocal{L}}{\parens{1 - \frac{P}{L}}} \s{d}P\\ &= k \s{d}t \\ \int \frac{1}{P} + \frac{\reciprocal{L}}{\parens{1 - \frac{P}{L}}} &= \int k \s{d}t \\ \ln\; P - \reciprocal{L} \ln\parens{1 - \frac{P}{L}} &= k t + C\\ \ln\parens{\frac {P} {L \cdot \parens{1 - \frac{P}{L}}} } } } } \h2{Second Order Homogeneous Linear Differential Equations with Constant Coefficients} \layout{ \note{ \h3{Properties} \ul { \li{If \{f(x)} and \{g(x)} are solutions; then \{f + g} is also a solution. Therefore, the most general solution to some second order homogeneous linear differential equations with constant coefficients would be \{y = C_1 f(x) + C_2 g(x)}. This is why, when you find two solutions to the characteristic equation \{r_1} and \{r_2} respectively, we write it like so.} } } \note{ \h3{\{r_1 = r_2}} Given some: \equation{ a y^{\prime\prime} + b y^\prime + c y = 0\;\text{where $a \neq 0$}\\ } We can presume that \{y} is of the form \{e^{r t}}, and therefore: \equation{ \text{if}\\ y &= e^{r t}\\ \text{then}\\ y^{\prime} &= r e^{r t}\\ y^{\prime\prime} &= r^2 e^{r t} } Substituting this back into the original equation, we have: \equation{ 0 &= a r^2 e^{r t} + b r e^{r t} + c e^{r t}\\ &= e^{r t} \left(a r^2 + b r + c\right) } Where: \equation{ &= \underbrace{e^{r t}}_\text{never $0$} \underbrace{\left(a r^2 + b r + c\right)}_\text{?} } So therefore: \equation{ &\underbrace{a r^2 + b r + c = 0} _\text{characteristic equation}\\ &r = \frac {-b \pm \sqrt{b^2 - 4ac}} {2a} \implies r_1, r_2 } Where the general solution is of the form: \equation{ \left. \array{{ll} y &=\; &C_1\; e^{r_1 t} &+\; C_2\; e^{r_2 t}\\ y^\prime &=\; &C_1\; r_1\; e^{r_1 t} &+\; C_2\; r_2\; e^{r_2 t}\\ y^{\prime\prime} &=\; &C_1\; \left(r_1\right)^2\; e^{r_1 t} &+\; C_2\; \left(r_2\right)^2\; e^{r_2 t} } \right\}\text{$\forall r_1 \; r_2$ where $r_1 \neq r_2$} } } } \h2{Parametric Equations} \layout{ \note{ \h3{First Derivative Formula} To find the derivative of a given function defined parametrically by the equations \{x = u(t)} and \{y = v(t)}. \equation{ \text{Given}\\ x &= u(t)\\ y &= v(t)\\ \text{Therefore}\\ \frac{\mathrm{d}y}{\mathrm{d}x} &= \frac {\frac{\mathrm{d}y}{\mathrm{d}t}} {\frac{\mathrm{d}x}{\mathrm{d}t}} = \frac{v^\prime(t)}{u^\prime(t)} } } \note{ \h3{Second Derivative Formula} To find the second derivative of a given function defined parametrically by the equations \{x = u(t)} and \{y = v(t)}. \equation{ \text{Given}\\ x &= u(t)\\ y &= v(t)\\ \text{Therefore}\\ \frac{\mathrm{d}{^2} y}{\mathrm{d}x^2} &= \frac{\mathrm{d}}{\mathrm{d}x}\left(\frac{\mathrm{d}y}{\mathrm{d}x}\right)\\ &= \frac { \frac{d}{\mathrm{d}t}\left( \frac{\mathrm{d}y}{\mathrm{d}x} \right) } {\frac{\mathrm{d}x}{\mathrm{d}t}}\\ &= \frac { \frac{\mathrm{d}}{\mathrm{d}t} \left(\frac{v^\prime(t)}{u^\prime(t)}\right) } {u^\prime(t)}\\ &= \underbrace { \frac { \frac{\mathrm{d}}{\mathrm{d}t} \left(\frac{v^\prime(t)}{u^\prime(t)}\right) } { \frac{\mathrm{d}}{\mathrm{d}t} u(t) } = \frac { \frac{\mathrm{d}}{\mathrm{d}t} } { \frac{\mathrm{d}}{\mathrm{d}t} } \frac { \left(\frac{v^\prime(t)}{u^\prime(t)}\right) } { u(t) } }_{\text{notice the common $\frac{\mathrm{d}}{\mathrm{d}t}$}} } The above shows different ways of representing \{\frac{\mathrm{d}^{2}y}{\mathrm{d}x^2}}. (I.e. it doesn't correspond to some final solution.) } \note{ \h3{Arc Length} Formula for the arc length of a parametric curve over the interval \{[a, b]}. \equation{ \int_a^b \sqrt{ \left(\frac{\mathrm{d}x}{\mathrm{d}t}\right)^2 + \left(\frac{\mathrm{d}y}{\mathrm{d}t}\right)^2 } \mathrm{d}t } } }