Num Weekly Summary
- W01/VL01
- date:
 - summary:
 
 - W01/VL02
- date:
 - summary:
 
 - W02/VL03
- date:
 - summary:
 
 - W02/VL04:
- date:
 - summary:
 
 - W03/VL05
- date: 31/10/2023, Tue
 - summary: \(LR\)-Decomposition, uniqueness, existence, algorithm, complexity, error analysis
- Script pages: 21 (satz 2.5) - 28 (2.3 Error Analysis of LR)
 - LR Zerlegung ist eindeutig
 - Product of two triangular matrices is also triangular (left or right)
 - Inverse of a triangular matrix is triangular (left or right)
 - Product of two left-triangular matrices with \(a_{ii} = 1\) is also a left-triangular matrix with \(\tilde{a}_{ii} = 1\)
 - \(LR\)-Decomposition method using Gauss decomposition (Alg 2.7) (?).
 - Existence of the \(LR\)-decomposition (Satz 2.8)
 - Practical version of \(LR\)-decomposition Algorithm (not Alg 2.7) saves space (Alg 2.9).
 - Complexity of Alg 2.9 and solving a LSE (De: LGS)
 - Error analysis of \(LR\)-Decomposition.
 
 
 - W03/VL06
- date: 02/11/23, Thu
 - summary: Script 28 - 33. Error estimation of Alg 2.9, forwards and backwards error, conditioning of a function, conditioning number.
- Error estimation of Alg 2.9 with proof (Lemma 2.13 & Satz 2.14)
 - Aposteriori error estimation of Alg 2.9 (Satz 2.16)
 - relationship between backwards and forwards error, a python example demonstrating that they aren’t necessarily related (?)
 - Conditioning of a function (Ch 2.4.1).
 - Conditioning number
 
 
 - W04/VL07
- date: 07/11/23, Tue
 - summary: Script 33 - 42.
- Question: How much effect does the error in \(A\) & \(b\) have on the solutuion of an LEQ/LGS?
 - Conditioning of a matrix & its proof. (Prop 2.20)
 - Error analysis of LEQ/LGS’s (Ch 2.4.3)
 - Convergent Matrix sequences and matrix series (Ch 2.4.4.)
- Neumann Series (2.4.9, P:37)
 - Spectral radius of a matrix
 
 - Convergent Matrix sequences are used in the proof of error estimation of LEQ/LGS’s. Proof of 2.2.1
 - Pivoting Strategies to increase \(LR\)-decomposition algorithm stability (Ch. 2.5) via exchanging rows during the steps of the algorithm.
- Permutation Matrix (Def 2.25)
 
 
 
 - W04/VL08
- date: 09/11/23, Thu
 - summary: Script 43 - 51
- note: numbering shifted 1 up, due to a new example
 - Alg. 2.27: \(LR\)-Decomposition with column pivot search.
- Satz 2.30: \(\forall\) regular matrix \(A \in \mathbb{R}^{n\times n}\, \exists\) a Permutations matrix \(P_{\pi}\), s.t \(LR = P_{\pi}A\) and its proof.
 - A problematic matrix for this method: Wilkinson matrix. Such problems can be avoided with “Full Pivot Search/Vollpivotsuche”.
 - But fullpivot search has the disatvantageous complexity \(\mathcal{O}(n^3)\)
 
 - Cholesky Decomposition: An efficient method for Symmetric Positive Definite (SPD) Matrices.
- Definition & Properties of SPD matrices. (Def 2.32 & Satz 2.33)
 - Satz 3.34: \(A\in\mathbb{R}^{n\times n} \text{\, SPD \, }\Rightarrow \exists \text{\, upper triangular} \, R\in \mathbb{R}^{n\times n} \text{\, s.t.\, } A = R^TR\) (Cholesky decomposition)
 - Proof of Satz 3.34
 - Algorithm for Cholesky Decomposition (Alg 2.35) and its complexity (\(\frac{n^3}{3} + \mathcal{O}(n^2)\)). (2 times more efficient than usual decommposition)
 
 - \(LR\)-Decomposition for Band-matrices
- Sparce matrices (schwach besetzte matrix) \(\approx\) many entries are 0.
 - how sparce matrices are stored efficiently: For example
 
A = 0 5 0 C = (2, 3) {non-nul columns} 0 0 0 R = (1, 3) {non-null rows} 0 0 7 X = (5, 7) {actual entries in these coordinates}- Definition of a band matrix (Def 2.38)
 - \(LR\)-decomposition for Band matrices.
 
 
 
 - W05/VL09
- date: 14.11.23, Tue
 - summary: Skript 51 - 55 (togehter with topics form appendix)
- Review of some LA topics (from Appendix):
- Def of orthogonal matrix
 - Def of unitary matrix
 - Lemma A.57/Satz A.58: Singular value decomposition (SVD) and its proof.
 - Programming example demonstrating uses of SVD for image compression.
 - Prop A.61 regarding SVD (?)
 
 - Intro to new Ch 3 - Interpolation & Approximation
- Overview of different types of interpolating functions: polynomial, rational (polynomial), spline, neural network
 - Intro Polynomial Interpolation:
- Def of vector space of polynomials of degree \(n\): \(\mathbb{P}_n\).
 - Def 3.1: Lagrange Interpolation Polynomials
 
 
 
 - Review of some LA topics (from Appendix):
 
 - W05/VL10
- date: 16.11.23, Thu
 - summary: Skript: 55 - 64
- Lagrange interpolation
- Lagrange polynomials \(\{l_i\}_{i=0\dots n}\) Form a basis for \(\mathbb{P}_n\) & its proof.
 
 - General interpolation with a general basis \(\{b_i\}_{i=0\dots n}\) \(\Rightarrow\) Vandermonde Matrix. note: Vandermonde matrix for Lagrange Basis is simply the identity matrix. Vandermonde matrix is very badly conditioned for the monomial basis \(\{x^i\}_{i=0}^n\)
 - Error analysis of Lagrange interpolation. Certain properties of the function that is to be interpolated determine the precision of the error analysis (like the smoothness of the function.) (Satz 3.6)
 - Neville’s Schema
 
 - Lagrange interpolation
 
 - W06/VL11
- date: 21.11.23, Tue
 - summary: Skript 64 - 71
- Newton Interpolation (Ch 3.5)
- Dividierende Differenzen
 - Lemma 3.11 and its proof via induction.
 
 - Intepolation error
 - Hermite Interpolation
 - (3.7 is skipped)
 - 3.8 Conditioning and Approximation
 - Weierstrass Theorem (3.25)
 
 - Newton Interpolation (Ch 3.5)
 
 - W06/VL12
- date: 23.11.23, Thu
 - summary: Skript 71 - 81
- review of error of polynomial interpolation: Bsp 3.21, 3.22.
 - Runge Phenomenon
 - Weierstrass Theorem (3.25) and its relationship to polynomial interpolation.
 - Error of the best approximation (Satz 3.26)
- Def 3.23: Lebesgue Constant \(\Lambda_n\)
 
 - Conditioning of the interpolation, what happens when we interpolate \(f + \delta\) instead of \(f\)?
 - If the function is only known to be continuous (but not necessarily differentiable), there is another measure for how strong a function “fluctuates”: Modulus of continuity (Stetigkeitsmodul) (Def 3.27)
 - Jackson’s Theorem (3.28) given without proof.
 - Corollary (3.29) - its proof.
 - How to keep \(\Lambda_n\) small? \(\rightarrow\) Chebyshev Interpolation, Chebyshev Points
- Goal: find appropriate points \(\{x_i\}_{i=1\dots n}\) s.t. \(\Lambda_n\) is minimal.
 - Definition of Chebyshev Points (Def 3.30): \(x_i^{(n + 1)} = \cos{(\frac{2i + 1}{2n + 2}\pi)}\) 

 - Definition of Chebyshev Polynomials (Def 3.32)
 - Lemma 3.33 - some properties of chebyshev polynomials
 
 
 
 - W07/VL13
- date: 28.11.23
 - summary: Skript 81 - 96 (some pages were skipped)
- Clenshaw-Curtis points
 - Satz 3.34
 - Proofs regarding Lebesgue Constats are skipped.
 - Lemma 3.40:
 - Skip until Spline Interpolation
 - Spline Interpolation:
- Error of spline interpolation
 - Cubic splines