You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Lecture 14. Low Rank Changes in A and Its Inverse
14. Low Rank Changes in A and Its Inverse
The video discusses the concept of low-rank matrices and their importance in function matrices, particularly the matrix inversion formula which finds the inverse of an N by n matrix in terms of a simpler 1 by 1 matrix. The formula is useful in finding the inverse of matrices that have low rank perturbations and can simplify the process of finding inverses. The speaker showcases how the formula works by presenting the formula for the second matrix and shows how the same logic was applied to arrive at the answer. The video also discusses practical applications of this formula, particularly in least squares problems and the Kalman filter.
Lecture 15. Matrices A(t) Depending on t, Derivative = dA/dt
15. Matrices A(t) Depending on t, Derivative = dA/dt
This video covers various topics related to matrices, including changes in matrices and their inverse, as well as changes in eigenvalues and singular values over time. The speaker explains key formulas for calculating these changes and emphasizes the importance of understanding calculus in linear algebra. Additionally, the lecture discusses the importance of normalization and explores interlacing theorems for eigenvalues in both symmetric and rank 1 matrices. Lastly, the video concludes with a review of the topics covered and a promise to expand on them in future lectures.
possible, they can still derive inequalities to understand how big the change could be. The lecture also covers the setup of the matrix A, which depends on time (T) and the inverse A inverse.
Lecture 16. Derivatives of Inverse and Singular Values
16. Derivatives of Inverse and Singular Values
This video covers a variety of topics including the derivative of the inverse and singular values of a matrix, interlacing, and the nuclear norm of a matrix. The speaker presents a formula for the derivative of singular values, using the SVD, to understand how a matrix changes over time, while establishing bounds for changes in eigenvalues in symmetric matrices. Vial's inequality is introduced as a way to estimate the lambda values of a matrix, and basis pursuit is used in matrix completion problems. The speaker also discusses the idea that the nuclear norm of a matrix comes from a norm that is not quite a norm and introduces the concept of Lasso and compressed sensing to be discussed in the next lecture.
Lecture 17: Rapidly Decreasing Singular Values
Lecture 17: Rapidly Decreasing Singular Values
The lecture focuses on matrices and their ranks, and how rapidly decreasing singular values are prevalent in computational mathematics. The lecturer examines low rank matrices and demonstrates how they have a lot of zeros in their sequence of singular values, making it more efficient to send the matrix to a friend in low rank form than in full rank form. They also introduce the numerical rank of a matrix, which is defined by allowing some wiggle room to define the tolerance of singular values of a matrix. By sampling smooth functions, which can be well-approximated by polynomials, the numerical rank can be low, resulting in a low-rank approximation of the matrix X. The lecture also includes examples of Gaussian and Vandermonde matrices to explain how they can lead to matrices of low rank, and discusses the usefulness of Zolotarev numbers in bounding singular values.
Lecture 18: Counting Parameters in SVD, LU, QR, Saddle Points
Lecture 18: Counting Parameters in SVD, LU, QR, Saddle Points
In this lecture, the speaker reviews various matrix factorizations such as L&U, Q&R, and eigenvector matrices and counts the number of free parameters in each of these matrices. They also discuss the computation of Qs versus SVD and count the number of parameters in the SVD for a rank-R matrix. The lecturer also explains the concept of saddle points in matrices and how to find them using optimization techniques and Lagrange multipliers. Lastly, the lecturer discusses the sign of eigenvalues of a symmetric matrix and how the Rayleigh quotient can help determine the maximum value and corresponding eigenvector of the matrix.
Lecture 19. Saddle Points Continued, Maxmin Principle
19. Saddle Points Continued, Maxmin Principle
In this video, the speaker continues discussing saddle points and how to find minimum and maximum values using the Rayleigh quotient in two-dimensional space. The interlacing theorem is explained, which involves writing saddle points as the maximum of a minimum to quickly find maxima and minima. The speaker also warns against overfitting when fitting data with a high-degree polynomial and discusses two open-ended labs for the class, involving saddle points and a simple neural network. The concepts of mean and variance in statistics and sample variance and covariance are explained, with the speaker noting that the covariance matrix for totally dependent outputs would be not invertible, and for polling scenarios with multiple people living in one house, some covariance is expected but not entirely independent.
Lecture 20. Definitions and Inequalities
20. Definitions and Inequalities
In this section of the video, the speaker discusses various concepts in probability theory including expected value, variance, and covariance matrices. Markov's inequality and Chebyshev's inequality were also introduced as fundamental tools for estimating probabilities. The speaker then proceeds to explain the relationship between Markov's inequality and Chebychev's inequality, illustrating how they lead to the same result. The concept of covariance and covariance matrix, a fundamental tool in probability theory, was also introduced. The video also explores the idea of joint probabilities and tensors, explaining how gluing coins together adds dependence and alters the probabilities. Finally, the speaker discusses the properties of the covariance matrix, emphasizing that it's always positive semi-definite and is a combination of rank 1 positive semi-definite matrices.
Lecture 21: Minimizing a Function Step by Step
Lecture 21: Minimizing a Function Step by Step
This video lecture discusses the basic algorithms used for minimizing a function and their convergence rates, particularly Newton's method and steepest descent. It also highlights the importance of convexity, which ensures that the function has one minimum, and introduces the concept of convex sets and convex functions. The lecturer explains how to test for convexity in a function, which determines whether it has saddle points or local minimums, as opposed to a global minimum. The video concludes with a discussion of Levenberg Marquardt, a cheaper version of Newton's method that is not fully second-order.
Lecture 22. Gradient Descent: Downhill to a Minimum
22. Gradient Descent: Downhill to a Minimum
In the video, "Gradient Descent: Downhill to a Minimum," the speaker discusses the importance of gradient descent in optimization and deep learning, where the goal is to minimize a function. The speaker introduces the gradient and the Hessian, and illustrates the steps of steepest descent using a quadratic function. The speaker also discusses how to interpret the gradient and the Hessian, as well as their role in measuring convexity. The speaker delves into choosing the appropriate learning rate, stressing the importance of the condition number in controlling the speed of convergence. The video also provides practical examples and formulas to help understand the concept of gradient descent, including the heavy ball method.
Lecture 23. Accelerating Gradient Descent (Use Momentum)
23. Accelerating Gradient Descent (Use Momentum)
This video discusses the concept of momentum in accelerating gradient descent. The presenter explains the basic gradient descent formula and shows how adding momentum can result in faster descent than the ordinary method, ultimately yielding significant improvements. They also discuss a continuous model of steepest descent and explain how it can be analyzed as a second-order differential equation with a momentum term. The presenter emphasizes the importance of minimizing both eigenvalues when using momentum to minimize the largest eigenvalue by choosing values for s and beta to make the eigenvalues of the matrix as small as possible. They also discuss Nesterov's method and suggest that it may be possible to get further improvements by going back two or three steps or more.