Optimization Engineering For Machine Learning and AI
Optimization Engineering For Machine Learning and AI
Optimization is a core fundamental area for machine learning and AI in general. Moreover, Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing or maximizing convex functions over convex sets. Many classes of convex optimization problems admit polynomial-time algorithms, whereas mathematical optimization is in general NP-hard.
In the first lesson/lecture of this course, we will talk about the following points:
What is Optimization?
Examples on Optimization
Factors of Optimization
Reliable/Efficient Problems
Goals & Topics of this Course
Brief History on Optimization
=======
In the second lesson/lecture, we will be covering important points on convex sets, which are the following:
00:00:00 Affine Combination
00:01:33 Affine Set
00:08:21 Convex Combination
00:09:25 Convex Set
00:13:45 Convex Hull
00:16:28 Example 1-Convex Cones
00:16:55 Conic Combination
00:20:47 Example 2-Hyperplanes
00:24:22 Example 3-Euclidean Ball
00:26:37 Example 4-Ellipsoid
00:30:40 Norms
00:35:51 Example 5-Polyhedra
00:41:18 Example 6-Positive Semidefinite cone
00:54:31 Operations preserving convexity
01:15:10 Closed & Open set
01:21:10 Solid sets
01:26:25 Pointed set
01:26:57 Proper cones
01:27:28 Generalized Inequalities
01:34:12 Minimum & Minimal Elements
01:46:28 Partial Order
01:48:53 Properties of Generalized Inequalities
01:53:09 Dual Cones
02:04:31 Dual Inequalities
=======
In the third lesson/lecture of this course on convex optimization, we will be covering important points on convex functions, which are the following:
00:01:14 Definition of Convex Function
00:03:31 Examples of Convex Function
00:13:50 Convexity in Higher Dimensions
00:24:30 First-order Condition
00:27:08 Second-order Conditions
00:35:17 Epigraphs
00:37:25 Jensen's Inequality
00:39:49 Operations preserving Convexity
00:52:21 Conjugate Convex function
01:02:09 Quasi Convex functions
01:11:14 Log-Convex functions
01:16:16 Convexity with respect to generalized inequalities
=======
In Lecture 4 of this course on convex optimization, we will be covering the fundamental principles of convex optimization, which include the following:
00:00 Standard form
04:19 Feasible point
05:07 Globally Optimum point
05:50 Locally Optimum point
15:04 Explicit & Implicit constraints
30:10 Optimality criterion for differentiable cost functions
34:48 Supporting Hyperplane
=======
In Lecture 5 of this course on convex optimization, we will be covering Linear Programming and the Simplex algorithm, which was introduced by George Dantzig. The outline of the lecture is as follows:
00:00:00 What is a Linear Program (LP) ?
00:07:24 LP feasible set
00:10:22 LP forms
00:10:50 Standard form LP
00:10:50 Standard form LP
00:11:24 Slack variables
00:13:08 Inequality form LP
00:13:34 Omitting inequality constraints
00:20:38 LP Example: The Diet Problem
00:25:49 The SIMPLEX Algorithm: Method and the usage of Non-basic, Slack, and Artificial variables
00:33:59 The SIMPLEX Algorithm - Example: Iteration 0
00:40:37 The SIMPLEX Algorithm - Example: Iteration 1
00:48:18 The SIMPLEX Algorithm - Example: Iteration 2
00:55:27 The SIMPLEX Algorithm - Example: Iteration 3
01:00:13 MATLAB: Implementing SIMPLEX
01:53:15 MATLAB: Verifying with linprog
01:58:48 George Bernard Dantzig
01:59:12 SIMPLEX: Geometric Interpretation
02:01:09 SIMPLEX: Time Complexity
=======
In Lecture 6 of this course on convex optimization, we will cover the essentials of Quadratic Programming.The outline of the lecture is as follows:
00:00 Intro
00:32 What is a Quadratic Program (QP) ?
03:24 QP reformulation
06:05 Illustrating the optimal solution
16:54 Solving a QP on MATLAB
25:43 Outro
=======
In Lecture 7 of this course on convex optimization, we will cover the essentials of Quadratically Constrained Quadratic Programs, i.e. QCQPs.The outline of the lecture is as follows:
00:00 Intro
00:33 What is a Quadratically Constrained Quadratic Program (QCQP) ?
05:16 QCQP Feasible Set
06:01 MATLAB Illustration of QCQP Feasible Set
13:39 QCQP Application 1: Minimizing a linear function over a centered ellipsoid
30:42 QCQP Application 2: Minimizing a linear function over an uncentered ellipsoid
37:16 QCQP Application 3: Minimizing a quadratic function over a centered ellipsoid
42:36 Outro
=======
In Lecture 8 of this course on convex optimization, we will cover Second Order Cone Programming, i.e. SOCPs. The outline of the lecture is as follows:
00:00:00 What is Second Order Cone Programming (SOCP) ?
00:02:37 QCQP as an SOCP
00:20:25 Robust Linear Programming as an SOCP
00:31:06 Linear Programming with Random Constraints as an SOCP
00:41:09 Sum of Norms minimization as an SOCP
00:47:27 Max of Norms minimization as an SOCP
00:49:40 Hyperbolic Constrained Programs as SOCPs
00:58:59 Quadratic Fractional Problems as SOCPs
01:02:16 Outro
==========
In Lecture 9 of this course on convex optimization, we will cover Geometric Programs, i.e. GPs. The outline of the lecture is as follows:
00:00 Intro
01:51 Monomials and Posynomials
10:45 GP problem formulation (polynomial form)
19:50 Relevant papers
23:12 GP in convex form
29:27 Example 1: Frobenius Norm Diagonal Scaling
33:25 Example 2: Volume Maximization given aspect ratios and area limitations
38:12 Summary
=====
In Lecture 10 of this course on convex optimization, we will cover Generalized Geometric Programs, i.e. GPs. The outline of the lecture is as follows:
00:00 Intro
01:16 Generalized Posynomials
08:46 Generalized Geometric Program (GGP)
09:45 GGP as a GP
17:40 Example 1: Floor Planning (GGP)
23:48 Example 2: Power Control (GP)
33:00 Example 3: Digital Circuit Design (GGP)
37:26 Mixed-Integer Geometric Program
39:27 Outro
======
In Lecture 11 of this course on convex optimization, we will cover Semidefinite programming, i.e. SDPs. The outline of the lecture is as follows:
00:00 Intro
01:05 Generalized Inequality Constraints
05:18 Conic Programs
07:59 Linear Matrix Inequality (LMI)
09:56 LMI brief history (Lyapunov, Kalman, Ricatti etc..)
18:10 Semidefinite Programming (SDP)
21:56 SOCP as SDP
29:30 Eigenvalue Minimization
32:43 Matrix Norm Minimization
34:39 Outro
=====
In Lecture 12 of this course on convex optimization, we will cover various topics related to Vector optimization, such as Pareto optimal points and the Pareto frontier, which is a well known boundary studied in Game theory, risk and trade-off analysis, portfolio analysis, etc. The topics covered are outlined as follows:
00:00:00 Intro
00:01:55 What is Vector Optimization ?
00:06:38 Optimal points & the set of achievable objective values
00:13:27 Pareto optimal points
00:18:56 BLUE estimator (example)
00:28:09 Scalarization
00:32:03 Pareto Frontier (Boundary)
00:38:28 Minimal Upper Bound on a set of matrices (example)
00:43:36 Plotting a Pareto front of regularized least squares on MATLAB (1st way: the genetic algorithm)
00:47:43 Plotting a Pareto front of regularized least squares on MATLAB (2nd way: using fminsearch)
00:53:43 Multicriterion optimization
01:01:39 Scalarization for Multicriterion optimization
01:06:51 Analytical Pareto Front of Regularized Least Squares
01:09:44 Plotting a Pareto front of regularized least squares on MATLAB (3rd way: Analytically)
01:12:08 Outro
======
In Lecture 13 of this course on convex optimization, we will cover various topics related to Vector optimization, such as Pareto optimal points and the Pareto frontier, which is a well known boundary studied in Game theory, risk and trade-off analysis, portfolio analysis, etc. The topics covered are outlined as follows:
00:00 Intro
00:29 Reminder: Multicriterion Optimization
03:17 Multicriterion Optimization: A closer look
09:02 Optimal Trade-off Analysis
12:38 Outro
========
In Lecture 14 of this course on Convex Optimization, we introduce the Lagrangian duality theory. In essence, for each optimization problem (convex or not), we can associate a certain function referred to as the Lagrangian function. This function, in turn, has a dual function (which serves as an infimum over the variable of interest x). It turns out that, for any optimisation problem, the dual function is a lower bound on the optimal value of the optimisation problem in hand. This lecture focuses on many examples that derive the Lagrangian and the associated dual functions. MATLAB implementations are also presented to give useful insights. This lecture is outlined as follows:
00:00 Intro
01:00 Lagrangian function and duality
04:02 Lagrangian dual function
06:46 Lower bound on the optimal value
09:16 MATLAB: Lower bound verification
15:28 Example 1 - Least Squares
17:48 Example 2 - Linear Programming
20:48 Example 3 - Two-way Partitioning
26:04 Relationship between conjugate function and the dual function
31:22 Example 4 - Equality Constrained Norm minimization
33:37 Example 5 - Entropy Maximization
35:44 Outro
========
In Lecture 15 of this course on Convex Optimization, we talk about a very very important topic in convex optimisation that is the Lagrange Dual Problem. This lecture is outlined as follows:
00:00:00 Intro
00:00:44 Revision: Lagrange Dual Function
00:01:30 The Dual Problem
00:06:54 Example 1: Dual Problem of Standard LP
00:08:59 Example 2: Dual Problem of Inequality LP
00:13:59 Weak Duality
00:16:24 Example 3: The 2-way Partitioning Problem (Revisited)
00:21:42 Strong Duality
00:23:15 Slater’s Condition
00:24:32 What is a Relative Interior (Convex Analysis by Tyrell Rockefellar) ?
00:28:16 Generalization of Slater’s Condition
00:29:26 Example 4: Duality of LS problems
00:38:33 Example 5: Duality of LP problems
00:54:52 Example 6: Duality of QCQP problems
00:59:22 Example 7 : Duality of the Entropy Maximization Problem
01:03:48 Example 8 : Duality of the Trust Region Problem (non-convex problem)
01:11:51 Outro
======
In Lecture 16 of this course on Convex Optimization, we talk about a very practical topic, when it comes to numerical optimization algorithms, and that is the ε-suboptimal inequality, which could report how good of an estimate we have. Said differently, the Lagrangian dual feasible points (λ,ν) provides a proof or certificate of the dual gap.
00:00 Intro
01:59 Lagrangian & Dual Functions
03:40 How good of an estimate ? (Certificate)
07:09 ε-suboptimal
10:09 Stopping Criterion
17:23 Outro
==============
In Lecture 17 of this course on Convex Optimization, we talk about Complementary Slackness, which could be used a test for optimality, or it could even tell us which constraints are active and which are not !!
This lecture is outlined as follows:
00:00 Intro
00:46 What is Complementary Slackness ?
08:15 A Genie Example
14:45 Another Genie Example
16:00 Outro
=====
In Lecture 18 of this course on Convex Optimization, we talk about KKT conditions for nonconvex and convex optimization problems.
This lecture is outlined as follows:
00:00 Intro
00:57 KKT Conditions for NonConvex Problems
04:32 KKT Conditions for Convex Problems
07:48 Example
10:47 Outro
=======
In Lecture 19 of this course on Convex Optimization, we talk about Perturbation and Sensitivity Analysis of general and convex optimization problems.
This lecture is outlined as follows:
00:00 Intro
02:34 Perturbed Optimization Problem
16:33 Global Perturbation Behavior
37:35 Local Perturbation Behavior
49:36 Shadow Price Interpretation
53:40 Outro
=======
In Lecture 20 of this course on Convex Optimization, we talk about Equivalent Reformulations of general and convex optimization problems.
This lecture is outlined as follows:
00:00:00 Intro
00:01:46 Reformulation 1: Introducing new variables
00:25:06 Log-Sum-Exponential Cost
00:33:23 Norm Minimization
00:49:39 Reformulation 1 (cont'd): Introducing constraint variables
01:05:11 Reformulation 2: Cost Transformation
01:14:23 Reformulation 3: Constraint Absorption
01:32:05 Summary
========
In Lecture 21 of this course on Convex Optimization, we talk about the theorem of weak alternatives of general optimization problems.
This lecture is outlined as follows:
00:00 Introduction
04:02 Feasibility Problem
05:41 Optimization Feasibility Problem
07:55 Dual Function
08:41 Note on Strong Alternatives
10:43 Dual Problem
13:12 Weak Duality
13:41 Relating (S) with (T)
15:16 Weak Alternatives
17:31 Why Weak Alternatives ?
19:33 Summary
23:18 Outro
======
In Lecture 22 of this course on Convex Optimization, we talk about the theorem of strong alternatives of convex optimization problems.
This lecture is outlined as follows:
01:43 Introduction
02:13 Strengthening Weak Alternatives
03:21 Strong Alternatives Conditions
08:27 Strict Inequality Case
09:49 Strong Alternatives of Linear Inequalities
13:31 Strong Alternatives of Strict Linear Inequalities
22:46 Strong Alternatives of Intersection of Ellipsoids
34:53 Summary
38:46 Outro
=======
In Lecture 23 of this course on Convex Optimization, we focus on algorithms that solve unconstrained minimization type problems. The lecture evolves around unconstrained minimization problems that might or might not enjoy closed form solutions. Descent methods are discussed along with exact line search and backtracking. MATLAB implementations are given along the way.
This lecture is outlined as follows:
00:00:00 Introduction
00:01:06 Unconstrained Minimization
00:01:36 Iterative Algorithm Assumptions
00:04:28 Gradient Equivalence
00:09:04 Unconstrained Least Squares
00:20:13 Unconstrained Geometric Program
00:28:10 Initial Subset Assumption
00:35:16 Intuitive Solution of Logarithmic Barrier Minimization
00:40:42 Generalization of Logarithmic Barriers
00:42:57 Descent Methods
00:50:42 Gradient Descent
00:52:59 Exact Line Search
00:56:23 Backtracking
01:00:25 MATLAB: Gradient Descent with Exact Line Search
01:17:35 MATLAB: Gradient Descent with Backtracking
01:20:12 MATLAB: Gradient Descent with Explicit Step Size Update
01:28:07 Summary
01:30:59 Outro
===============
References:
[1] Boyd, Stephen, and Lieven Vandenberghe. Convex optimization. Cambridge university press, 2004.
[2] Nesterov, Yurii. Introductory lectures on convex optimization: A basic course. Vol. 87. Springer Science & Business Media, 2013. Reference no. 3:
[3] Ben-Tal, Ahron, and Arkadi Nemirovski. Lectures on modern convex optimization: analysis, algorithms, and engineering applications. Vol. 2. Siam, 2001.
A master class to learn convex optimization for ML and its applications to different fields and areas of engineering
Url: View Details
What you will learn
- Convex optimization theory and concepts for machine learning and AI
- Engineering mathematics of convex optimization for ML, DL, and AI
- Convex optimization methods and techniques in ML, DL, and AI
Rating: 5
Level: All Levels
Duration: 22 hours
Instructor: RS Academy
Courses By: 0-9 A B C D E F G H I J K L M N O P Q R S T U V W X Y Z
About US
The display of third-party trademarks and trade names on this site does not necessarily indicate any affiliation or endorsement of coursescompany.com.
View Sitemap