You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In much of this lecture, we'll think of $X$ as a matrix of **data** in which
40
38
41
-
* each column is an **individual** -- a time period or person, depending on the application
39
+
* each column is an **individual** -- a time period or person, depending on the application
42
40
43
-
* each row is a **random variable** describing an attribute of a time period or a person, depending on the application
41
+
* each row is a **random variable** describing an attribute of a time period or a person, depending on the application
44
42
45
43
46
44
We'll be interested in two situations
47
45
48
-
* A **short and fat** case in which $m << n$, so that there are many more columns (individuals) than rows (attributes).
46
+
* A **short and fat** case in which $m << n$, so that there are many more columns (individuals) than rows (attributes).
49
47
50
-
* A **tall and skinny** case in which $m >> n$, so that there are many more rows (attributes) than columns (individuals).
48
+
* A **tall and skinny** case in which $m >> n$, so that there are many more rows (attributes) than columns (individuals).
51
49
52
50
53
51
We'll apply a **singular value decomposition** of $X$ in both situations.
@@ -116,29 +114,20 @@ Thus,
116
114
117
115
We'll apply this circle of ideas later in this lecture when we study Dynamic Mode Decomposition.
118
116
119
-
120
-
121
-
122
-
123
117
**Road Ahead**
124
118
125
119
What we have described above is called a **full** SVD.
126
120
127
-
128
-
129
121
In a **full** SVD, the shapes of $U$, $\Sigma$, and $V$ are $\left(m, m\right)$, $\left(m, n\right)$, $\left(n, n\right)$, respectively.
130
122
131
123
Later we'll also describe an **economy** or **reduced** SVD.
132
124
133
125
Before we study a **reduced** SVD we'll say a little more about properties of a **full** SVD.
134
126
135
-
136
127
## Four Fundamental Subspaces
137
128
138
-
139
129
Let ${\mathcal C}$ denote a column space, ${\mathcal N}$ denote a null space, and ${\mathcal R}$ denote a row space.
140
130
141
-
142
131
Let's start by recalling the four fundamental subspaces of an $m \times n$
143
132
matrix $X$ of rank $p$.
144
133
@@ -263,14 +252,12 @@ $$
263
252
\end{aligned}
264
253
$$ (eq:fourspaceSVD)
265
254
266
-
267
-
268
255
Since $U$ and $V$ are both orthonormal matrices, collection {eq}`eq:fourspaceSVD` asserts that
269
256
270
-
* $U_L$ is an orthonormal basis for the column space of $X$
271
-
* $U_R$ is an orthonormal basis for the null space of $X^\top $
272
-
* $V_L$ is an orthonormal basis for the row space of $X$
273
-
* $V_R$ is an orthonormal basis for the null space of $X$
257
+
* $U_L$ is an orthonormal basis for the column space of $X$
258
+
* $U_R$ is an orthonormal basis for the null space of $X^\top $
259
+
* $V_L$ is an orthonormal basis for the row space of $X$
260
+
* $V_R$ is an orthonormal basis for the null space of $X$
274
261
275
262
276
263
We have verified the four claims in {eq}`eq:fourspaceSVD` simply by performing the multiplications called for by the right side of {eq}`eq:fullSVDpartition` and reading them.
@@ -286,8 +273,6 @@ Sometimes these properties are described with the following two pairs of orthogo
286
273
* ${\mathcal C}(X)$ is the orthogonal complement of $ {\mathcal N}(X^\top )$
287
274
* ${\mathcal R}(X)$ is the orthogonal complement ${\mathcal N}(X)$
288
275
289
-
290
-
291
276
Let's do an example.
292
277
293
278
@@ -340,13 +325,13 @@ Suppose that we want to construct the best rank $r$ approximation of an $m \tim
340
325
341
326
By best, we mean a matrix $X_r$ of rank $r < p$ that, among all rank $r$ matrices, minimizes
342
327
343
-
$$ || X - X_r || $$
328
+
$$
329
+
|| X - X_r ||
330
+
$$
344
331
345
332
where $ || \cdot || $ denotes a norm of a matrix $X$ and where $X_r$ belongs to the space of all rank $r$ matrices
346
333
of dimension $m \times n$.
347
334
348
-
349
-
350
335
Three popular **matrix norms** of an $m \times n$ matrix $X$ can be expressed in terms of the singular values of $X$
351
336
352
337
* the **spectral** or $l^2$ norm $|| X ||_2 = \max_{||y|| \neq 0} \frac{||X y ||}{||y||} = \sigma_1$
@@ -369,12 +354,6 @@ You can read about the Eckart-Young theorem and some of its uses [here](https://
369
354
370
355
We'll make use of this theorem when we discuss principal components analysis (PCA) and also dynamic mode decomposition (DMD).
371
356
372
-
373
-
374
-
375
-
376
-
377
-
378
357
## Full and Reduced SVD's
379
358
380
359
Up to now we have described properties of a **full** SVD in which shapes of $U$, $\Sigma$, and $V$ are $\left(m, m\right)$, $\left(m, n\right)$, $\left(n, n\right)$, respectively.
@@ -385,7 +364,6 @@ Thus, note that because we assume that $X$ has rank $p$, there are only $p$ nonz
385
364
386
365
A **reduced** SVD uses this fact to express $U$, $\Sigma$, and $V$ as matrices with shapes $\left(m, p\right)$, $\left(p, p\right)$, $\left( n, p\right)$.
0 commit comments