from matplotlib import pyplot as plt
import numpy as np
12345)
np.random.seed(
= np.random.randint(1, 3, (5, 4))
a1 = np.random.randint(1, 3, (4, 8))
a2
#constructing our sample matrix A
= a1@a2 + 0.1*np.random.rand(5, 8) A
Unsupervised Learning with Linear Algebra
Introduction to Singular Value Decomposition (SVD)
In one of our lectures, we discussed the concept of matrix factorization methods - where we decompose a matrix into two or more simpler matrices. This is an example of dimensionality reduction, and these simpler matrices allow us to capture the underlying structure and relationships within the data.
Previously, we have discussed Principal Component Analysis (PCA) and Nonnegative Matrix Factorization (NMF) as matrix factorization methods. However, for this blog post we will be looking at: Singular Value Decomposition (SVD). A SVD of a real matrix \(A \in \mathbb{R}^{m \times n}\) is:\[ A = UDV^{T} \] In this decomposition:
- D: \(D \in \mathbb{R}^{m \times n} \rightarrow\) whose diagonal contains singular values \(\sigma_i\) of \(A\). The singular values quantify the importance of each singular vector in capturing the underlying structure and variations in the data. Larger singular values indicate more significant contributions to the data’s variation, while smaller singular values represent less significant contributions.
- U: \(U \in \mathbb{R}^{m \times m} \rightarrow \text{Orthogonal Matrix}\)
- V: \(V \in \mathbb{R}^{n \times n} \rightarrow \text{Orthogonal Matrix}\)
Matrix Approximation using SVD
In this section of the blog post, we aim to:- Get acquainted with numpy’s implementation of SVD
- Develop an appreciation of how SVD allows us to approximate a matrix \(A\) using much smaller representations of matrices
Getting Acquainted with Numpy’s Implementation of SVD
Before we decompose a matrix \(A\) using SVD, we need to construct a sample matrix \(A\):
Now, we can go ahead and visualize this matrix:
A
array([[11.09940146, 11.06768737, 12.07908225, 14.01709143, 14.00268493,
11.08003702, 10.09037225, 11.00246762],
[ 7.04917473, 8.05262552, 9.0596366 , 10.00519575, 10.08950895,
8.07282662, 7.081835 , 8.05002228],
[11.08101894, 10.00959685, 12.021895 , 14.02587191, 14.04681058,
10.04593732, 11.07095098, 10.0178053 ],
[ 9.05314499, 10.01677422, 11.07688139, 12.09281705, 12.06094937,
9.01501835, 8.04896267, 10.0377345 ],
[10.08486014, 11.09110972, 13.03838487, 14.03154959, 14.05683942,
10.0187818 , 10.01258415, 11.06875958]])
With matrices, it is often of utility to visualize them as an image to get an understanding of the underlying pattern:
= "Greys")
plt.imshow(A, cmap = plt.gca().axis("off") a
Now that we have our matrix \(A\), we can use numpy’s implementation of SVD, to get \(U\), \(V\), and sigma which has been discussed above. The only difference is that we get sigma and not \(D\). This means that we get a numpy array of the singular values of \(A\), but we have to construct a diagonal matrix - \(D\) - ourselves, where the diagonals contain the singular values of \(A\). We can wrap up all of these functionalities in a single getValues
function, which will call the diagonalize
function to create the matrix \(D\) out of sigma:
def getValues(matrix):
= np.linalg.svd(matrix)
U, sigma, V = diagonalize(matrix, sigma)
D return U, D, V
def diagonalize(matrix, sigma):
# creating the D matrix in the SVD
# here the matrix of 0s will have the same shape as A
= np.zeros_like(matrix, dtype=float)
D # putting the singular values along the diagonal of D
min(matrix.shape),:min(matrix.shape)] = np.diag(sigma)
D[:return D
= getValues(A) U, D, V
Now, let us try to reconstruct \(A\) using the decomposed matrices as displayed in the formula above and compare it to the original \(A\):
@ D @ V U
array([[11.09940146, 11.06768737, 12.07908225, 14.01709143, 14.00268493,
11.08003702, 10.09037225, 11.00246762],
[ 7.04917473, 8.05262552, 9.0596366 , 10.00519575, 10.08950895,
8.07282662, 7.081835 , 8.05002228],
[11.08101894, 10.00959685, 12.021895 , 14.02587191, 14.04681058,
10.04593732, 11.07095098, 10.0178053 ],
[ 9.05314499, 10.01677422, 11.07688139, 12.09281705, 12.06094937,
9.01501835, 8.04896267, 10.0377345 ],
[10.08486014, 11.09110972, 13.03838487, 14.03154959, 14.05683942,
10.0187818 , 10.01258415, 11.06875958]])
A
array([[11.09940146, 11.06768737, 12.07908225, 14.01709143, 14.00268493,
11.08003702, 10.09037225, 11.00246762],
[ 7.04917473, 8.05262552, 9.0596366 , 10.00519575, 10.08950895,
8.07282662, 7.081835 , 8.05002228],
[11.08101894, 10.00959685, 12.021895 , 14.02587191, 14.04681058,
10.04593732, 11.07095098, 10.0178053 ],
[ 9.05314499, 10.01677422, 11.07688139, 12.09281705, 12.06094937,
9.01501835, 8.04896267, 10.0377345 ],
[10.08486014, 11.09110972, 13.03838487, 14.03154959, 14.05683942,
10.0187818 , 10.01258415, 11.06875958]])
They are the same!
Approximating \(A\) using Smaller Matrices
One of the main reasons why SVD is so useful is because we can get a pretty good approximation of the matrix \(A\) by choosing a smaller subset of these decomposed matrices: \(U\), \(D\), and \(V\). For instance, in our original situation:\[ A = UDV^{T} = \{m \times m\} \times \{m \times n\} \times \{n \times n\} = \{m \times n\} \]
However, we can choose a subset say \(k\) = 3, where we will only:
- Pick the first \(k\) columns of U
- Pick the top \(k\) singular values of D
- Pick the first \(k\) rows of V
Which gives us:
\[
A = UDV^{T} = \{m \times k\} \times \{k \times k\} \times \{k \times n\} = \{m \times n\}
\]
Therefore, we can get a \(m \times n\) approximation of \(A\) using a smaller subset of these decomposed matrices. The way this works is that in SVD, the obtained singular values are present in decreasing order of “importance”. This means that the first singular value is typically the most important in terms of capturing the primary patterns, structures, and variations of the data, and this “importance” keeps on decreasing as we go down. Therefore, depending on our value of \(k\), we can get a pretty good approximation of the original matrix using a relatively smaller subset of the decomposed matrices. To see this in practice, let us create a function that will allow us to visualize our: original matrix \(A\) and the reconstructed matrix \(A\) with a smaller subset of the decomposed matrices, as images. This will allow us to get an appreciation of how similar the reconstructed matrix \(A\) is to the original one, while saving us a lot of space because we used a smaller subset of the matrices:
def compareImages(A, A_):
= plt.subplots(1, 2, figsize = (15, 5))
fig, axarr
0].imshow(A, cmap = "Greys")
axarr[0].axis("off")
axarr[0].set(title = "Original Image")
axarr[
1].imshow(A_, cmap = "Greys")
axarr[1].axis("off")
axarr[1].set(title = "Reconstructed Image")
axarr[
#creating subsets of the decomposed matrices
= 1
k = U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
A_
# visualizing the original and reconstructed matrix A (images)
compareImages(A, A_)
- \(A_\text{original} = \{5 \times 8\} = 40 \text{ units}\), to
- \(A_\text{reconstructed} = U_{\text{k}}D_{\text{k}}V_{\text{k}}^{T} = \{5 \times 1\} + \{1 \times 1\} + \{1 \times 8\} = 14 \text{ units}\)
Image Compression using Matrix Approximation
Now that we have understood how using SVD, we can get a fairly good approximation of the matrix \(A\) using a much smaller subset of the decomposed matrices, we can go ahead and apply to this to image compression. In simple words, what we are trying to do is:- Represent an image as a matrix,
- Decompose the matrix using SVD,
- Use a smaller subset of these decomposed matrices to reconstruct the original image (matrix)

Now, let us get started:
import PIL
import urllib
#function to read an image and save it as a numpy array
def read_image(url):
return np.array(PIL.Image.open(urllib.request.urlopen(url)))
= "https://i.pinimg.com/originals/1f/f9/68/1ff9682f61e99f217bb67a61f02ecb56.jpg"
url
= read_image(url)
img
np.shape(img)
(1200, 1920, 3)
Therefore, we can see that our image is stored as a numpy object of shape: \(1200 \times 1920 \times 3\), this means that there are \(1200\) rows, \(1920\) columns, and the \(3\) is representative of the RGB channels - since at each pixel there is a numerical value for all red, green, and blue. To simplify our task of image compression using SVD, we want to deal with a 2-Dimensional matrix. To do this, we can convert this image to greyscale in which case it will have dimensions: \(1200 \times 1920\) - \(1200\) rows, \(1920\) columns, and each value will range from \(0\) to \(255\). Where, \(0\) = Black, and \(255\) = White. Converting our image to greyscale:
def to_greyscale(image):
return 1 - np.dot(image[...,:3], [0.2989, 0.5870, 0.1140])
= to_greyscale(img)
grey_img
np.shape(grey_img)
(1200, 1920)
As mentioned, the image is now a \(1200 \times 1920\) matrix. Now, let us visualize the original and greyscale image side-by-side:
= plt.subplots(1, 2, figsize = (15, 5))
fig, axarr
0].imshow(img)
axarr[0].axis("off")
axarr[0].set(title = "Original Image")
axarr[
1].imshow(grey_img, cmap = "Greys")
axarr[1].axis("off")
axarr[1].set(title = "Greyscale Image") axarr[
[Text(0.5, 1.0, 'Greyscale Image')]
Reconstructing the Image from its Singular Value Decomposition
Now, we will write the function svd_reconstruct
which will reconstruct an image using \(k\) singular values:
def svd_reconstruct(image, k):
= getValues(image)
U, D, V = U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
grey_img_recon return grey_img_recon
Now, we can go ahead and visualize how our reconstructed image would look with, let’s say, \(k = 50\) singular values:
50)) compareImages(grey_img, svd_reconstruct(grey_img,
Experimenting with Different Values of \(k\)
In this, part of the code, we first want to modify our svd_reconstruct
to incorporate the calculation of how much space we are saving for differing values of \(k\). The formula for this would be:
\[
\text{Percent Storage} = \frac{\text{Storage Needed for Reconstructed}}{\text{Storage Needed for Original}} = \frac{\{(m \times k) + (k \times k) + (k \times n)\} \times 8 \text{ bits}}{m \times n \times 8 \text{ bits}}
\]
\[
= (\frac{mk + kk + kn}{mn}) \times 100
\]
We know, that the matrix of our original image contains \(m \times n\) values - each of which is 8 bytes, if we perform SVD and choose \(k\) singular values, then we have \(mk + kk + kn\) values - each of which is 8 bytes! Thus, the above formula represents the way in which we can achieve percent storage.
Note: Percent Storage is a measure of what percentage of the original size is the reconstructed image using! So, % Storage = \(30\)% means that the reconstructed image’s storage size is \(30\)% of that of the original!
def svd_reconstruct(image, k):
= getValues(image)
U, D, V = U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
grey_img_recon
#code for calculating percent storage
= (image.shape[0] * k) + (k * k) + (k * image.shape[1])
reconstructedStorage = image.shape[0] * image.shape[1]
originalStorage = np.round((reconstructedStorage/originalStorage) * 100, 2)
percentStorage
#returning reconstructed image, and % storage
return grey_img_recon, percentStorage
Now, we want to write our svd_experiment
function which will help us look at the effects that varying size of \(k\) has on: the quality of the image, and the % storage:
def svd_experiment(image):
= 5
rows = 4
cols = plt.subplots(rows, cols, figsize = (25, 10))
fig, axarr = [13 * i for i in range(1, 21)]
k = 0
index for i in range(rows):
for j in range(cols):
#using the value of k at a particular index and then incrementing
= svd_reconstruct(image, k[index])
reconstructedImg, perStorage +=1
index#showing the reconstructed image
= "Greys")
axarr[i][j].imshow(reconstructedImg, cmap "off")
axarr[i][j].axis(set(title = f"k = {k[index-1]} Singular Values, % Storage = {perStorage}%")
axarr[i][j].
fig.tight_layout()
svd_experiment(grey_img)
Therefore, we can see that as the number of \(k\) decrease, the quality of the picture becomes blurrier, however the % storage falls. For example in our situation for \(k = 13\), the reconstructed image’s size is \(1.77\)% of the original image. However, the quality of the image is really low.
To get an appreciation for SVD, we should realize that for around \(k = 104\), the reconstructed image becomes pretty similar to the original, but the reconstructed image only takes 14.55% the storage that the original one took!
Optional Extras
In this section of the blog post, we want to modify oursvd_reconstruct
function to allow the user to:
-
Specify a Desired Compression Factor and Select the Number of Components \(k\) based on this Selection:
\[ \text{Compression Factor (CF)} = \frac{\text{Size of Compressed Bits}}{\text{Size of Original Bits}} \] \[ \text{CF} =\frac{mk + k^{2} + kn}{mn} = \frac{k^2 + k(m+n)}{mn} \] \[ k^2 + k(m+n) = (mn) \times \text{ CF} \] \[ k^2 + k(m+n) - [(mn) \times \text{ CF}] = 0 \] Therefore, we can solve this quadratic equation and get the value for \(k\), for the user provided compression factor!
-
Specify a Desired Threshold epsilon for the Singular Values: selecting all singular values \(i\), such that:
\[ \sigma_{i} \geq \epsilon_{\text{threshold}} \]
Note: The following implementation assumes that the user will either only specify the (1) compression factor, or only the (2) epsilon factor, or only the explicit value of (3) k - Singular Values
def svd_reconstruct(image, k, cf = 0, epsilon = 0):
= 0
grey_img_recon
# checking for when compression factor is specified
if(cf!=0):
# specifying the coefficients of the quadratic equation
= 1
a = image.shape[0] + image.shape[1]
b = (-1) * cf * image.shape[0] * image.shape[1]
c # only looking at the positive value - therefore ignoring the negative
= (-b + np.sqrt((b*b)-(4*a*c)))/(2*a)
k # rounding the value of k - since we can only have whole values - and converting to integer
= np.round(k)
k = k.astype(int)
k # reconstructing our grey image, based on this value of k
= getValues(image)
U, D, V = U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
grey_img_recon
# checking for when epsilon value is specified
elif(epsilon!=0):
= np.linalg.svd(image)
U, sigma, V # finding k - the number of singular values which are
# above epsilon - the specified threshold
= (sigma[sigma>epsilon]).size
k = diagonalize(image, sigma)
D # reconstructing our grey image, based on this value of k
= U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
grey_img_recon
# checking for default case when only
elif(epsilon==0 and cf==0):
= getValues(image)
U, D, V = U[:,:k]
U_ = D[:k, :k]
D_ = V[:k, :]
V_ = U_ @ D_ @ V_
grey_img_recon
#code for calculating percent storage
= (image.shape[0] * k) + (k * k) + (k * image.shape[1])
reconstructedStorage = image.shape[0] * image.shape[1]
originalStorage = np.round((reconstructedStorage/originalStorage) * 100, 2)
percentStorage
#returning reconstructed image, and % storage
return grey_img_recon, percentStorage
Experiments with Different Values of Compression Factor
Now, we can go ahead get a reconstructed image with a compression factor of say \(0.05\)! This means that the reconstructed image, will be \(0.05 * 100 = 5\)% of the storage of the actual image:
= svd_reconstruct(grey_img, 0, 0.05)
reconstructedImage, perStorage compareImages(grey_img, reconstructedImage)
Now, we can go ahead get a reconstructed image with a compression factor of say \(0.3\)! This means that the reconstructed image, will be \(0.3 * 100 = 30\)% of the storage of the actual image:
= svd_reconstruct(grey_img, 0, 0.3)
reconstructedImage, perStorage compareImages(grey_img, reconstructedImage)
Therefore, we can see that as the compression factor goes up, the reconstructed image becomes more like the original image!
Experiments with Different Values of Epsilon - \(\epsilon\)
Finally, we can experiment providing an \(\epsilon\) - epsilon value, which provides a threshold and only the singular values larger than this specified value will be selected. Let us say, we choose \(\epsilon = 13000\) as our threshold. Therefore, only the singular values which are larger than \(13000\) will be used in the image reconstruction:
= svd_reconstruct(grey_img, 0, 0, 13000)
reconstructedImage, perStorage compareImages(grey_img, reconstructedImage)
Let us say, we choose \(\epsilon = 130\) as our threshold. Therefore, only the singular values which are larger than \(130\) will be used in the image reconstruction:
= svd_reconstruct(grey_img, 0, 0, 130)
reconstructedImage, perStorage compareImages(grey_img, reconstructedImage)
Therefore, we can see that as our threshold for the singular values goes down, more and more singular values are used in the image reconstruction, and therefore the reconstructed image becomes more like the original!