K-Means Clustering for Image Quantization
In this post we explore image quantization by applying the k-means algorithm to a sample flower image. We will explore how different values of k affects the quality of the resulting image and computational complexity of the algorithm. Let’s dive right in!
Implemented Modules
All modules within this section have been personally implemented by the author using NumPy
and Matplotlib
libraries. You can find the source code on GitHub.
Documentation
src.image
Image
: Represents an image with attributes such as RGB matrix, RGB vector, shape, and title.load_image(image_path: str) -> Image
: Loads an image from the given file path, converts it to RGB format, and returns an Image object.rgb_vector_to_image(rgb_vector: np.ndarray, image_shape: tuple) -> Image
: Converts an RGB vector to an Image object with the specified shape.display_images(*images: Image, figsize=(10, 5))
: Displays images using matplotlib.plot_rgb_vector(rgb_vector: np.ndarray, figsize=(10, 5), title=None, alpha=0.6, s=4)
: Plots a 3D scatter plot of RGB vectors.
1
from src import image
src.model
KMeans
: Implements the -means clustering algorithm with parameters for the number of clusters (n_clusters), tolerance (tol), maximum iterations (max_iter), and random state (random_state).KMeans.fit(X: np.ndarray) -> KMeans
: Fits the KMeans model to the input data X using the specified parameters.KMeans.labels_
: An array of shape (n_samples,) where each element represents the cluster index assigned to the corresponding sample.KMeans.cluster_centers_
: An array of shape (n_clusters, n_features) representing the final cluster centers after fitting the model. Each row corresponds to the centroid of a cluster in the feature space.plot_k_metrics(k_values: list[int], clusters_cost_list: list[float], execution_time_list: list[float], figsize: tuple = (10, 4))
: Plots the clusters cost and execution time of -means for different values of .
1
from src import model
Try it yourself
To try out these modules, simply follow these steps:
Download the source code here
Go to the /image_quantization folder
Install required packages:
pip install -r requirements.txt
Create a notebook or Python file in the /image_quantization folder
You’re all set! You can import the modules as
from src import image, model
Image Quantization
One of the applications of
Let’s explore this concept by applying the
Loading the Image
1
2
3
4
5
import numpy as np
img = image.load_image("img/flower.jpg")
img.title = "Original Image"
image.display_images(img)
The image is loaded as matrix
1
img.rgb_matrix.shape
(417, 640, 3)
To fit the
1
img.rgb_vector.shape
(266880, 3)
We can visualize the feature vectors in the 3-dimensional space:
1
image.plot_rgb_vector(img.rgb_vector, title="Feature vectors $$ X $$")
We will experiment with three different values for
-means
First we train the model with
1
2
k8_means = model.KMeans(n_clusters=8, random_state=44)
%time k8_means.fit(img.rgb_vector)
CPU times: total: 10.9 s
Wall time: 11.1 s<src.model.KMeans at 0x1237c937210>
We can inspect executed iterations:
1
k8_means.iterations
72
And we can access the labels and centroids (or centers) found:
1
k8_means.labels_
array([6, 6, 6, …, 6, 6, 6], dtype=int64)
1
k8_means.cluster_centers_
array([
[42.39915168, 67.65893264, 84.11651684],
[ 87.46747399, 127.68052914, 119.33007048],
[191.76564391, 194.46557984, 189.38218125],
[114.46671081, 156.33419082, 139.91508329],
[ 60.50504249, 99.93731908, 100.45057895],
[152.42347775, 174.89832302, 153.38496535],
[ 24.28310161, 36.94729513, 65.34303884],
[223.21554659, 224.39480287, 222.18355735]
])
As we can observe, each centroid is a 3-dimensional floating-point array. Consequently, we need to cast this array to an integer array.
1
2
k8_colors = np.uint8(k8_means.cluster_centers_)
k8_colors
array([
[ 42, 67, 84],
[ 87, 127, 119],
[191, 194, 189],
[114, 156, 139],
[ 60, 99, 100],
[152, 174, 153],
[ 24, 36, 65],
[223, 224, 222]
], dtype=uint8)
Let’s plot the centers in the 3-dimensional space:
1
image.plot_rgb_vector(k8_colors, alpha=1, s=100, title="8-means centroids")
To obtain the quantized image, we simply replace each pixel’s feature vector with the corresponding centroid of the cluster to which it belongs:
1
k8_rgb_vector = k8_colors[k8_means.labels_]
Finally, we can display the quantized image:
1
2
3
k8_image = image.rgb_vector_to_image(k8_rgb_vector, img.shape)
k8_image.title = "8-mean Image"
image.display_images(k8_image)
To accelerate this process we create a function quantize_image()
, that given the image to quantize and the value of
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
def quantize_image(img: image.Image, k: int, random_state: int):
kmeans = model.KMeans(n_clusters=k, random_state=random_state)
# Training
print("[Training]")
%time kmeans.fit(img.rgb_vector)
print(f"Iterations: {kmeans.iterations}")
# Results
print("\n[Results]")
centroids = np.uint8(kmeans.cluster_centers_)
print(f"{k}-means centroids:")
image.plot_rgb_vector(centroids, alpha=1, s=100, title=f"{k}-means centroids")
quantized_img = image.rgb_vector_to_image(centroids[kmeans.labels_], img.shape)
quantized_img.title = f"{k}-mean image"
print("Quantized image:")
image.display_images(quantized_img)
return quantized_img
-means
We can now train the
1
k16_image = quantize_image(img, k=16, random_state=44)
[Training]
CPU times: total: 25.8 s
Wall time: 26.2 s
Iterations: 100[Results]
16-means centroids:
Quantized image:
-means
Finally, we quantize the image with
1
k32_image = quantize_image(img, k=32, random_state=44)
[Training]
CPU times: total: 36.3 s
Wall time: 36.5 s
Iterations: 76[Results]
32-means centroids:
Quantized image:
Comparison
By comparing the results obtained with
1
image.display_images(k8_image, k16_image, k32_image, figsize=(20, 15))
It’s worth noting that, in addition to the visual impact, the computational complexity of the algorithm increases with a larger value of
Quality measure and execution time
We can quantify the image quality with a clusters cost function, which measures the squared Euclidean distance between each feature vector
1
2
3
4
5
6
7
def clusters_cost(X, centroids, labels):
cost = 0
for j in range(len(centroids)):
cluster = X[np.where(labels == j)]
distance = np.linalg.norm(cluster - centroids[j])**2
cost += np.sum(distance)
return cost
If we analyze our dataset
1
len(np.unique(img.rgb_vector, axis=0))
25786
A quantization with
The model.plot_k_metrics()
function plots the change in clusters cost and execution time for different values of record_k_metrics()
function. We utilize the highly optimized sklearn.cluster.KMeans
class to significantly improve the execution time. Please note that our implementation, model.KMeans
, has an identical interface to sklearn.cluster.KMeans
, except for the parameter n_init
, which is not implemented in model.KMeans
.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
from sklearn.cluster import KMeans
from time import time
def record_k_metrics(
img: image.Image,
k_values: list[int],
random_state: int,
):
# Train the model for each k
execution_time_list = []
clusters_cost_list = []
for k in k_values:
kmeans = KMeans(n_clusters=k, random_state=random_state, n_init=5)
# Execution time
start_time = time()
kmeans.fit(img.rgb_vector)
end_time = time()
execution_time_list.append(end_time - start_time)
# Clusters cost
centroids = np.uint8(kmeans.cluster_centers_)
cost = clusters_cost(img.rgb_vector, centroids, kmeans.labels_)
clusters_cost_list.append(cost)
return clusters_cost_list, execution_time_list
We can visualize the clusters cost and execution time for
1
2
3
k_values = list(range(2, 65, 2))
clusters_cost_list, execution_time_list = record_k_metrics(img, k_values, random_state=44)
model.plot_k_metrics(k_values, clusters_cost_list, execution_time_list)
As
Conclusion
In this exploration of image quantization using the
Experimenting with
Further, we explored the influence of plot_k_metrics()
function facilitated a comprehensive view of these metrics, shedding light on the intricate relationship between the number of clusters and computational performance.
This project, though a humble beginning, marks a significant step in understanding the nuances of image quantization. Your feedback and thoughts are invaluable as we continue to refine and expand our exploration of machine learning concepts.
Thank you for joining me on this journey, and I look forward to hearing your insights and suggestions.