Skip to content
Snippets Groups Projects
Commit ebe1428f authored by Mirko Birbaumer's avatar Mirko Birbaumer
Browse files

Solutions to Exercises Block 1 executed

parent b8c75e14
No related branches found
No related tags found
No related merge requests found
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Exercise 1 - K-Nearest Neighbor Classifier for MNIST # Exercise 1 - K-Nearest Neighbor Classifier for MNIST
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
In this exercise, we'll apply KNN Classifiers to the MNIST dataset. The aim of the exercise is to get acquainted with the MNIST dataset. In this exercise, we'll apply KNN Classifiers to the MNIST dataset. The aim of the exercise is to get acquainted with the MNIST dataset.
This guide uses [tf.keras](https://www.tensorflow.org/guide/keras), a high-level API to build and train models in TensorFlow. This guide uses [tf.keras](https://www.tensorflow.org/guide/keras), a high-level API to build and train models in TensorFlow.
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
## Install and import dependencies ## Install and import dependencies
We'll need [TensorFlow Datasets](https://www.tensorflow.org/datasets/), an API that simplifies downloading and accessing datasets, and provides several sample datasets to work with. We're also using a few helper libraries. We'll need [TensorFlow Datasets](https://www.tensorflow.org/datasets/), an API that simplifies downloading and accessing datasets, and provides several sample datasets to work with. We're also using a few helper libraries.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
!pip install -U tensorflow_datasets !pip install -U tensorflow_datasets
``` ```
%% Output %% Output
Collecting tensorflow_datasets Requirement already up-to-date: tensorflow_datasets in /opt/conda/lib/python3.7/site-packages (4.6.0)
Downloading tensorflow_datasets-4.5.2-py3-none-any.whl (4.2 MB) Requirement already satisfied, skipping upgrade: absl-py in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.2.0)
 |████████████████████████████████| 4.2 MB 5.1 MB/s eta 0:00:01
[?25hRequirement already satisfied, skipping upgrade: six in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.14.0)
Collecting dill
Downloading dill-0.3.4-py2.py3-none-any.whl (86 kB)
 |████████████████████████████████| 86 kB 1.8 MB/s eta 0:00:01
[?25hCollecting importlib-resources; python_version < "3.9"
Downloading importlib_resources-5.4.0-py3-none-any.whl (28 kB)
Requirement already satisfied, skipping upgrade: typing-extensions; python_version < "3.8" in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (3.7.4.3) Requirement already satisfied, skipping upgrade: typing-extensions; python_version < "3.8" in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (3.7.4.3)
Collecting promise Requirement already satisfied, skipping upgrade: dill in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (0.3.5.1)
Downloading promise-2.3.tar.gz (19 kB) Requirement already satisfied, skipping upgrade: termcolor in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (2.0.1)
Collecting tensorflow-metadata Requirement already satisfied, skipping upgrade: toml in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (0.10.2)
Downloading tensorflow_metadata-1.6.0-py3-none-any.whl (48 kB)
 |████████████████████████████████| 48 kB 6.0 MB/s eta 0:00:01
[?25hRequirement already satisfied, skipping upgrade: tqdm in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (4.45.0)
Requirement already satisfied, skipping upgrade: requests>=2.19.0 in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (2.23.0)
Requirement already satisfied, skipping upgrade: termcolor in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.1.0)
Requirement already satisfied, skipping upgrade: protobuf>=3.12.2 in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (3.19.4)
Requirement already satisfied, skipping upgrade: numpy in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.19.1) Requirement already satisfied, skipping upgrade: numpy in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.19.1)
Requirement already satisfied, skipping upgrade: absl-py in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.0.0) Requirement already satisfied, skipping upgrade: protobuf>=3.12.2 in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (3.20.2)
Requirement already satisfied, skipping upgrade: importlib-resources; python_version < "3.9" in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (5.9.0)
Requirement already satisfied, skipping upgrade: tensorflow-metadata in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.10.0)
Requirement already satisfied, skipping upgrade: etils[epath] in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (0.8.0)
Requirement already satisfied, skipping upgrade: tqdm in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (4.45.0)
Requirement already satisfied, skipping upgrade: requests>=2.19.0 in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (2.23.0)
Requirement already satisfied, skipping upgrade: six in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (1.14.0)
Requirement already satisfied, skipping upgrade: promise in /opt/conda/lib/python3.7/site-packages (from tensorflow_datasets) (2.3)
Requirement already satisfied, skipping upgrade: zipp>=3.1.0; python_version < "3.10" in /opt/conda/lib/python3.7/site-packages (from importlib-resources; python_version < "3.9"->tensorflow_datasets) (3.1.0) Requirement already satisfied, skipping upgrade: zipp>=3.1.0; python_version < "3.10" in /opt/conda/lib/python3.7/site-packages (from importlib-resources; python_version < "3.9"->tensorflow_datasets) (3.1.0)
Collecting googleapis-common-protos<2,>=1.52.0 Requirement already satisfied, skipping upgrade: googleapis-common-protos<2,>=1.52.0 in /opt/conda/lib/python3.7/site-packages (from tensorflow-metadata->tensorflow_datasets) (1.56.4)
Downloading googleapis_common_protos-1.55.0-py2.py3-none-any.whl (212 kB) Requirement already satisfied, skipping upgrade: idna<3,>=2.5 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (2.9)
 |████████████████████████████████| 212 kB 40.3 MB/s eta 0:00:01 Requirement already satisfied, skipping upgrade: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (1.25.11)
[?25hRequirement already satisfied, skipping upgrade: idna<3,>=2.5 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (2.9)
Requirement already satisfied, skipping upgrade: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (1.25.9)
Requirement already satisfied, skipping upgrade: chardet<4,>=3.0.2 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (3.0.4)
Requirement already satisfied, skipping upgrade: certifi>=2017.4.17 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (2020.6.20) Requirement already satisfied, skipping upgrade: certifi>=2017.4.17 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (2020.6.20)
Building wheels for collected packages: promise Requirement already satisfied, skipping upgrade: chardet<4,>=3.0.2 in /opt/conda/lib/python3.7/site-packages (from requests>=2.19.0->tensorflow_datasets) (3.0.4)
Building wheel for promise (setup.py) ... [?25ldone WARNING: You are using pip version 20.2.4; however, version 22.2.2 is available.
[?25h Created wheel for promise: filename=promise-2.3-py3-none-any.whl size=21495 sha256=9675d5917247a843670272d14b85c27811d276d4f06faeb1bc0a545e304f40d5
Stored in directory: /home/jovyan/.cache/pip/wheels/29/93/c6/762e359f8cb6a5b69c72235d798804cae523bbe41c2aa8333d
Successfully built promise
Installing collected packages: dill, importlib-resources, promise, googleapis-common-protos, tensorflow-metadata, tensorflow-datasets
Successfully installed dill-0.3.4 googleapis-common-protos-1.55.0 importlib-resources-5.4.0 promise-2.3 tensorflow-datasets-4.5.2 tensorflow-metadata-1.6.0
WARNING: You are using pip version 20.2.4; however, version 22.0.3 is available.
You should consider upgrading via the '/opt/conda/bin/python3 -m pip install --upgrade pip' command. You should consider upgrading via the '/opt/conda/bin/python3 -m pip install --upgrade pip' command.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
# Import TensorFlow # Import TensorFlow
# FOR COLAB USERS: # FOR COLAB USERS:
# If you run this notebook in Colab, then execute the following line (uncomment it) # If you run this notebook in Colab, then execute the following line (uncomment it)
# %tensorflow_version 2.x # %tensorflow_version 2.x
# If you run this noteook in your tensorflow 2.x environment, then # If you run this noteook in your tensorflow 2.x environment, then
# verify you have version > 2.0 # verify you have version > 2.0
import tensorflow as tf import tensorflow as tf
print(tf.__version__) print(tf.__version__)
# Now you should get version 2.x # Now you should get version 2.x
# If you still get version 1.x, then execute (uncomment) the following lines and run the cell # If you still get version 1.x, then execute (uncomment) the following lines and run the cell
#!pip uninstall tensorflow #!pip uninstall tensorflow
#!pip install --upgrade pip #!pip install --upgrade pip
#!pip install --upgrade tensorflow #!pip install --upgrade tensorflow
#!python3 -c "import tensorflow as tf;print(tf.reduce_sum(tf.random.normal([1000, 1000])))" #!python3 -c "import tensorflow as tf;print(tf.reduce_sum(tf.random.normal([1000, 1000])))"
#try: #try:
# import tensorflow as tf # import tensorflow as tf
#except Exception: #except Exception:
# pass # pass
#print(tf.__version__) #print(tf.__version__)
``` ```
%% Output %% Output
2.7.1 2.7.1
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
from __future__ import absolute_import, division, print_function, unicode_literals from __future__ import absolute_import, division, print_function, unicode_literals
# Import TensorFlow Datasets # Import TensorFlow Datasets
import tensorflow as tf import tensorflow as tf
import tensorflow_datasets as tfds import tensorflow_datasets as tfds
tfds.disable_progress_bar() tfds.disable_progress_bar()
# Helper libraries # Helper libraries
import math import math
import numpy as np import numpy as np
import matplotlib.pyplot as plt import matplotlib.pyplot as plt
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
import logging import logging
logger = tf.get_logger() logger = tf.get_logger()
logger.setLevel(logging.ERROR) logger.setLevel(logging.ERROR)
!python -V !python -V
``` ```
%% Output %% Output
Python 3.7.6 Python 3.7.6
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
## Import the MNIST dataset ## Import the MNIST dataset
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
This guide uses the [MNIST](http://yann.lecun.com/exdb/mnist/) dataset—often used as the "Hello, World" of machine learning programs for computer vision. The MNIST dataset contains images of handwritten digits (0, 1, 2, etc) This guide uses the [MNIST](http://yann.lecun.com/exdb/mnist/) dataset—often used as the "Hello, World" of machine learning programs for computer vision. The MNIST dataset contains images of handwritten digits (0, 1, 2, etc)
We will use 60,000 images to train the network and 10,000 images to evaluate how accurately the network learned to classify images. You can access the MNIST directly from TensorFlow, using the [Datasets](https://www.tensorflow.org/datasets) API: We will use 60,000 images to train the network and 10,000 images to evaluate how accurately the network learned to classify images. You can access the MNIST directly from TensorFlow, using the [Datasets](https://www.tensorflow.org/datasets) API:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
dataset, metadata = tfds.load('mnist', as_supervised=True, with_info=True) dataset, metadata = tfds.load('mnist', as_supervised=True, with_info=True)
train_dataset, test_dataset = dataset['train'], dataset['test'] train_dataset, test_dataset = dataset['train'], dataset['test']
``` ```
%% Output %% Output
Downloading and preparing dataset 11.06 MiB (download: 11.06 MiB, generated: 21.00 MiB, total: 32.06 MiB) to /home/jovyan/tensorflow_datasets/mnist/3.0.1... Downloading and preparing dataset 11.06 MiB (download: 11.06 MiB, generated: 21.00 MiB, total: 32.06 MiB) to ~/tensorflow_datasets/mnist/3.0.1...
Dataset mnist downloaded and prepared to /home/jovyan/tensorflow_datasets/mnist/3.0.1. Subsequent calls will reuse this data. Dataset mnist downloaded and prepared to ~/tensorflow_datasets/mnist/3.0.1. Subsequent calls will reuse this data.
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Loading the dataset returns metadata as well as a *training dataset* and *test dataset*. Loading the dataset returns metadata as well as a *training dataset* and *test dataset*.
* The model is trained using `train_dataset`. * The model is trained using `train_dataset`.
* The model is tested against `test_dataset`. * The model is tested against `test_dataset`.
The images are 28 $\times$ 28 arrays, with pixel values in the range `[0, 255]`. The *labels* are an array of integers, in the range `[0, 9]`. These correspond to the handwritten numbers. The images are 28 $\times$ 28 arrays, with pixel values in the range `[0, 255]`. The *labels* are an array of integers, in the range `[0, 9]`. These correspond to the handwritten numbers.
Each image is mapped to a single label. Since the *class names* are not included with the dataset, store them here to use later when plotting the images: Each image is mapped to a single label. Since the *class names* are not included with the dataset, store them here to use later when plotting the images:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
class_names = ['Zero', 'One', 'Two', 'Three', 'Four', 'Five', class_names = ['Zero', 'One', 'Two', 'Three', 'Four', 'Five',
'Six', 'Seven', 'Eight', 'Nine'] 'Six', 'Seven', 'Eight', 'Nine']
``` ```
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
### Explore the data ### Explore the data
Let's explore the format of the dataset before training the model. The following shows there are 60,000 images in the training set, and 10000 images in the test set: Let's explore the format of the dataset before training the model. The following shows there are 60,000 images in the training set, and 10000 images in the test set:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_train_examples = metadata.splits['train'].num_examples num_train_examples = metadata.splits['train'].num_examples
num_test_examples = metadata.splits['test'].num_examples num_test_examples = metadata.splits['test'].num_examples
print("Number of training examples: {}".format(num_train_examples)) print("Number of training examples: {}".format(num_train_examples))
print("Number of test examples: {}".format(num_test_examples)) print("Number of test examples: {}".format(num_test_examples))
``` ```
%% Output %% Output
Number of training examples: 60000 Number of training examples: 60000
Number of test examples: 10000 Number of test examples: 10000
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Let's plot an image to see what it looks like. Let's plot an image to see what it looks like.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
# Take a single image, and remove the color dimension by reshaping # Take a single image, and remove the color dimension by reshaping
for image, label in test_dataset.take(1): for image, label in test_dataset.take(1):
break break
image = image.numpy().reshape((28,28)) image = image.numpy().reshape((28,28))
# Plot the image - voila an example of a handwritten digit # Plot the image - voila an example of a handwritten digit
plt.figure() plt.figure()
plt.imshow(image, cmap=plt.cm.binary) plt.imshow(image, cmap=plt.cm.binary)
plt.colorbar() plt.colorbar()
plt.grid(False) plt.grid(False)
plt.show() plt.show()
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Display the first 25 images from the *test set* and display the class name below each image. Verify that the data is in the correct format and we're ready to build and train the network. Display the first 25 images from the *test set* and display the class name below each image. Verify that the data is in the correct format and we're ready to build and train the network.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
plt.figure(figsize=(10,10)) plt.figure(figsize=(10,10))
i = 0 i = 0
for (image, label) in test_dataset.take(25): for (image, label) in test_dataset.take(25):
image = image.numpy().reshape((28,28)) image = image.numpy().reshape((28,28))
plt.subplot(5,5,i+1) plt.subplot(5,5,i+1)
plt.xticks([]) plt.xticks([])
plt.yticks([]) plt.yticks([])
plt.grid(False) plt.grid(False)
plt.imshow(image, cmap=plt.cm.binary) plt.imshow(image, cmap=plt.cm.binary)
plt.xlabel(class_names[label]) plt.xlabel(class_names[label])
i += 1 i += 1
plt.show() plt.show()
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
## Import the Fashion MNIST dataset ## Import the Fashion MNIST dataset
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
If numbers are not your thing then use the [Fashion MNIST](https://github.com/zalandoresearch/fashion-mnist) dataset, which contains 70,000 grayscale images in 10 categories. The images show individual articles of clothing at low resolution (28 $\times$ 28 pixels), as seen here: If numbers are not your thing then use the [Fashion MNIST](https://github.com/zalandoresearch/fashion-mnist) dataset, which contains 70,000 grayscale images in 10 categories. The images show individual articles of clothing at low resolution (28 $\times$ 28 pixels), as seen here:
<table> <table>
<tr><td> <tr><td>
<img src="https://tensorflow.org/images/fashion-mnist-sprite.png" <img src="https://tensorflow.org/images/fashion-mnist-sprite.png"
alt="Fashion MNIST sprite" width="600"> alt="Fashion MNIST sprite" width="600">
</td></tr> </td></tr>
<tr><td align="center"> <tr><td align="center">
<b>Figure 1.</b> <a href="https://github.com/zalandoresearch/fashion-mnist">Fashion-MNIST samples</a> (by Zalando, MIT License).<br/>&nbsp; <b>Figure 1.</b> <a href="https://github.com/zalandoresearch/fashion-mnist">Fashion-MNIST samples</a> (by Zalando, MIT License).<br/>&nbsp;
</td></tr> </td></tr>
</table> </table>
You may use Fashion MNIST for variety, and because it's a slightly more challenging problem than regular MNIST. Both datasets are relatively small and are used to verify that an algorithm works as expected. They're good starting points to test and debug code. You may use Fashion MNIST for variety, and because it's a slightly more challenging problem than regular MNIST. Both datasets are relatively small and are used to verify that an algorithm works as expected. They're good starting points to test and debug code.
We will use 60,000 images to train the network and 10,000 images to evaluate how accurately the network learned to classify images. You can access the Fashion MNIST directly from TensorFlow, using the [Datasets](https://www.tensorflow.org/datasets) API: We will use 60,000 images to train the network and 10,000 images to evaluate how accurately the network learned to classify images. You can access the Fashion MNIST directly from TensorFlow, using the [Datasets](https://www.tensorflow.org/datasets) API:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
dataset, metadata = tfds.load('fashion_mnist', as_supervised=True, with_info=True) dataset, metadata = tfds.load('fashion_mnist', as_supervised=True, with_info=True)
train_dataset, test_dataset = dataset['train'], dataset['test'] train_dataset, test_dataset = dataset['train'], dataset['test']
``` ```
%% Output %% Output
Downloading and preparing dataset 29.45 MiB (download: 29.45 MiB, generated: 36.42 MiB, total: 65.87 MiB) to /home/jovyan/tensorflow_datasets/fashion_mnist/3.0.1... Downloading and preparing dataset 29.45 MiB (download: 29.45 MiB, generated: 36.42 MiB, total: 65.87 MiB) to ~/tensorflow_datasets/fashion_mnist/3.0.1...
Dataset fashion_mnist downloaded and prepared to /home/jovyan/tensorflow_datasets/fashion_mnist/3.0.1. Subsequent calls will reuse this data. Dataset fashion_mnist downloaded and prepared to ~/tensorflow_datasets/fashion_mnist/3.0.1. Subsequent calls will reuse this data.
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Loading the dataset returns metadata as well as a *training dataset* and *test dataset*. Loading the dataset returns metadata as well as a *training dataset* and *test dataset*.
* The model is trained using `train_dataset`. * The model is trained using `train_dataset`.
* The model is tested against `test_dataset`. * The model is tested against `test_dataset`.
The images are 28 $\times$ 28 arrays, with pixel values in the range `[0, 255]`. The *labels* are an array of integers, in the range `[0, 9]`. These correspond to the *class* of clothing the image represents: The images are 28 $\times$ 28 arrays, with pixel values in the range `[0, 255]`. The *labels* are an array of integers, in the range `[0, 9]`. These correspond to the *class* of clothing the image represents:
<table> <table>
<tr> <tr>
<th>Label</th> <th>Label</th>
<th>Class</th> <th>Class</th>
</tr> </tr>
<tr> <tr>
<td>0</td> <td>0</td>
<td>T-shirt/top</td> <td>T-shirt/top</td>
</tr> </tr>
<tr> <tr>
<td>1</td> <td>1</td>
<td>Trouser</td> <td>Trouser</td>
</tr> </tr>
<tr> <tr>
<td>2</td> <td>2</td>
<td>Pullover</td> <td>Pullover</td>
</tr> </tr>
<tr> <tr>
<td>3</td> <td>3</td>
<td>Dress</td> <td>Dress</td>
</tr> </tr>
<tr> <tr>
<td>4</td> <td>4</td>
<td>Coat</td> <td>Coat</td>
</tr> </tr>
<tr> <tr>
<td>5</td> <td>5</td>
<td>Sandal</td> <td>Sandal</td>
</tr> </tr>
<tr> <tr>
<td>6</td> <td>6</td>
<td>Shirt</td> <td>Shirt</td>
</tr> </tr>
<tr> <tr>
<td>7</td> <td>7</td>
<td>Sneaker</td> <td>Sneaker</td>
</tr> </tr>
<tr> <tr>
<td>8</td> <td>8</td>
<td>Bag</td> <td>Bag</td>
</tr> </tr>
<tr> <tr>
<td>9</td> <td>9</td>
<td>Ankle boot</td> <td>Ankle boot</td>
</tr> </tr>
</table> </table>
Each image is mapped to a single label. Since the *class names* are not included with the dataset, store them here to use later when plotting the images: Each image is mapped to a single label. Since the *class names* are not included with the dataset, store them here to use later when plotting the images:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
class_names = ['T-shirt/top', 'Trouser', 'Pullover', 'Dress', 'Coat', class_names = ['T-shirt/top', 'Trouser', 'Pullover', 'Dress', 'Coat',
'Sandal', 'Shirt', 'Sneaker', 'Bag', 'Ankle boot'] 'Sandal', 'Shirt', 'Sneaker', 'Bag', 'Ankle boot']
``` ```
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
### Explore the data ### Explore the data
Let's explore the format of the dataset before training the model. The following shows there are 60,000 images in the training set, and 10000 images in the test set: Let's explore the format of the dataset before training the model. The following shows there are 60,000 images in the training set, and 10000 images in the test set:
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_train_examples = metadata.splits['train'].num_examples num_train_examples = metadata.splits['train'].num_examples
num_test_examples = metadata.splits['test'].num_examples num_test_examples = metadata.splits['test'].num_examples
print("Number of training examples: {}".format(num_train_examples)) print("Number of training examples: {}".format(num_train_examples))
print("Number of test examples: {}".format(num_test_examples)) print("Number of test examples: {}".format(num_test_examples))
``` ```
%% Output %% Output
Number of training examples: 60000 Number of training examples: 60000
Number of test examples: 10000 Number of test examples: 10000
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Let's plot an image to see what it looks like. Let's plot an image to see what it looks like.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
# Take a single image, and remove the color dimension by reshaping # Take a single image, and remove the color dimension by reshaping
for image, label in test_dataset.take(1): for image, label in test_dataset.take(1):
break break
image = image.numpy().reshape((28,28)) image = image.numpy().reshape((28,28))
# Plot the image - voila a piece of fashion clothing # Plot the image - voila a piece of fashion clothing
plt.figure() plt.figure()
plt.imshow(image, cmap=plt.cm.binary) plt.imshow(image, cmap=plt.cm.binary)
plt.colorbar() plt.colorbar()
plt.grid(False) plt.grid(False)
plt.show() plt.show()
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Display the first 25 images from the *test set* and display the class name below each image. Verify that the data is in the correct format and we're ready to build and train the network. Display the first 25 images from the *test set* and display the class name below each image. Verify that the data is in the correct format and we're ready to build and train the network.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
plt.figure(figsize=(10,10)) plt.figure(figsize=(10,10))
i = 0 i = 0
for (image, label) in test_dataset.take(25): for (image, label) in test_dataset.take(25):
image = image.numpy().reshape((28,28)) image = image.numpy().reshape((28,28))
plt.subplot(5,5,i+1) plt.subplot(5,5,i+1)
plt.xticks([]) plt.xticks([])
plt.yticks([]) plt.yticks([])
plt.grid(False) plt.grid(False)
plt.imshow(image, cmap=plt.cm.binary) plt.imshow(image, cmap=plt.cm.binary)
plt.xlabel(class_names[label]) plt.xlabel(class_names[label])
i += 1 i += 1
plt.show() plt.show()
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
Decide whether you want to work with the traditional or Fashin MNIST dataset, then extract 5000 training examples and Decide whether you want to work with the traditional or Fashin MNIST dataset, then extract 5000 training examples and
500 test examples. 500 test examples.
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
i=0 i=0
for (image, label) in train_dataset.take(5000): for (image, label) in train_dataset.take(5000):
if i==0: if i==0:
X_train = image.numpy().reshape((1,28*28)) X_train = image.numpy().reshape((1,28*28))
y_train = np.array([label]) y_train = np.array([label])
else: else:
X_train = np.concatenate([X_train, image.numpy().reshape((1,28*28))], axis=0) X_train = np.concatenate([X_train, image.numpy().reshape((1,28*28))], axis=0)
y_train = np.concatenate([y_train, np.array([label])], axis=0) y_train = np.concatenate([y_train, np.array([label])], axis=0)
i+=1 i+=1
print("Shape of image training data : ", X_train.shape) print("Shape of image training data : ", X_train.shape)
print("Shape of training data labels : ", y_train.shape) print("Shape of training data labels : ", y_train.shape)
``` ```
%% Output %% Output
Shape of image training data : (5000, 784) Shape of image training data : (5000, 784)
Shape of training data labels : (5000,) Shape of training data labels : (5000,)
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
j=0 j=0
for (image, label) in test_dataset.take(500): for (image, label) in test_dataset.take(500):
if j==0: if j==0:
X_test = image.numpy().reshape((1,28*28)) X_test = image.numpy().reshape((1,28*28))
y_test = np.array([label]) y_test = np.array([label])
else: else:
X_test = np.concatenate([X_test, image.numpy().reshape((1,28*28))], axis=0) X_test = np.concatenate([X_test, image.numpy().reshape((1,28*28))], axis=0)
y_test = np.concatenate([y_test, np.array([label])], axis=0) y_test = np.concatenate([y_test, np.array([label])], axis=0)
j+=1 j+=1
print("Shape of image test data : ", X_test.shape) print("Shape of image test data : ", X_test.shape)
print("Shape of test data labels : ", y_test.shape) print("Shape of test data labels : ", y_test.shape)
``` ```
%% Output %% Output
Shape of image test data : (500, 784) Shape of image test data : (500, 784)
Shape of test data labels : (500,) Shape of test data labels : (500,)
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Exercises # Exercises
1. Apply Nearest Neighbour with L1 distance to this subset of the dataset and determine the accuracy on the 1. Apply Nearest Neighbour with L1 distance to this subset of the dataset and determine the accuracy on the
test dataset and plot the confusion matrix. test dataset and plot the confusion matrix.
2. Apply K-Nearest Neighbour with $k=5$ and L2 distance to this subset of the dataset and determine the accuracy on the test dataset and plot the confusion matrix. 2. Apply K-Nearest Neighbour with $k=5$ and L2 distance to this subset of the dataset and determine the accuracy on the test dataset and plot the confusion matrix.
3. Determine by means of 5-fold cross-validation the best value of $k$ in the set $\{1,4,5,10,12,18,20\}$. 3. Determine by means of 5-fold cross-validation the best value of $k$ in the set $\{1,4,5,10,12,18,20\}$.
4. Scale the pixel values to the interval $[0, 1]$ and compute the test accuracy for the best value of k determined in exercise 3. 4. Scale the pixel values to the interval $[0, 1]$ and compute the test accuracy for the best value of k determined in exercise 3.
5. Implement the Cosine distance measure in the k-nearest neighbour classifier. The cosine distance between two vectors $a$ and $b$ can be computed by 5. Implement the Cosine distance measure in the k-nearest neighbour classifier. The cosine distance between two vectors $a$ and $b$ can be computed by
```python ```python
from numpy.linalg import norm from numpy.linalg import norm
from numpy import dot from numpy import dot
dists[a,b] = 1 - dot(a, b)/(norm(a)*norm(b)) dists[a,b] = 1 - dot(a, b)/(norm(a)*norm(b))
``` ```
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Solution 1 # Solution 1
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
class KNearestNeighbor(): class KNearestNeighbor():
""" a kNN classifier with L2 distance """ """ a kNN classifier with L2 distance """
def __init__(self): def __init__(self):
pass pass
def train(self, X, y): def train(self, X, y):
""" """
Train the classifier. For k-nearest neighbors this is just Train the classifier. For k-nearest neighbors this is just
memorizing the training data. memorizing the training data.
Inputs: Inputs:
- X: A numpy array of shape (num_train, D) containing the training data - X: A numpy array of shape (num_train, D) containing the training data
consisting of num_train samples each of dimension D. consisting of num_train samples each of dimension D.
- y: A numpy array of shape (N,) containing the training labels, where - y: A numpy array of shape (N,) containing the training labels, where
y[i] is the label for X[i]. y[i] is the label for X[i].
""" """
self.X_train = X.astype('float') self.X_train = X.astype('float')
self.y_train = y self.y_train = y
def predict(self, X, k=1, num_loops=0): def predict(self, X, k=1, num_loops=0):
""" """
Predict labels for test data using this classifier. Predict labels for test data using this classifier.
Inputs: Inputs:
- X: A numpy array of shape (num_test, D) containing test data consisting - X: A numpy array of shape (num_test, D) containing test data consisting
of num_test samples each of dimension D. of num_test samples each of dimension D.
- k: The number of nearest neighbors that vote for the predicted labels. - k: The number of nearest neighbors that vote for the predicted labels.
- num_loops: Determines which implementation to use to compute distances - num_loops: Determines which implementation to use to compute distances
between training points and testing points. between training points and testing points.
Returns: Returns:
- y: A numpy array of shape (num_test,) containing predicted labels for the - y: A numpy array of shape (num_test,) containing predicted labels for the
test data, where y[i] is the predicted label for the test point X[i]. test data, where y[i] is the predicted label for the test point X[i].
""" """
if num_loops == 0: if num_loops == 0:
dists = self.compute_distances_no_loops(X) dists = self.compute_distances_no_loops(X)
elif num_loops == 1: elif num_loops == 1:
dists = self.compute_distances_one_loop(X) dists = self.compute_distances_one_loop(X)
elif num_loops == 2: elif num_loops == 2:
dists = self.compute_distances_two_loops(X) dists = self.compute_distances_two_loops(X)
else: else:
raise ValueError('Invalid value %d for num_loops' % num_loops) raise ValueError('Invalid value %d for num_loops' % num_loops)
return self.predict_labels(dists, k=k) return self.predict_labels(dists, k=k)
def compute_distances_two_loops(self, X): def compute_distances_two_loops(self, X):
""" """
Compute the distance between each test point in X and each Compute the distance between each test point in X and each
training point in self.X_train using a nested loop over both training point in self.X_train using a nested loop over both
the training data and the test data. the training data and the test data.
Inputs: Inputs:
- X: A numpy array of shape (num_test, D) containing test data. - X: A numpy array of shape (num_test, D) containing test data.
Returns: Returns:
- dists: A numpy array of shape (num_test, num_train) where - dists: A numpy array of shape (num_test, num_train) where
dists[i, j] is the Euclidean distance between the ith test dists[i, j] is the Euclidean distance between the ith test
point and the jth training point. point and the jth training point.
""" """
num_test = X.shape[0] num_test = X.shape[0]
num_train = self.X_train.shape[0] num_train = self.X_train.shape[0]
dists = np.zeros((num_test, num_train)) dists = np.zeros((num_test, num_train))
X = X.astype('float') X = X.astype('float')
for i in range(num_test): for i in range(num_test):
for j in range(num_train): for j in range(num_train):
dists[i, j] = np.sqrt(np.sum(np.square(self.X_train[j,:] - X[i,:]))) dists[i, j] = np.sqrt(np.sum(np.square(self.X_train[j,:] - X[i,:])))
return dists return dists
def compute_distances_one_loop(self, X): def compute_distances_one_loop(self, X):
""" """
Compute the distance between each test point in X and each training point Compute the distance between each test point in X and each training point
in self.X_train using a single loop over the test data. in self.X_train using a single loop over the test data.
Input / Output: Same as compute_distances_two_loops Input / Output: Same as compute_distances_two_loops
""" """
num_test = X.shape[0] num_test = X.shape[0]
num_train = self.X_train.shape[0] num_train = self.X_train.shape[0]
dists = np.zeros((num_test, num_train)) dists = np.zeros((num_test, num_train))
X = X.astype('float') X = X.astype('float')
for i in range(num_test): for i in range(num_test):
dists[i, :] = np.sqrt(np.sum(np.square(self.X_train - X[i,:]), axis = 1)) dists[i, :] = np.sqrt(np.sum(np.square(self.X_train - X[i,:]), axis = 1))
return dists return dists
def compute_distances_no_loops(self, X): def compute_distances_no_loops(self, X):
""" """
Compute the distance between each test point in X and each training point Compute the distance between each test point in X and each training point
in self.X_train using no explicit loops. in self.X_train using no explicit loops.
Input / Output: Same as compute_distances_two_loops Input / Output: Same as compute_distances_two_loops
""" """
num_test = X.shape[0] num_test = X.shape[0]
num_train = self.X_train.shape[0] num_train = self.X_train.shape[0]
dists = np.zeros((num_test, num_train)) dists = np.zeros((num_test, num_train))
X=X.astype('float') X=X.astype('float')
# Most "elegant" solution leads however to memory issues # Most "elegant" solution leads however to memory issues
# dists = np.sqrt(np.square((self.X_train[:, np.newaxis, :] - X)).sum(axis=2)).T # dists = np.sqrt(np.square((self.X_train[:, np.newaxis, :] - X)).sum(axis=2)).T
# split (p-q)^2 to p^2 + q^2 - 2pq # split (p-q)^2 to p^2 + q^2 - 2pq
dists = np.sqrt((X**2).sum(axis=1)[:, np.newaxis] + (self.X_train**2).sum(axis=1) - 2 * X.dot(self.X_train.T)) dists = np.sqrt((X**2).sum(axis=1)[:, np.newaxis] + (self.X_train**2).sum(axis=1) - 2 * X.dot(self.X_train.T))
return dists return dists
def predict_labels(self, dists, k=1): def predict_labels(self, dists, k=1):
""" """
Given a matrix of distances between test points and training points, Given a matrix of distances between test points and training points,
predict a label for each test point. predict a label for each test point.
Inputs: Inputs:
- dists: A numpy array of shape (num_test, num_train) where dists[i, j] - dists: A numpy array of shape (num_test, num_train) where dists[i, j]
gives the distance betwen the ith test point and the jth training point. gives the distance betwen the ith test point and the jth training point.
Returns: Returns:
- y: A numpy array of shape (num_test,) containing predicted labels for the - y: A numpy array of shape (num_test,) containing predicted labels for the
test data, where y[i] is the predicted label for the test point X[i]. test data, where y[i] is the predicted label for the test point X[i].
""" """
num_test = dists.shape[0] num_test = dists.shape[0]
y_pred = np.zeros(num_test, dtype='float64') y_pred = np.zeros(num_test, dtype='float64')
for i in range(num_test): for i in range(num_test):
# A list of length k storing the labels of the k nearest neighbors to # A list of length k storing the labels of the k nearest neighbors to
# the ith test point. # the ith test point.
closest_y = [] closest_y = []
# get the k indices with smallest distances # get the k indices with smallest distances
min_indices = np.argsort(dists[i,:])[:k] min_indices = np.argsort(dists[i,:])[:k]
closest_y = np.bincount(self.y_train[min_indices]) closest_y = np.bincount(self.y_train[min_indices])
# predict the label of the nearest example # predict the label of the nearest example
y_pred[i] = np.argmax(closest_y) y_pred[i] = np.argmax(closest_y)
return y_pred return y_pred
class KNearestNeighbor_L1(KNearestNeighbor): class KNearestNeighbor_L1(KNearestNeighbor):
""" a kNN classifier with L1 distance """ """ a kNN classifier with L1 distance """
def __init__(self): def __init__(self):
super().__init__() super().__init__()
def compute_distances_one_loop(self, X): def compute_distances_one_loop(self, X):
""" """
We overwrite the compute_distance_one_loop method of the parent class We overwrite the compute_distance_one_loop method of the parent class
KNearestNeighbor. KNearestNeighbor.
Compute the distance between each test point in X and each training point Compute the distance between each test point in X and each training point
in self.X_train using one loop and the L1 distance measure. in self.X_train using one loop and the L1 distance measure.
Input / Output: Same as compute_distances_two_loops Input / Output: Same as compute_distances_two_loops
""" """
num_test = X.shape[0] num_test = X.shape[0]
num_train = self.X_train.shape[0] num_train = self.X_train.shape[0]
dists = np.zeros((num_test, num_train)) dists = np.zeros((num_test, num_train))
X = X.astype('float') X = X.astype('float')
for i in range(num_test): for i in range(num_test):
dists[i, :] = (np.sum(np.abs(self.X_train - X[i,:]), axis = 1)) dists[i, :] = (np.sum(np.abs(self.X_train - X[i,:]), axis = 1))
return dists return dists
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
classifier = KNearestNeighbor_L1() classifier = KNearestNeighbor_L1()
classifier.train(X_train, y_train) classifier.train(X_train, y_train)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
dists = classifier.compute_distances_no_loops(X_test) dists = classifier.compute_distances_no_loops(X_test)
dists.shape dists.shape
``` ```
%% Output %% Output
(500, 5000) (500, 5000)
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
y_test_pred = classifier.predict_labels(dists, k=1) y_test_pred = classifier.predict_labels(dists, k=1)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_test = X_test.shape[0] num_test = X_test.shape[0]
num_correct = np.sum(y_test_pred == y_test) num_correct = np.sum(y_test_pred == y_test)
accuracy = float(num_correct) / num_test accuracy = float(num_correct) / num_test
print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy)) print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy))
``` ```
%% Output %% Output
Got 399 / 500 correct => accuracy: 0.798000 Got 399 / 500 correct => accuracy: 0.798000
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
# utility function for plotting confusion matrix # utility function for plotting confusion matrix
import matplotlib.pyplot as plt import matplotlib.pyplot as plt
import seaborn as sns import seaborn as sns
from sklearn.metrics import confusion_matrix from sklearn.metrics import confusion_matrix
def plot_confmat(y_true, y_pred): def plot_confmat(y_true, y_pred):
""" """
Plot the confusion matrix and save to user_files dir Plot the confusion matrix and save to user_files dir
""" """
conf_matrix = confusion_matrix(y_true, y_pred) conf_matrix = confusion_matrix(y_true, y_pred)
fig = plt.figure(figsize=(9,9)) fig = plt.figure(figsize=(9,9))
ax = fig.add_subplot(111) ax = fig.add_subplot(111)
sns.heatmap(conf_matrix, sns.heatmap(conf_matrix,
annot=True, annot=True,
fmt='.0f') fmt='.0f')
plt.title('Confusion matrix') plt.title('Confusion matrix')
ax.set_xticklabels(class_names) ax.set_xticklabels(class_names)
ax.set_yticklabels(class_names) ax.set_yticklabels(class_names)
plt.ylabel('True') plt.ylabel('True')
plt.xlabel('Predicted') plt.xlabel('Predicted')
plot_confmat(y_test, y_test_pred) plot_confmat(y_test, y_test_pred)
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Solution 2 # Solution 2
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
classifier = KNearestNeighbor() classifier = KNearestNeighbor()
classifier.train(X_train, y_train) classifier.train(X_train, y_train)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
dists = classifier.compute_distances_no_loops(X_test) dists = classifier.compute_distances_no_loops(X_test)
dists.shape dists.shape
``` ```
%% Output %% Output
(500, 5000) (500, 5000)
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
y_test_pred = classifier.predict_labels(dists, k=5) y_test_pred = classifier.predict_labels(dists, k=5)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_test = X_test.shape[0] num_test = X_test.shape[0]
num_correct = np.sum(y_test_pred == y_test) num_correct = np.sum(y_test_pred == y_test)
accuracy = float(num_correct) / num_test accuracy = float(num_correct) / num_test
print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy)) print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy))
``` ```
%% Output %% Output
Got 400 / 500 correct => accuracy: 0.800000 Got 400 / 500 correct => accuracy: 0.800000
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
plot_confmat(y_test, y_test_pred) plot_confmat(y_test, y_test_pred)
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Solution 3 # Solution 3
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_folds = 5 num_folds = 5
k_choices = [1, 4, 5, 10, 12, 18, 20] k_choices = [1, 4, 5, 10, 12, 18, 20]
X_train_folds = [] X_train_folds = []
y_train_folds = [] y_train_folds = []
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_train = X_train.shape[0] num_train = X_train.shape[0]
fold_size = np.ceil(num_train/num_folds).astype('int') fold_size = np.ceil(num_train/num_folds).astype('int')
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
X_train_folds = np.split(X_train, [(i + 1)*fold_size for i in np.arange(num_folds)]) X_train_folds = np.split(X_train, [(i + 1)*fold_size for i in np.arange(num_folds)])
y_train_folds = np.split(y_train, [(i + 1)*fold_size for i in np.arange(num_folds)]) y_train_folds = np.split(y_train, [(i + 1)*fold_size for i in np.arange(num_folds)])
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
k_to_accuracies = {} k_to_accuracies = {}
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
for k in k_choices: for k in k_choices:
k_to_accuracies[k] = [] k_to_accuracies[k] = []
classifier = KNearestNeighbor() classifier = KNearestNeighbor()
for i in range(num_folds): for i in range(num_folds):
X_cv_training = np.concatenate([x for k, x in enumerate(X_train_folds) if k!=i], axis=0) X_cv_training = np.concatenate([x for k, x in enumerate(X_train_folds) if k!=i], axis=0)
y_cv_training = np.concatenate([x for k, x in enumerate(y_train_folds) if k!=i], axis=0) y_cv_training = np.concatenate([x for k, x in enumerate(y_train_folds) if k!=i], axis=0)
classifier.train(X_cv_training, y_cv_training) classifier.train(X_cv_training, y_cv_training)
dists = classifier.compute_distances_no_loops(X_train_folds[i]) dists = classifier.compute_distances_no_loops(X_train_folds[i])
y_test_pred = classifier.predict_labels(dists, k=k) y_test_pred = classifier.predict_labels(dists, k=k)
k_to_accuracies[k].append(np.mean(y_train_folds[i] == y_test_pred)) k_to_accuracies[k].append(np.mean(y_train_folds[i] == y_test_pred))
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
for k in sorted(k_to_accuracies): for k in sorted(k_to_accuracies):
for accuracy in k_to_accuracies[k]: for accuracy in k_to_accuracies[k]:
print('k = %d, accuracy = %f' % (k, accuracy)) print('k = %d, accuracy = %f' % (k, accuracy))
``` ```
%% Output %% Output
k = 1, accuracy = 0.809000 k = 1, accuracy = 0.809000
k = 1, accuracy = 0.809000 k = 1, accuracy = 0.809000
k = 1, accuracy = 0.812000 k = 1, accuracy = 0.812000
k = 1, accuracy = 0.780000 k = 1, accuracy = 0.780000
k = 1, accuracy = 0.790000 k = 1, accuracy = 0.790000
k = 4, accuracy = 0.811000 k = 4, accuracy = 0.811000
k = 4, accuracy = 0.834000 k = 4, accuracy = 0.834000
k = 4, accuracy = 0.842000 k = 4, accuracy = 0.842000
k = 4, accuracy = 0.806000 k = 4, accuracy = 0.806000
k = 4, accuracy = 0.799000 k = 4, accuracy = 0.799000
k = 5, accuracy = 0.804000 k = 5, accuracy = 0.804000
k = 5, accuracy = 0.828000 k = 5, accuracy = 0.828000
k = 5, accuracy = 0.839000 k = 5, accuracy = 0.839000
k = 5, accuracy = 0.807000 k = 5, accuracy = 0.807000
k = 5, accuracy = 0.805000 k = 5, accuracy = 0.805000
k = 10, accuracy = 0.810000 k = 10, accuracy = 0.810000
k = 10, accuracy = 0.827000 k = 10, accuracy = 0.827000
k = 10, accuracy = 0.832000 k = 10, accuracy = 0.832000
k = 10, accuracy = 0.805000 k = 10, accuracy = 0.805000
k = 10, accuracy = 0.803000 k = 10, accuracy = 0.803000
k = 12, accuracy = 0.808000 k = 12, accuracy = 0.808000
k = 12, accuracy = 0.818000 k = 12, accuracy = 0.818000
k = 12, accuracy = 0.833000 k = 12, accuracy = 0.833000
k = 12, accuracy = 0.802000 k = 12, accuracy = 0.802000
k = 12, accuracy = 0.802000 k = 12, accuracy = 0.802000
k = 18, accuracy = 0.804000 k = 18, accuracy = 0.804000
k = 18, accuracy = 0.810000 k = 18, accuracy = 0.810000
k = 18, accuracy = 0.832000 k = 18, accuracy = 0.832000
k = 18, accuracy = 0.798000 k = 18, accuracy = 0.798000
k = 18, accuracy = 0.791000 k = 18, accuracy = 0.791000
k = 20, accuracy = 0.799000 k = 20, accuracy = 0.799000
k = 20, accuracy = 0.806000 k = 20, accuracy = 0.806000
k = 20, accuracy = 0.822000 k = 20, accuracy = 0.822000
k = 20, accuracy = 0.797000 k = 20, accuracy = 0.797000
k = 20, accuracy = 0.787000 k = 20, accuracy = 0.787000
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
# plot the raw observations # plot the raw observations
for k in k_choices: for k in k_choices:
accuracies = k_to_accuracies[k] accuracies = k_to_accuracies[k]
plt.scatter([k] * len(accuracies), accuracies) plt.scatter([k] * len(accuracies), accuracies)
# plot the trend line with error bars that correspond # plot the trend line with error bars that correspond
# to standard deviation # to standard deviation
accuracies_mean = np.array([np.mean(v) for k,v in accuracies_mean = np.array([np.mean(v) for k,v in
sorted(k_to_accuracies.items())]) sorted(k_to_accuracies.items())])
accuracies_std = np.array([np.std(v) for k,v in sorted(k_to_accuracies.items())]) accuracies_std = np.array([np.std(v) for k,v in sorted(k_to_accuracies.items())])
plt.errorbar(k_choices, accuracies_mean, yerr=accuracies_std) plt.errorbar(k_choices, accuracies_mean, yerr=accuracies_std)
plt.title('Cross-validation on k') plt.title('Cross-validation on k')
plt.xlabel('k') plt.xlabel('k')
plt.ylabel('Cross-validation accuracy') plt.ylabel('Cross-validation accuracy')
plt.show() plt.show()
``` ```
%% Output %% Output
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Solution 4 # Solution 4
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
X_train =X_train/255 X_train =X_train/255
X_test = X_test/255 X_test = X_test/255
classifier = KNearestNeighbor() classifier = KNearestNeighbor()
classifier.train(X_train, y_train) classifier.train(X_train, y_train)
dists = classifier.compute_distances_no_loops(X_test) dists = classifier.compute_distances_no_loops(X_test)
dists.shape dists.shape
y_test_pred = classifier.predict_labels(dists, k=1) y_test_pred = classifier.predict_labels(dists, k=1)
num_test = X_test.shape[0] num_test = X_test.shape[0]
num_correct = np.sum(y_test_pred == y_test) num_correct = np.sum(y_test_pred == y_test)
accuracy = float(num_correct) / num_test accuracy = float(num_correct) / num_test
print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy)) print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy))
``` ```
%% Output %% Output
Got 399 / 500 correct => accuracy: 0.798000 Got 399 / 500 correct => accuracy: 0.798000
%% Cell type:markdown id: tags: %% Cell type:markdown id: tags:
# Solution 5 # Solution 5
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
from numpy.linalg import norm from numpy.linalg import norm
from numpy import dot from numpy import dot
class KNearestNeighbor_cosine(): class KNearestNeighbor_cosine():
""" a kNN classifier with L2 distance """ """ a kNN classifier with L2 distance """
def __init__(self): def __init__(self):
pass pass
def train(self, X, y): def train(self, X, y):
""" """
Train the classifier. For k-nearest neighbors this is just Train the classifier. For k-nearest neighbors this is just
memorizing the training data. memorizing the training data.
Inputs: Inputs:
- X: A numpy array of shape (num_train, D) containing the training data - X: A numpy array of shape (num_train, D) containing the training data
consisting of num_train samples each of dimension D. consisting of num_train samples each of dimension D.
- y: A numpy array of shape (N,) containing the training labels, where - y: A numpy array of shape (N,) containing the training labels, where
y[i] is the label for X[i]. y[i] is the label for X[i].
""" """
self.X_train = X.astype('float') self.X_train = X.astype('float')
self.y_train = y self.y_train = y
def predict(self, X, k=1): def predict(self, X, k=1):
""" """
Predict labels for test data using this classifier. Predict labels for test data using this classifier.
Inputs: Inputs:
- X: A numpy array of shape (num_test, D) containing test data consisting - X: A numpy array of shape (num_test, D) containing test data consisting
of num_test samples each of dimension D. of num_test samples each of dimension D.
- k: The number of nearest neighbors that vote for the predicted labels. - k: The number of nearest neighbors that vote for the predicted labels.
- num_loops: Determines which implementation to use to compute distances - num_loops: Determines which implementation to use to compute distances
between training points and testing points. between training points and testing points.
Returns: Returns:
- y: A numpy array of shape (num_test,) containing predicted labels for the - y: A numpy array of shape (num_test,) containing predicted labels for the
test data, where y[i] is the predicted label for the test point X[i]. test data, where y[i] is the predicted label for the test point X[i].
""" """
dists = self.compute_distances_two_loops(X) dists = self.compute_distances_two_loops(X)
return self.predict_labels(dists, k=k) return self.predict_labels(dists, k=k)
def compute_distances_two_loops(self, X): def compute_distances_two_loops(self, X):
""" """
Compute the distance between each test point in X and each Compute the distance between each test point in X and each
training point in self.X_train using a nested loop over both training point in self.X_train using a nested loop over both
the training data and the test data. the training data and the test data.
Inputs: Inputs:
- X: A numpy array of shape (num_test, D) containing test data. - X: A numpy array of shape (num_test, D) containing test data.
Returns: Returns:
- dists: A numpy array of shape (num_test, num_train) where - dists: A numpy array of shape (num_test, num_train) where
dists[i, j] is the Euclidean distance between the ith test dists[i, j] is the Euclidean distance between the ith test
point and the jth training point. point and the jth training point.
""" """
num_test = X.shape[0] num_test = X.shape[0]
num_train = self.X_train.shape[0] num_train = self.X_train.shape[0]
dists = np.zeros((num_test, num_train)) dists = np.zeros((num_test, num_train))
X = X.astype('float') X = X.astype('float')
for i in range(num_test): for i in range(num_test):
for j in range(num_train): for j in range(num_train):
dists[i, j] = 1 - dot(self.X_train[j,:] , X[i,:])/(norm(self.X_train[j,:])*norm(self.X_train[i,:])) dists[i, j] = 1 - dot(self.X_train[j,:] , X[i,:])/(norm(self.X_train[j,:])*norm(self.X_train[i,:]))
return dists return dists
def predict_labels(self, dists, k=1): def predict_labels(self, dists, k=1):
""" """
Given a matrix of distances between test points and training points, Given a matrix of distances between test points and training points,
predict a label for each test point. predict a label for each test point.
Inputs: Inputs:
- dists: A numpy array of shape (num_test, num_train) where dists[i, j] - dists: A numpy array of shape (num_test, num_train) where dists[i, j]
gives the distance betwen the ith test point and the jth training point. gives the distance betwen the ith test point and the jth training point.
Returns: Returns:
- y: A numpy array of shape (num_test,) containing predicted labels for the - y: A numpy array of shape (num_test,) containing predicted labels for the
test data, where y[i] is the predicted label for the test point X[i]. test data, where y[i] is the predicted label for the test point X[i].
""" """
num_test = dists.shape[0] num_test = dists.shape[0]
y_pred = np.zeros(num_test, dtype='float64') y_pred = np.zeros(num_test, dtype='float64')
for i in range(num_test): for i in range(num_test):
# A list of length k storing the labels of the k nearest neighbors to # A list of length k storing the labels of the k nearest neighbors to
# the ith test point. # the ith test point.
closest_y = [] closest_y = []
# get the k indices with smallest distances # get the k indices with smallest distances
min_indices = np.argsort(dists[i,:])[:k] min_indices = np.argsort(dists[i,:])[:k]
closest_y = np.bincount(self.y_train[min_indices]) closest_y = np.bincount(self.y_train[min_indices])
# predict the label of the nearest example # predict the label of the nearest example
y_pred[i] = np.argmax(closest_y) y_pred[i] = np.argmax(closest_y)
return y_pred return y_pred
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
classifier = KNearestNeighbor_cosine() classifier = KNearestNeighbor_cosine()
classifier.train(X_train, y_train) classifier.train(X_train, y_train)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
dists = classifier.compute_distances_two_loops(X_test) dists = classifier.compute_distances_two_loops(X_test)
dists.shape dists.shape
``` ```
%% Output %% Output
(500, 5000) (500, 5000)
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
y_test_pred = classifier.predict_labels(dists, k=1) y_test_pred = classifier.predict_labels(dists, k=1)
``` ```
%% Cell type:code id: tags: %% Cell type:code id: tags:
``` python ``` python
num_test = X_test.shape[0] num_test = X_test.shape[0]
num_correct = np.sum(y_test_pred == y_test) num_correct = np.sum(y_test_pred == y_test)
accuracy = float(num_correct) / num_test accuracy = float(num_correct) / num_test
print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy)) print('Got %d / %d correct => accuracy: %f' % (num_correct, num_test, accuracy))
``` ```
%% Output %% Output
Got 192 / 500 correct => accuracy: 0.384000 Got 394 / 500 correct => accuracy: 0.788000
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment