Browse Source

HASY: Add k-NN (k=3, k=5)

Martin Thoma 8 years ago
parent
commit
4dcb875bf6
1 changed files with 10 additions and 5 deletions
  1. 10 5
      publications/hasy/content.tex

+ 10 - 5
publications/hasy/content.tex

@@ -201,8 +201,8 @@ of any classifier being evaluated on \dbName{} as follows:
 \subsection{Model Baselines}
 Eight standard algorithms were evaluated by their accuracy on the raw image
 data. The neural networks were implemented with
-Tensorflow~\cite{tensorflow2015-whitepaper}. All other algorithms are
-implemented in sklearn~\cite{scikit-learn}. \Cref{table:classifier-results}
+Tensorflow~0.12.1~\cite{tensorflow2015-whitepaper}. All other algorithms are
+implemented in sklearn~0.18.1~\cite{scikit-learn}. \Cref{table:classifier-results}
 shows the results of the models being trained and tested on MNIST and also for
 \dbNameVersion{}:
 \begin{table}[h]
@@ -215,6 +215,8 @@ shows the results of the models being trained and tested on MNIST and also for
     Random Forest & \SI{96.41}{\percent} & \SI{62.4}{\percent} & \SI{62.1}{\percent} -- \SI{62.8}{\percent}\\% & \SI{19.0}{\second}\\
     MLP (1 Layer) & \SI{89.09}{\percent} & \SI{62.2}{\percent} & \SI{61.7}{\percent} -- \SI{62.9}{\percent}\\% &  \SI{7.8}{\second}\\
     LDA           & \SI{86.42}{\percent} & \SI{46.8}{\percent} & \SI{46.3}{\percent} -- \SI{47.7}{\percent}\\% &   \SI{0.2}{\second}\\
+    $k$-NN ($k=3$)& \SI{92.84}{\percent} & \SI{28.4}{\percent} & \SI{27.4}{\percent} -- \SI{29.1}{\percent}\\% & \SI{196.2}{\second}\\
+    $k$-NN ($k=5$)& \SI{92.88}{\percent} & \SI{27.4}{\percent} & \SI{26.9}{\percent} -- \SI{28.3}{\percent}\\% & \SI{196.2}{\second}\\
     QDA           & \SI{55.61}{\percent} & \SI{25.4}{\percent} & \SI{24.9}{\percent} -- \SI{26.2}{\percent}\\% & \SI{94.7}{\second}\\
     Decision Tree & \SI{65.40}{\percent} & \SI{11.0}{\percent} & \SI{10.4}{\percent} -- \SI{11.6}{\percent}\\% &   \SI{0.0}{\second}\\
     Naive Bayes   & \SI{56.15}{\percent} &  \SI{8.3}{\percent} & \SI{7.9}{\percent} -- \hphantom{0}\SI{8.7}{\percent}\\% & \SI{24.7}{\second}\\
@@ -225,9 +227,12 @@ shows the results of the models being trained and tested on MNIST and also for
              % The test time is the time needed for all test samples in average.
              The number of
              test samples differs between the folds, but is $\num{16827} \pm
-             166$. The decision tree
-             was trained with a maximum depth of 5. The exact structure
-             of the CNNs is explained in~\cref{subsec:CNNs-Classification}.}
+             166$. The decision tree was trained with a maximum depth of~5. The
+             exact structure of the CNNs is explained
+             in~\cref{subsec:CNNs-Classification}. For $k$ nearest neighbor,
+             the amount of samples per class had to be reduced to 50 for HASY
+             due to the extraordinary amount of testing time this algorithm
+             needs.}
     \label{table:classifier-results}
 \end{table}