소스 검색

HASY: Add k-NN (k=3, k=5)

Martin Thoma 8 년 전
부모
커밋
4dcb875bf6
1개의 변경된 파일10개의 추가작업 그리고 5개의 파일을 삭제
  1. 10 5
      publications/hasy/content.tex

+ 10 - 5
publications/hasy/content.tex

@@ -201,8 +201,8 @@ of any classifier being evaluated on \dbName{} as follows:
 \subsection{Model Baselines}
 \subsection{Model Baselines}
 Eight standard algorithms were evaluated by their accuracy on the raw image
 Eight standard algorithms were evaluated by their accuracy on the raw image
 data. The neural networks were implemented with
 data. The neural networks were implemented with
-Tensorflow~\cite{tensorflow2015-whitepaper}. All other algorithms are
-implemented in sklearn~\cite{scikit-learn}. \Cref{table:classifier-results}
+Tensorflow~0.12.1~\cite{tensorflow2015-whitepaper}. All other algorithms are
+implemented in sklearn~0.18.1~\cite{scikit-learn}. \Cref{table:classifier-results}
 shows the results of the models being trained and tested on MNIST and also for
 shows the results of the models being trained and tested on MNIST and also for
 \dbNameVersion{}:
 \dbNameVersion{}:
 \begin{table}[h]
 \begin{table}[h]
@@ -215,6 +215,8 @@ shows the results of the models being trained and tested on MNIST and also for
     Random Forest & \SI{96.41}{\percent} & \SI{62.4}{\percent} & \SI{62.1}{\percent} -- \SI{62.8}{\percent}\\% & \SI{19.0}{\second}\\
     Random Forest & \SI{96.41}{\percent} & \SI{62.4}{\percent} & \SI{62.1}{\percent} -- \SI{62.8}{\percent}\\% & \SI{19.0}{\second}\\
     MLP (1 Layer) & \SI{89.09}{\percent} & \SI{62.2}{\percent} & \SI{61.7}{\percent} -- \SI{62.9}{\percent}\\% &  \SI{7.8}{\second}\\
     MLP (1 Layer) & \SI{89.09}{\percent} & \SI{62.2}{\percent} & \SI{61.7}{\percent} -- \SI{62.9}{\percent}\\% &  \SI{7.8}{\second}\\
     LDA           & \SI{86.42}{\percent} & \SI{46.8}{\percent} & \SI{46.3}{\percent} -- \SI{47.7}{\percent}\\% &   \SI{0.2}{\second}\\
     LDA           & \SI{86.42}{\percent} & \SI{46.8}{\percent} & \SI{46.3}{\percent} -- \SI{47.7}{\percent}\\% &   \SI{0.2}{\second}\\
+    $k$-NN ($k=3$)& \SI{92.84}{\percent} & \SI{28.4}{\percent} & \SI{27.4}{\percent} -- \SI{29.1}{\percent}\\% & \SI{196.2}{\second}\\
+    $k$-NN ($k=5$)& \SI{92.88}{\percent} & \SI{27.4}{\percent} & \SI{26.9}{\percent} -- \SI{28.3}{\percent}\\% & \SI{196.2}{\second}\\
     QDA           & \SI{55.61}{\percent} & \SI{25.4}{\percent} & \SI{24.9}{\percent} -- \SI{26.2}{\percent}\\% & \SI{94.7}{\second}\\
     QDA           & \SI{55.61}{\percent} & \SI{25.4}{\percent} & \SI{24.9}{\percent} -- \SI{26.2}{\percent}\\% & \SI{94.7}{\second}\\
     Decision Tree & \SI{65.40}{\percent} & \SI{11.0}{\percent} & \SI{10.4}{\percent} -- \SI{11.6}{\percent}\\% &   \SI{0.0}{\second}\\
     Decision Tree & \SI{65.40}{\percent} & \SI{11.0}{\percent} & \SI{10.4}{\percent} -- \SI{11.6}{\percent}\\% &   \SI{0.0}{\second}\\
     Naive Bayes   & \SI{56.15}{\percent} &  \SI{8.3}{\percent} & \SI{7.9}{\percent} -- \hphantom{0}\SI{8.7}{\percent}\\% & \SI{24.7}{\second}\\
     Naive Bayes   & \SI{56.15}{\percent} &  \SI{8.3}{\percent} & \SI{7.9}{\percent} -- \hphantom{0}\SI{8.7}{\percent}\\% & \SI{24.7}{\second}\\
@@ -225,9 +227,12 @@ shows the results of the models being trained and tested on MNIST and also for
              % The test time is the time needed for all test samples in average.
              % The test time is the time needed for all test samples in average.
              The number of
              The number of
              test samples differs between the folds, but is $\num{16827} \pm
              test samples differs between the folds, but is $\num{16827} \pm
-             166$. The decision tree
-             was trained with a maximum depth of 5. The exact structure
-             of the CNNs is explained in~\cref{subsec:CNNs-Classification}.}
+             166$. The decision tree was trained with a maximum depth of~5. The
+             exact structure of the CNNs is explained
+             in~\cref{subsec:CNNs-Classification}. For $k$ nearest neighbor,
+             the amount of samples per class had to be reduced to 50 for HASY
+             due to the extraordinary amount of testing time this algorithm
+             needs.}
     \label{table:classifier-results}
     \label{table:classifier-results}
 \end{table}
 \end{table}