Przeglądaj źródła

documents/write-math-ba-paper: Fixed some spelling mistakes

Martin Thoma 10 lat temu
rodzic
commit
c9def13de2

+ 4 - 1
documents/write-math-ba-paper/README.md

@@ -1,3 +1,6 @@
+[Download compiled PDF](https://github.com/MartinThoma/LaTeX-examples/blob/master/documents/write-math-ba-paper/write-math-ba-paper.pdf)
+
 ## Spell checking
 ## Spell checking
 * Spell checking `aspell --lang=en --mode=tex check write-math-ba-paper.tex`
 * Spell checking `aspell --lang=en --mode=tex check write-math-ba-paper.tex`
-* Spell checking with `http://www.reverso.net/spell-checker`
+* Spell checking with `http://www.reverso.net/spell-checker`
+* https://github.com/devd/Academic-Writing-Check

BIN
documents/write-math-ba-paper/write-math-ba-paper.pdf


+ 27 - 26
documents/write-math-ba-paper/write-math-ba-paper.tex

@@ -75,22 +75,21 @@ Daniel Kirsch describes in~\cite{Kirsch} a system called Detexify which uses
 time warping to classify on-line handwritten symbols and claims to achieve a
 time warping to classify on-line handwritten symbols and claims to achieve a
 TOP-3 error of less than $\SI{10}{\percent}$ for a set of $\num{100}$~symbols.
 TOP-3 error of less than $\SI{10}{\percent}$ for a set of $\num{100}$~symbols.
 He also published his data on \url{https://github.com/kirel/detexify-data},
 He also published his data on \url{https://github.com/kirel/detexify-data},
-which was collected by a crowd-sourcing approach via
+which was collected by a crowdsourcing approach via
 \url{http://detexify.kirelabs.org}. Those recordings as well as some recordings
 \url{http://detexify.kirelabs.org}. Those recordings as well as some recordings
 which were collected by a similar approach via \url{http://write-math.com} were
 which were collected by a similar approach via \url{http://write-math.com} were
 used to train and evaluated different classifiers. A complete description of
 used to train and evaluated different classifiers. A complete description of
 all involved software, data and experiments is given in~\cite{Thoma:2014}.
 all involved software, data and experiments is given in~\cite{Thoma:2014}.
 
 
 \section{Steps in Handwriting Recognition}
 \section{Steps in Handwriting Recognition}
-The following steps are used in all classifiers which are described in the
-following:
+The following steps are used in many classifiers:
 
 
 \begin{enumerate}
 \begin{enumerate}
     \item \textbf{Preprocessing}: Recorded data is never perfect. Devices have
     \item \textbf{Preprocessing}: Recorded data is never perfect. Devices have
-          errors and people make mistakes while using devices. To tackle these
-          problems there are preprocessing algorithms to clean the data. The
-          preprocessing algorithms can also remove unnecessary variations of
-          the data that do not help in the classification process, but hide
+          errors and people make mistakes while using the devices. To tackle
+          these problems there are preprocessing algorithms to clean the data.
+          The preprocessing algorithms can also remove unnecessary variations
+          of the data that do not help in the classification process, but hide
           what is important. Having slightly different sizes of the same symbol
           what is important. Having slightly different sizes of the same symbol
           is an example of such a variation. Four preprocessing algorithms that
           is an example of such a variation. Four preprocessing algorithms that
           clean or normalize recordings are explained in
           clean or normalize recordings are explained in
@@ -117,15 +116,16 @@ following:
           improve the performance of learning algorithms.
           improve the performance of learning algorithms.
 \end{enumerate}
 \end{enumerate}
 
 
-After these steps, we are faced with a classification learning task which consists of
-two parts:
+After these steps, we are faced with a classification learning task which
+consists of two parts:
 \begin{enumerate}
 \begin{enumerate}
     \item \textbf{Learning} parameters for a given classifier. This process is
     \item \textbf{Learning} parameters for a given classifier. This process is
           also called \textit{training}.
           also called \textit{training}.
     \item \textbf{Classifying} new recordings, sometimes called
     \item \textbf{Classifying} new recordings, sometimes called
           \textit{evaluation}. This should not be confused with the evaluation
           \textit{evaluation}. This should not be confused with the evaluation
           of the classification performance which is done for multiple
           of the classification performance which is done for multiple
-          topologies, preprocessing queues, and features in \Cref{ch:Evaluation}.
+          topologies, preprocessing queues, and features in
+          \Cref{ch:Evaluation}.
 \end{enumerate}
 \end{enumerate}
 
 
 The classification learning task can be solved with \glspl{MLP} if the number
 The classification learning task can be solved with \glspl{MLP} if the number
@@ -141,7 +141,7 @@ and feature extraction easier, more effective or faster. It does so by resolving
 errors in the input data, reducing duplicate information and removing irrelevant
 errors in the input data, reducing duplicate information and removing irrelevant
 information.
 information.
 
 
-Preprocessing algorithms fall in two groups: Normalization and noise
+Preprocessing algorithms fall into two groups: Normalization and noise
 reduction algorithms.
 reduction algorithms.
 
 
 A very important normalization algorithm in single-symbol recognition is
 A very important normalization algorithm in single-symbol recognition is
@@ -157,12 +157,12 @@ Another normalization preprocessing algorithm is resampling. As the data points
 on the pen trajectory are generated asynchronously and with different
 on the pen trajectory are generated asynchronously and with different
 time-resolutions depending on the used hardware and software, it is desirable
 time-resolutions depending on the used hardware and software, it is desirable
 to resample the recordings to have points spread equally in time for every
 to resample the recordings to have points spread equally in time for every
-recording. This was done with linear interpolation of the $(x,t)$ and $(y,t)$
+recording. This was done by linear interpolation of the $(x,t)$ and $(y,t)$
 sequences and getting a fixed number of equally spaced points per stroke.
 sequences and getting a fixed number of equally spaced points per stroke.
 
 
 \textit{Connect strokes} is a noise reduction algorithm. It happens sometimes
 \textit{Connect strokes} is a noise reduction algorithm. It happens sometimes
 that the hardware detects that the user lifted the pen where the user certainly
 that the hardware detects that the user lifted the pen where the user certainly
-didn't do so. This can be detected by measuring the euclidean distance between
+didn't do so. This can be detected by measuring the Euclidean distance between
 the end of one stroke and the beginning of the next stroke. If this distance is
 the end of one stroke and the beginning of the next stroke. If this distance is
 below a threshold, then the strokes are connected.
 below a threshold, then the strokes are connected.
 
 
@@ -207,19 +207,20 @@ activation functions can be varied. The learning algorithm is parameterized by
 the learning rate $\eta \in (0, \infty)$, the momentum $\alpha \in [0, \infty)$
 the learning rate $\eta \in (0, \infty)$, the momentum $\alpha \in [0, \infty)$
 and the number of epochs.
 and the number of epochs.
 
 
-The topology of \glspl{MLP} will be denoted in the following by separating
-the number of neurons per layer with colons. For example, the notation $160{:}500{:}500{:}500{:}369$
-means that the input layer gets 160~features, there are three hidden layers
-with 500~neurons per layer and one output layer with 369~neurons.
-
-\glspl{MLP} training can be executed in
-various different ways, for example with \gls{SLP}.
-In case of a \gls{MLP} with the topology $160{:}500{:}500{:}500{:}369$,
-\gls{SLP} works as follows: At first a \gls{MLP} with one hidden layer ($160{:}500{:}369$)
-is trained. Then the output layer is discarded, a new hidden layer and a new
-output layer is added and it is trained again, resulting in a $160{:}500{:}500{:}369$
-\gls{MLP}. The output layer is discarded again, a new hidden layer is added and
-a new output layer is added and the training is executed again.
+The topology of \glspl{MLP} will be denoted in the following by separating the
+number of neurons per layer with colons. For example, the notation
+$160{:}500{:}500{:}500{:}369$ means that the input layer gets 160~features,
+there are three hidden layers with 500~neurons per layer and one output layer
+with 369~neurons.
+
+\glspl{MLP} training can be executed in various different ways, for example
+with \gls{SLP}. In case of a \gls{MLP} with the topology
+$160{:}500{:}500{:}500{:}369$, \gls{SLP} works as follows: At first a \gls{MLP}
+with one hidden layer ($160{:}500{:}369$) is trained. Then the output layer is
+discarded, a new hidden layer and a new output layer is added and it is trained
+again, resulting in a $160{:}500{:}500{:}369$ \gls{MLP}. The output layer is
+discarded again, a new hidden layer is added and a new output layer is added
+and the training is executed again.
 
 
 Denoising auto-encoders are another way of pretraining. An
 Denoising auto-encoders are another way of pretraining. An
 \textit{auto-encoder} is a neural network that is trained to restore its input.
 \textit{auto-encoder} is a neural network that is trained to restore its input.