| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215 |
- \documentclass[a4paper,9pt]{scrartcl}
- \usepackage{amssymb, amsmath} % needed for math
- \usepackage[utf8]{inputenc} % this is needed for umlauts
- \usepackage[USenglish]{babel} % this is needed for umlauts
- \usepackage[T1]{fontenc} % this is needed for correct output of umlauts in pdf
- \usepackage[margin=2.5cm]{geometry} %layout
- \usepackage{hyperref} % hyperlinks
- \usepackage{color}
- \usepackage{framed}
- \usepackage{enumerate} % for advanced numbering of lists
- \usepackage{csquotes} % for enquote
- \newcommand\titletext{Peer-Review of\\"Deep Neuronal Networks for Semantiv Segmentation in Medical
- Informatics"}
- \title{\titletext}
- \author{Martin Thoma}
- \hypersetup{
- pdfauthor = {Martin Thoma},
- pdfkeywords = {peer review},
- pdftitle = {Lineare Algebra}
- }
- \usepackage{microtype}
- \begin{document}
- \maketitle
- \section{Introduction}
- This is a peer-review of \enquote{Deep Neuronal Networks for Semantiv
- Segmentation in Medical Informatics} by Marvin Teichmann. The reviewed document
- is available under \href{https://github.com/MarvinTeichmann/seminar-pixel-exact-classification.git}{https://github.com/MarvinTeichmann/seminar-pixel-exact-classification.git}, version
- \texttt{b1bdb4802c8e268ebf7ca66adb7f806e29afb413}.
- \section{Summary of the Content}
- The author wants to describe how convolutional networks can be used for
- semantic segmentation tasks in medicine. To do so, he introduces Convolutional
- Neural Networks.
- As the introduction, section~2 (Computer Vision Tasks) and section~5
- (Application in Medical Informatics) are not written yet, it can only be said
- that the plan of writing them is good.
- The author expects the reader to know how neural networks work in general, but
- gives a detailed introduction into CNNs. He continues with explaining fully
- convolutional networks (FCNs). This leads in a natural fashion to the
- application of neural networks for segmentation.
- \section{Overall Feedback}
- Gramatical errors make it sometimes difficult to understand relatively easy
- sentences. Also, the missing parts make it difficult to see if there is a
- consistent overall structure.
- I recommend adding more source to claims made in the paper.
- The overall structure seems to be logical, definitions are given most of the
- time (see the feedback below for some exceptions where it should be added).
- \section{Major Remarks}
- \subsection{Section 3 / 3.1: CNNs}
- \begin{itemize}
- \item What is \enquote{stationarity of statistics}?
- \item What are \enquote{translation invariance functions}?
- \item The term \enquote{Kernel} and \enquote{reception field} were neither
- introduced nor a source was given where the reader could find
- definitions.
- \item What is a \enquote{channel size}? Do you mean the number of channels
- or the channel dimension?
- \item What is $F_{nm}$? A function, but on which domain does it operate and
- to which domain does it map? What does this function mean? Is it
- an activation function?
- \item What does $n << h,w$ mean? $n \ll \min(h, w)$?
- \item It was not explained what \enquote{a sliding window fashion} means.
- \item I miss an~image in section 3.1 (definitions and notation).
- \end{itemize}
- \subsection{Section 3.2: Layer types}
- \begin{itemize}
- \item I've never heard of activation layers. Do you mean fully connected
- layers? If not, then you should probably cite a publication which
- calls it like that.
- \item \enquote{curtained weights} - what is that? (The problem might be
- my lack of knowledge of the English language). However, I think
- you should cite a source here for the claim that this is possible.
- \item \enquote{a variety of tasks including edge and area detection,
- contrast sharpening and image blurring}: I miss a source.
- \item \enquote{big ($k \geq 7$). [KSH12, SZ14, SLJ + 14].} - What exactly
- do you cite here?
- \item An image with a tiny example would make the pooling layer much
- easier to understand. However, you can also cite a source which
- explains this well.
- \item The sentence \enquote{Firstly it naturally reduces the spatial dimension
- enabling the network to learn more compact representation if the data and decreasing the
- amount of parameters in the succeeding layers.} sounds wrong. You forgot something
- At \enquote{if the data}.
- \item The sentence is gramatically wrong and makes it hard to understand
- \enquote{Secondly it introduces robust translation invariant.}.
- \item \enquote{Minor shifts in the input data will not result in the same activation after pooling.}
- Not? I thought that was the advantage of pooling, that you get
- invariant?
- \item \enquote{Recently ReLU Nonlinearities [KSH12](AlexNet, Bolzmann)}:
- It is possible to make that easier to read:
- \enquote{Recently ReLU nonlinearities, as introduced by~[KSH12](AlexNet, Bolzmann)}
- - However, I'm not too sure what you mean with \enquote{Bolzmann}.
- \item It was not explained / defined what ReLU means / is.
- \end{itemize}
- \subsection{Section 4: Neural Networks for Segmentation}
- \begin{itemize}
- \item \enquote{After the overwhelming successes of DCNNs in image classification}: Add source
- \item \enquote{in combination with traditional classifiers} - What are \enquote{traditional} classifiers?
- \item \enquote{Other authors used the idea described in Section 2} - Don't make me jump back. Can you give that idea a short name? Then you can write something like \enquote{the idea of sliding windows}. As you wrote about sliding windows in the rest of the sentence, I guess restrucuting the sentence might help.
- \item \enquote{are currently the state-of-the art in several semantic segmentation benchmarks.} - name at least one.
- \end{itemize}
- \subsection{Section 4.1: Sliding Window efficiency in CNNs}
- \begin{itemize}
- \item \enquote{The input image will be down sampled by a factor of s corresponding to the product of all strides being applied in $C'$.} - I don't think that is obvious. Please explain it or give a source for that claim.
- \item \enquote{shift-and-stitch} - What is that?
- \end{itemize}
- \subsection{Section 4.2: FCNs}
- \begin{itemize}
- \item \enquote{builds up on the ideas presented of Section 4.1} - which ones?
- The \textit{sliding-window-as-a-convoluton} idea and which other idea?
- \item \enquote{they are not trying to avoid downsampling as part of the progress}
- - do you mean process?
- \item Explain what an \enquote{upsampling layer} is.
- \end{itemize}
- \subsection{Section 4.2.1: Deconvolution}
- This section is still to be done.
- \subsection{Section 4.2.2: Skip-Architecture}
- An image would help, although I guess it is already easy to understand.
- \subsection{4.2.3 Transfer Learning}
- \begin{itemize}
- \item What is transfer lerning?
- \item What is VGG16 (cite paper) - same for AlexNet and GoogLeNet, if it
- wasn't done already. People who don't know what a CNN is will also
- not know what AlexNet / GoogLeNet is.
- \end{itemize}
- \subsection{4.3 Extensions of FCN}
- \begin{itemize}
- \item \enquote{Several extensions of FCN have been proposed} - give sources
- \item \enquote{of strong labeled data} what is \textbf{strong} labeled data?
- \end{itemize}
- \section{Minor Remarks}
- I stopped looking for typos in section 4.1.
- \begin{itemize}
- \item \enquote{we}: It is a single author. Why does he write \enquote{we}?
- \item should be lower case:
- \begin{itemize}
- \item \enquote{Architecture} should be lower case
- \item \enquote{Classification Challenge} should be lower case
- \item \enquote{Classification}, \enquote{Localization}, \enquote{Detection}, \enquote{Segmentation}
- \item \enquote{Tasks}
- \item \enquote{Layer}
- \item \enquote{Nonlinearities}
- \item \enquote{Semantic Segmentation}
- \end{itemize}
- \item typos (missing characters like commas, switched characters, \dots)
- \begin{itemize}
- \item \enquote{as fellows}
- \item \enquote{descripe}
- \item \enquote{architeture}
- \item \enquote{a translation invariance functions}
- \item \enquote{$f$ is than applied}
- \item \enquote{To archive that $f_{ks}$ is chosen}
- \item \enquote{an MLP}
- \item \enquote{In convolutional layers stride is usually choose to be $s = 1$ ,}
- \item \enquote{applies non-learnable function}
- \item \enquote{to learn nonlinear function} - \enquote{a} is missing
- \item \enquote{this models}
- \item \enquote{Fully Convolutional Networks (FCN)} - missing plural s in (FCNs)
- \item \enquote{FCN are an architecture} - mixed singular and plural. \enquote{A FCN is an architecture\dots}
- \item \enquote{approaches ConvNets} - comma missing
- \item \enquote{relevant} $\neq$ \enquote{relevance}
- \item \enquote{itself will be a ConvNet, that means} - replace the comma by a point. This sentence is too long.
- \item \enquote{only downside is, that} - remove comma
- \end{itemize}
- \item Typography
- \begin{itemize}
- \item Why don't you include \texttt{hyperref}? I really like being able
- to directly jump to the sections, without having to manually
- search them.
- \item I prefer $\mathbb{R}$ instead of $R$. This makes it more obvious
- that it is not a variable, but the set of real numbers.
- \item \verb+\ll+ is nicer than \verb+<<+: $\ll$ vs $<<$.
- \item \verb+exp+ ($exp$) are three variables. The function is \verb+\exp+ ($\exp$). Same for $\tanh$.
- \item \enquote{A recent break-trough has been achieved with} - That seems to be a good point to start a new paragraph.
- \end{itemize}
- \item \enquote{[...], the ImageNet Classification Challenge} should be
- followed by a comma
- \item \enquote{have broken new records}: either \enquote{have broken records}
- or something like \enquote{have set new records}
- \item \enquote{For the pooling layer typically s is choose to be k} - I would write \enquote{For the pooling layer $s$ is typically choosen to be equal to $k$}
- \item \enquote{to further computer vision tasks} - I'm not too sure if you can say \enquote{further} in this context
- \end{itemize}
- \end{document}
|