paper-peer-review.tex 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215
  1. \documentclass[a4paper,9pt]{scrartcl}
  2. \usepackage{amssymb, amsmath} % needed for math
  3. \usepackage[utf8]{inputenc} % this is needed for umlauts
  4. \usepackage[USenglish]{babel} % this is needed for umlauts
  5. \usepackage[T1]{fontenc} % this is needed for correct output of umlauts in pdf
  6. \usepackage[margin=2.5cm]{geometry} %layout
  7. \usepackage{hyperref} % hyperlinks
  8. \usepackage{color}
  9. \usepackage{framed}
  10. \usepackage{enumerate} % for advanced numbering of lists
  11. \usepackage{csquotes} % for enquote
  12. \newcommand\titletext{Peer-Review of\\"Deep Neuronal Networks for Semantiv Segmentation in Medical
  13. Informatics"}
  14. \title{\titletext}
  15. \author{Martin Thoma}
  16. \hypersetup{
  17. pdfauthor = {Martin Thoma},
  18. pdfkeywords = {peer review},
  19. pdftitle = {Lineare Algebra}
  20. }
  21. \usepackage{microtype}
  22. \begin{document}
  23. \maketitle
  24. \section{Introduction}
  25. This is a peer-review of \enquote{Deep Neuronal Networks for Semantiv
  26. Segmentation in Medical Informatics} by Marvin Teichmann. The reviewed document
  27. is available under \href{https://github.com/MarvinTeichmann/seminar-pixel-exact-classification.git}{https://github.com/MarvinTeichmann/seminar-pixel-exact-classification.git}, version
  28. \texttt{b1bdb4802c8e268ebf7ca66adb7f806e29afb413}.
  29. \section{Summary of the Content}
  30. The author wants to describe how convolutional networks can be used for
  31. semantic segmentation tasks in medicine. To do so, he introduces Convolutional
  32. Neural Networks.
  33. As the introduction, section~2 (Computer Vision Tasks) and section~5
  34. (Application in Medical Informatics) are not written yet, it can only be said
  35. that the plan of writing them is good.
  36. The author expects the reader to know how neural networks work in general, but
  37. gives a detailed introduction into CNNs. He continues with explaining fully
  38. convolutional networks (FCNs). This leads in a natural fashion to the
  39. application of neural networks for segmentation.
  40. \section{Overall Feedback}
  41. Gramatical errors make it sometimes difficult to understand relatively easy
  42. sentences. Also, the missing parts make it difficult to see if there is a
  43. consistent overall structure.
  44. I recommend adding more source to claims made in the paper.
  45. The overall structure seems to be logical, definitions are given most of the
  46. time (see the feedback below for some exceptions where it should be added).
  47. \section{Major Remarks}
  48. \subsection{Section 3 / 3.1: CNNs}
  49. \begin{itemize}
  50. \item What is \enquote{stationarity of statistics}?
  51. \item What are \enquote{translation invariance functions}?
  52. \item The term \enquote{Kernel} and \enquote{reception field} were neither
  53. introduced nor a source was given where the reader could find
  54. definitions.
  55. \item What is a \enquote{channel size}? Do you mean the number of channels
  56. or the channel dimension?
  57. \item What is $F_{nm}$? A function, but on which domain does it operate and
  58. to which domain does it map? What does this function mean? Is it
  59. an activation function?
  60. \item What does $n << h,w$ mean? $n \ll \min(h, w)$?
  61. \item It was not explained what \enquote{a sliding window fashion} means.
  62. \item I miss an~image in section 3.1 (definitions and notation).
  63. \end{itemize}
  64. \subsection{Section 3.2: Layer types}
  65. \begin{itemize}
  66. \item I've never heard of activation layers. Do you mean fully connected
  67. layers? If not, then you should probably cite a publication which
  68. calls it like that.
  69. \item \enquote{curtained weights} - what is that? (The problem might be
  70. my lack of knowledge of the English language). However, I think
  71. you should cite a source here for the claim that this is possible.
  72. \item \enquote{a variety of tasks including edge and area detection,
  73. contrast sharpening and image blurring}: I miss a source.
  74. \item \enquote{big ($k \geq 7$). [KSH12, SZ14, SLJ + 14].} - What exactly
  75. do you cite here?
  76. \item An image with a tiny example would make the pooling layer much
  77. easier to understand. However, you can also cite a source which
  78. explains this well.
  79. \item The sentence \enquote{Firstly it naturally reduces the spatial dimension
  80. enabling the network to learn more compact representation if the data and decreasing the
  81. amount of parameters in the succeeding layers.} sounds wrong. You forgot something
  82. At \enquote{if the data}.
  83. \item The sentence is gramatically wrong and makes it hard to understand
  84. \enquote{Secondly it introduces robust translation invariant.}.
  85. \item \enquote{Minor shifts in the input data will not result in the same activation after pooling.}
  86. Not? I thought that was the advantage of pooling, that you get
  87. invariant?
  88. \item \enquote{Recently ReLU Nonlinearities [KSH12](AlexNet, Bolzmann)}:
  89. It is possible to make that easier to read:
  90. \enquote{Recently ReLU nonlinearities, as introduced by~[KSH12](AlexNet, Bolzmann)}
  91. - However, I'm not too sure what you mean with \enquote{Bolzmann}.
  92. \item It was not explained / defined what ReLU means / is.
  93. \end{itemize}
  94. \subsection{Section 4: Neural Networks for Segmentation}
  95. \begin{itemize}
  96. \item \enquote{After the overwhelming successes of DCNNs in image classification}: Add source
  97. \item \enquote{in combination with traditional classifiers} - What are \enquote{traditional} classifiers?
  98. \item \enquote{Other authors used the idea described in Section 2} - Don't make me jump back. Can you give that idea a short name? Then you can write something like \enquote{the idea of sliding windows}. As you wrote about sliding windows in the rest of the sentence, I guess restrucuting the sentence might help.
  99. \item \enquote{are currently the state-of-the art in several semantic segmentation benchmarks.} - name at least one.
  100. \end{itemize}
  101. \subsection{Section 4.1: Sliding Window efficiency in CNNs}
  102. \begin{itemize}
  103. \item \enquote{The input image will be down sampled by a factor of s corresponding to the product of all strides being applied in $C'$.} - I don't think that is obvious. Please explain it or give a source for that claim.
  104. \item \enquote{shift-and-stitch} - What is that?
  105. \end{itemize}
  106. \subsection{Section 4.2: FCNs}
  107. \begin{itemize}
  108. \item \enquote{builds up on the ideas presented of Section 4.1} - which ones?
  109. The \textit{sliding-window-as-a-convoluton} idea and which other idea?
  110. \item \enquote{they are not trying to avoid downsampling as part of the progress}
  111. - do you mean process?
  112. \item Explain what an \enquote{upsampling layer} is.
  113. \end{itemize}
  114. \subsection{Section 4.2.1: Deconvolution}
  115. This section is still to be done.
  116. \subsection{Section 4.2.2: Skip-Architecture}
  117. An image would help, although I guess it is already easy to understand.
  118. \subsection{4.2.3 Transfer Learning}
  119. \begin{itemize}
  120. \item What is transfer lerning?
  121. \item What is VGG16 (cite paper) - same for AlexNet and GoogLeNet, if it
  122. wasn't done already. People who don't know what a CNN is will also
  123. not know what AlexNet / GoogLeNet is.
  124. \end{itemize}
  125. \subsection{4.3 Extensions of FCN}
  126. \begin{itemize}
  127. \item \enquote{Several extensions of FCN have been proposed} - give sources
  128. \item \enquote{of strong labeled data} what is \textbf{strong} labeled data?
  129. \end{itemize}
  130. \section{Minor Remarks}
  131. I stopped looking for typos in section 4.1.
  132. \begin{itemize}
  133. \item \enquote{we}: It is a single author. Why does he write \enquote{we}?
  134. \item should be lower case:
  135. \begin{itemize}
  136. \item \enquote{Architecture} should be lower case
  137. \item \enquote{Classification Challenge} should be lower case
  138. \item \enquote{Classification}, \enquote{Localization}, \enquote{Detection}, \enquote{Segmentation}
  139. \item \enquote{Tasks}
  140. \item \enquote{Layer}
  141. \item \enquote{Nonlinearities}
  142. \item \enquote{Semantic Segmentation}
  143. \end{itemize}
  144. \item typos (missing characters like commas, switched characters, \dots)
  145. \begin{itemize}
  146. \item \enquote{as fellows}
  147. \item \enquote{descripe}
  148. \item \enquote{architeture}
  149. \item \enquote{a translation invariance functions}
  150. \item \enquote{$f$ is than applied}
  151. \item \enquote{To archive that $f_{ks}$ is chosen}
  152. \item \enquote{an MLP}
  153. \item \enquote{In convolutional layers stride is usually choose to be $s = 1$ ,}
  154. \item \enquote{applies non-learnable function}
  155. \item \enquote{to learn nonlinear function} - \enquote{a} is missing
  156. \item \enquote{this models}
  157. \item \enquote{Fully Convolutional Networks (FCN)} - missing plural s in (FCNs)
  158. \item \enquote{FCN are an architecture} - mixed singular and plural. \enquote{A FCN is an architecture\dots}
  159. \item \enquote{approaches ConvNets} - comma missing
  160. \item \enquote{relevant} $\neq$ \enquote{relevance}
  161. \item \enquote{itself will be a ConvNet, that means} - replace the comma by a point. This sentence is too long.
  162. \item \enquote{only downside is, that} - remove comma
  163. \end{itemize}
  164. \item Typography
  165. \begin{itemize}
  166. \item Why don't you include \texttt{hyperref}? I really like being able
  167. to directly jump to the sections, without having to manually
  168. search them.
  169. \item I prefer $\mathbb{R}$ instead of $R$. This makes it more obvious
  170. that it is not a variable, but the set of real numbers.
  171. \item \verb+\ll+ is nicer than \verb+<<+: $\ll$ vs $<<$.
  172. \item \verb+exp+ ($exp$) are three variables. The function is \verb+\exp+ ($\exp$). Same for $\tanh$.
  173. \item \enquote{A recent break-trough has been achieved with} - That seems to be a good point to start a new paragraph.
  174. \end{itemize}
  175. \item \enquote{[...], the ImageNet Classification Challenge} should be
  176. followed by a comma
  177. \item \enquote{have broken new records}: either \enquote{have broken records}
  178. or something like \enquote{have set new records}
  179. \item \enquote{For the pooling layer typically s is choose to be k} - I would write \enquote{For the pooling layer $s$ is typically choosen to be equal to $k$}
  180. \item \enquote{to further computer vision tasks} - I'm not too sure if you can say \enquote{further} in this context
  181. \end{itemize}
  182. \end{document}