Add the missing fully connected layer 2 in the RNN example
@@ -116,6 +116,7 @@
" # Set forward pass.\n",
" def call(self, x, is_training=False):\n",
" x = self.fc1(x)\n",
+ " x = self.fc2(x)\n",
" x = self.out(x)\n",
" if not is_training:\n",
" # tf cross entropy expect logits without softmax, so only\n",