@@ -19,8 +19,8 @@ is positive or negative. This is a binary classification task.
1919Data
2020++++
2121
22- As previously mentionned , the provided scripts are used to train a LSTM
23- recurrent neural on the Large Movie Review Dataset dataset.
22+ As previously mentioned , the provided scripts are used to train a LSTM
23+ recurrent neural network on the Large Movie Review Dataset dataset.
2424
2525While the dataset is public, in this tutorial we provide a copy of the dataset
2626that has previously been preprocessed according to the needs of this LSTM
@@ -40,11 +40,13 @@ matrix associated with the connections between the neurons of the recurrent
4040hidden layer. This means that, the magnitude of weights in the transition
4141matrix can have a strong impact on the learning process.
4242
43- If the weights in this matrix are small, it can lead to a situation called
43+ If the weights in this matrix are small (or, more formally, if the leading
44+ eigenvalue of the weight matrix is small), it can lead to a situation called
4445*vanishing gradients* where the gradient signal gets so small that learning
4546either becomes very slow or stops working altogether. It can also make more
4647difficult the task of learning long-term dependencies in the data.
47- Conversely, if the weights in this matrix are large, it can lead to a
48+ Conversely, if the weights in this matrix are large (or, again, more formally,
49+ if the leading eigenvalue of the weight matrix is large), it can lead to a
4850situation where the gradient signal is so large that it can cause learning to
4951diverge. This is often referred to as *exploding gradients*.
5052
@@ -55,7 +57,7 @@ connection (a connection to itself), a forget gate and an output gate. The
5557self-recurrent connection has a weight of 1.0 and ensures that, barring any
5658outside interference, the state of a memory cell can remain constant from one
5759timestep to another. The gates serve to modulate the interactions between the
58- memory cell and itself and its environment. The input gate can allow incoming
60+ memory cell itself and its environment. The input gate can allow incoming
5961signal to alter the state of the memory cell or block it. On the other hand,
6062the output gate can allow the state of the memory cell to have an effect on
6163other neurons or prevent it. Finally, the forget gate can modulate the memory
0 commit comments