Skip to content

Commit 6a3f6d9

Browse files
committed
Fixed typos
1 parent 8cff23c commit 6a3f6d9

File tree

1 file changed

+7
-5
lines changed

1 file changed

+7
-5
lines changed

doc/lstm.txt

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -19,8 +19,8 @@ is positive or negative. This is a binary classification task.
1919
Data
2020
++++
2121

22-
As previously mentionned, the provided scripts are used to train a LSTM
23-
recurrent neural on the Large Movie Review Dataset dataset.
22+
As previously mentioned, the provided scripts are used to train a LSTM
23+
recurrent neural network on the Large Movie Review Dataset dataset.
2424

2525
While the dataset is public, in this tutorial we provide a copy of the dataset
2626
that has previously been preprocessed according to the needs of this LSTM
@@ -40,11 +40,13 @@ matrix associated with the connections between the neurons of the recurrent
4040
hidden layer. This means that, the magnitude of weights in the transition
4141
matrix can have a strong impact on the learning process.
4242

43-
If the weights in this matrix are small, it can lead to a situation called
43+
If the weights in this matrix are small (or, more formally, if the leading
44+
eigenvalue of the weight matrix is small), it can lead to a situation called
4445
*vanishing gradients* where the gradient signal gets so small that learning
4546
either becomes very slow or stops working altogether. It can also make more
4647
difficult the task of learning long-term dependencies in the data.
47-
Conversely, if the weights in this matrix are large, it can lead to a
48+
Conversely, if the weights in this matrix are large (or, again, more formally,
49+
if the leading eigenvalue of the weight matrix is large), it can lead to a
4850
situation where the gradient signal is so large that it can cause learning to
4951
diverge. This is often referred to as *exploding gradients*.
5052

@@ -55,7 +57,7 @@ connection (a connection to itself), a forget gate and an output gate. The
5557
self-recurrent connection has a weight of 1.0 and ensures that, barring any
5658
outside interference, the state of a memory cell can remain constant from one
5759
timestep to another. The gates serve to modulate the interactions between the
58-
memory cell and itself and its environment. The input gate can allow incoming
60+
memory cell itself and its environment. The input gate can allow incoming
5961
signal to alter the state of the memory cell or block it. On the other hand,
6062
the output gate can allow the state of the memory cell to have an effect on
6163
other neurons or prevent it. Finally, the forget gate can modulate the memory

0 commit comments

Comments
 (0)