Skip to content

Commit e6ade48

Browse files
author
Yoshua Bengio
committed
added pointers to ift6266 course notes on the same subjects
1 parent bb46237 commit e6ade48

3 files changed

Lines changed: 14 additions & 5 deletions

File tree

doc/gettingstarted.txt

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,7 @@ List of Symbols and acronyms
153153

154154
* :math:`D`: number of input dimensions.
155155
* :math:`D_h^{(i)}`: number of hidden units in the :math:`i`-th layer.
156-
* :math:`f_{\theta}(x)`, :math:`f(x)`: classification function associated with a model :math:`P(Y|x,\theta)`, defined as :math:`argmax_k P(Y=k|x,\theta)`.
156+
* :math:`f_{\theta}(x)`, :math:`f(x)`: classification function associated with a model :math:`P(Y|x,\theta)`, defined as :math:`{\rm argmax}_k P(Y=k|x,\theta)`.
157157
Note that we will often drop the :math:`\theta` subscript.
158158
* L: number of labels.
159159
* :math:`\mathcal{L}(\theta, \cal{D})`: log-likelihood :math:`\cal{D}`
@@ -189,7 +189,9 @@ utility of unsupervised *pre-training* is often evaluated on the basis of what
189189
performance can be achieved after supervised *fine-tuning*. This chapter
190190
reviews the basics of supervised learning for classification models, and covers
191191
the minibatch stochastic gradient descent algorithm that is used to fine-tune
192-
many of the models in the Deep Learning Tutorials.
192+
many of the models in the Deep Learning Tutorials. Have a look at these
193+
`introductory course notes on gradient-based learning <http://www.iro.umontreal.ca/~pift6266/H10/notes/gradient.html>`_
194+
for more basics on the notion of optimizing a training criterion using the gradient.
193195

194196

195197
.. _opt_learn_classifier:
@@ -228,7 +230,7 @@ In this tutorial, :math:`f` is defined as:
228230

229231
.. math::
230232

231-
f(x) = argmax_k P(Y=k | x, \theta)
233+
f(x) = {\rm argmax}_k P(Y=k | x, \theta)
232234

233235
In python, using Theano this can be written as :
234236

doc/intro.txt

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,12 @@
22
Deep Learning Tutorials
33
=======================
44

5+
Deep Learning is a new area of Machine Learning research, which
6+
has been introduced with the objective of moving Machine Learning
7+
closer to one of its original goals: Artificial Intelligence.
8+
See these course notes for a `brief introduction to Machine Learning for AI <http://www.iro.umontreal.ca/~pift6266/H10/notes/mlintro.html>`_
9+
and an `introduction to Deep Learning algorithms <http://www.iro.umontreal.ca/~pift6266/H10/notes/deepintro.html>`_.
10+
511
Deep Learning is about learning multiple levels of representation
612
and abstraction that help to
713
make sense of data such as images, sound, and text.
@@ -12,7 +18,7 @@ For more about deep learning algorithms, see for example:
1218
- The LISA `public wiki <http://www.iro.umontreal.ca/~lisa/twiki/bin/view.cgi/Public/WebHome>`_ has a `reading list <http://www.iro.umontreal.ca/~lisa/twiki/bin/view.cgi/Public/ReadingOnDeepNetworks>`_ and a `bibliography <http://www.iro.umontreal.ca/~lisa/twiki/bin/view.cgi/Public/DeepNetworksBibliography>`_.
1319
- Geoff Hinton has `readings <http://www.cs.toronto.edu/~hinton/deeprefs.html>`_ from last year's `NIPS tutorial <http://videolectures.net/jul09_hinton_deeplearn/>`_.
1420

15-
These tutorials will introduce you to some of the most important deep learning
21+
The tutorials presented here will introduce you to some of the most important deep learning
1622
algorithms and will also show you how to run them using Theano_. Theano is a python library that makes writing deep learning models easy, and gives the option of
1723
training them on a GPU.
1824

doc/mlp.txt

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,8 @@ input data into a space where it becomes linearly separable. This intermediate
3333
layer is referred to as a **hidden layer**. A single hidden layer is
3434
sufficient to make MLPs a **universal approximator**. However we will see later
3535
on that there are substantial benefits to using many such hidden layers, i.e. the
36-
very premise of **deep learning**.
36+
very premise of **deep learning**. See these course notes for an `introduction
37+
to MLPs, the back-propagation algorithm, and how to train MLPs <http://www.iro.umontreal.ca/~pift6266/H10/notes/mlp.html>`_.
3738

3839
This tutorial will again tackle the problem of MNIST digit classification.
3940

0 commit comments

Comments
 (0)