Skip to content

Commit 6a1a2a0

Browse files
Update rnn.md
1 parent 655b346 commit 6a1a2a0

File tree

1 file changed

+13
-12
lines changed

1 file changed

+13
-12
lines changed

rnn.md

Lines changed: 13 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ permalink: /rnn/
66

77
Table of Contents:
88

9-
- [Intro to RNN](#intro)
9+
- [Introduction to RNN](#intro)
1010
- [RNN example as Character-level language model](#char)
1111
- [Multilayer RNNs](#multi)
1212
- [Long-Short Term Memory (LSTM)](#lstm)
@@ -16,7 +16,7 @@ Table of Contents:
1616

1717
<a name='intro'></a>
1818

19-
## Intro to RNN
19+
## Introduction to RNN
2020

2121
In this lecture note, we're going to be talking about the Recurrent Neural Networks (RNNs). One
2222
great thing about the RNNs is that they offer a lot of flexibility on how we wire up the neural
@@ -25,25 +25,26 @@ input vector (red), then we process it with some hidden layers (green), and we
2525
fixed sized output vector (blue) as depicted in the leftmost model ("Vanilla" Neural Networks) in Figure 1.
2626
While **"Vanilla" Neural Networks** receive a single input and produce one label for that image, there are tasks where
2727
the model produce a sequence of outputs as shown in the one-to-many model in Figure 1. **Recurrent Neural Networks** allow
28-
us to operate over sequences of input, output, or both at the same time. An example of **one-to-many** model is image
29-
captioning where we are given a fixed sized image and produce a sequence of words that describe the content of that image through RNN (second model in Figure 1).
30-
An example of **many-to-one** task is action prediction where we look at a sequence of video frames instead of a single image and produce
28+
us to operate over sequences of input, output, or both at the same time.
29+
* An example of **one-to-many** model is image captioning where we are given a fixed sized image and produce a sequence of words that describe the content of that image through RNN (second model in Figure 1).
30+
* An example of **many-to-one** task is action prediction where we look at a sequence of video frames instead of a single image and produce
3131
a label of what action was happening in the video as shown in the third model in Figure 1. Another example of many-to-one task is
32-
sentiment classification in NLP where we are given a sequence of words
33-
of a sentence and then classify what sentiment (e.g. positive or negative) that sentence is.
34-
An example of **many-to-many** task is video-captioning where the input is a sequence of video frames and the output is caption that describes
32+
sentiment classification in NLP where we are given a sequence of words of a sentence and then classify what sentiment (e.g. positive or negative) that sentence is.
33+
* An example of **many-to-many** task is video-captioning where the input is a sequence of video frames and the output is caption that describes
3534
what was in the video as shown in the fourth model in Figure 1. Another example of many-to-many task is machine translation in NLP, where we can have an
36-
RNN that takes a sequence of words of a sentence in English, and then this RNN is asked to produce
37-
a sequence of words of a sentence in French. There is a also a variation of many-to-many task as shown in the last model in Figure 1,
35+
RNN that takes a sequence of words of a sentence in English, and then this RNN is asked to produce a sequence of words of a sentence in French.
36+
* There is a also a variation of many-to-many task as shown in the last model in Figure 1,
3837
where the model generates an output at every timestep. An example of this many-to-many task is video classification on a frame level
3938
where the model classifies every single frame of video with some number of classes. We should note that we don't want
4039
this prediction to only be a function of the current timestep (current frame of the video), but also all the timesteps (frames)
41-
that have come before this video. In general, RNNs allow us to wire up an architecture, where the prediction at every single timestep is a
40+
that have come before this video.
41+
42+
In general, RNNs allow us to wire up an architecture, where the prediction at every single timestep is a
4243
function of all the timesteps that have come before.
4344

4445
<div class="fig figcenter fighighlight">
4546
<img src="/assets/rnn/types.png" width="100%">
46-
<div class="figcaption">Figure 1. Different (non-exhaustive) types of Recurrent Neural Network architectures. Red boxes are input vectors. Green boxes are hidden layers. Blue boxes are output vectors.</div>
47+
<div class="figcaption"> <b> Figure 1.</b> Different (non-exhaustive) types of Recurrent Neural Network architectures. Red boxes are input vectors. Green boxes are hidden layers. Blue boxes are output vectors.</div>
4748
</div>
4849

4950
A Recurrent Neural Network is basically a blackbox (Figure 2), where it has a state and it receives through

0 commit comments

Comments
 (0)