File tree Expand file tree Collapse file tree 3 files changed +6
-6
lines changed
tensorflow/g3doc/tutorials Expand file tree Collapse file tree 3 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -390,7 +390,7 @@ summary_writer = tf.train.SummaryWriter(FLAGS.train_dir,
390
390
```
391
391
392
392
Lastly, the events file will be updated with new summary values every time the
393
- ` summary_op ` is run and the ouput passed to the writer's ` add_summary() `
393
+ ` summary_op ` is run and the output passed to the writer's ` add_summary() `
394
394
function.
395
395
396
396
``` python
Original file line number Diff line number Diff line change @@ -10,7 +10,7 @@ for an introduction to recurrent neural networks and LSTMs in particular.
10
10
11
11
In this tutorial we will show how to train a recurrent neural network on
12
12
a challenging task of language modeling. The goal of the problem is to fit a
13
- probabilistic model which assigns probablities to sentences. It does so by
13
+ probabilistic model which assigns probabilities to sentences. It does so by
14
14
predicting next words in a text given a history of previous words. For this
15
15
purpose we will use the Penn Tree Bank (PTB) dataset, which is a popular
16
16
benchmark for measuring quality of these models, whilst being small and
@@ -80,7 +80,7 @@ of unrolled steps.
80
80
This is easy to implement by feeding inputs of length ` num_steps ` at a time and
81
81
doing backward pass after each iteration.
82
82
83
- A simplifed version of the code for the graph creation for truncated
83
+ A simplified version of the code for the graph creation for truncated
84
84
backpropagation:
85
85
86
86
``` python
@@ -129,7 +129,7 @@ word_embeddings = tf.nn.embedding_lookup(embedding_matrix, word_ids)
129
129
The embedding matrix will be initialized randomly and the model will learn to
130
130
differentiate the meaning of words just by looking at the data.
131
131
132
- ### Loss Fuction
132
+ ### Loss Function
133
133
134
134
We want to minimize the average negative log probability of the target words:
135
135
Original file line number Diff line number Diff line change @@ -50,7 +50,7 @@ This basic architecture is depicted below.
50
50
Each box in the picture above represents a cell of the RNN, most commonly
51
51
a GRU cell or an LSTM cell (see the [ RNN Tutorial] ( ../../tutorials/recurrent/index.md )
52
52
for an explanation of those). Encoder and decoder can share weights or,
53
- as is more common, use a different set of parameters. Mutli -layer cells
53
+ as is more common, use a different set of parameters. Multi -layer cells
54
54
have been successfully used in sequence-to-sequence models too, e.g. for
55
55
translation [ Sutskever et al., 2014] ( http://arxiv.org/abs/1409.3215 ) .
56
56
@@ -203,7 +203,7 @@ sentence with a special PAD symbol. Then we'd need only one seq2seq model,
203
203
for the padded lengths. But on shorter sentence our model would be inefficient,
204
204
encoding and decoding many PAD symbols that are useless.
205
205
206
- As a compromise between contructing a graph for every pair of lengths and
206
+ As a compromise between constructing a graph for every pair of lengths and
207
207
padding to a single length, we use a number of * buckets* and pad each sentence
208
208
to the length of the bucket above it. In `translate.py` we use the following
209
209
default buckets.
You can’t perform that action at this time.
0 commit comments