Skip to content
This repository was archived by the owner on Jul 7, 2023. It is now read-only.

Commit 8287094

Browse files
committed
Reword comments
1 parent 9f02a51 commit 8287094

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

tensor2tensor/models/shake_shake.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -82,8 +82,8 @@ def model_fn_body(self, features):
8282
blocks_per_stage = (hparams.num_hidden_layers - 2) // 6
8383

8484
# For canonical Shake-Shake, the entry flow is a 3x3 convolution with 16
85-
# filters then a batch norm. Instead we use the one in SmallImageModality,
86-
# which also seems to include a layer norm.
85+
# filters then a batch norm. Instead we will rely on the one in
86+
# SmallImageModality, which seems to instead use a layer norm.
8787
x = inputs
8888
mode = hparams.mode
8989
with tf.variable_scope('shake_shake_stage_1'):
@@ -97,7 +97,7 @@ def model_fn_body(self, features):
9797

9898
# For canonical Shake-Shake, we should perform 8x8 average pooling and then
9999
# have a fully-connected layer (which produces the logits for each class).
100-
# Instead, we just use the Xception exit flow in ClassLabelModality.
100+
# Instead, we rely on the Xception exit flow in ClassLabelModality.
101101
#
102102
# Also, this model_fn does not return an extra_loss. However, TensorBoard
103103
# reports an exponential moving average for extra_loss, where the initial

0 commit comments

Comments
 (0)