This repository was archived by the owner on Jul 7, 2023. It is now read-only.
File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -82,8 +82,8 @@ def model_fn_body(self, features):
8282 blocks_per_stage = (hparams .num_hidden_layers - 2 ) // 6
8383
8484 # For canonical Shake-Shake, the entry flow is a 3x3 convolution with 16
85- # filters then a batch norm. Instead we use the one in SmallImageModality,
86- # which also seems to include a layer norm.
85+ # filters then a batch norm. Instead we will rely on the one in
86+ # SmallImageModality, which seems to instead use a layer norm.
8787 x = inputs
8888 mode = hparams .mode
8989 with tf .variable_scope ('shake_shake_stage_1' ):
@@ -97,7 +97,7 @@ def model_fn_body(self, features):
9797
9898 # For canonical Shake-Shake, we should perform 8x8 average pooling and then
9999 # have a fully-connected layer (which produces the logits for each class).
100- # Instead, we just use the Xception exit flow in ClassLabelModality.
100+ # Instead, we rely on the Xception exit flow in ClassLabelModality.
101101 #
102102 # Also, this model_fn does not return an extra_loss. However, TensorBoard
103103 # reports an exponential moving average for extra_loss, where the initial
You can’t perform that action at this time.
0 commit comments