Skip to content

Commit

Permalink
Fix typos (keras-team#9138)
Browse files Browse the repository at this point in the history
  • Loading branch information
taehoonlee authored and fchollet committed Jan 20, 2018
1 parent d4f6098 commit abe06d5
Show file tree
Hide file tree
Showing 5 changed files with 6 additions and 6 deletions.
2 changes: 1 addition & 1 deletion examples/mnist_denoising_autoencoder.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
'''Trains a denoising autoenconder on MNIST dataset.
'''Trains a denoising autoencoder on MNIST dataset.
Denoising is one of the classic applications of autoencoders.
The denoising process removes unwanted noise that corrupted the
Expand Down
4 changes: 2 additions & 2 deletions keras/applications/nasnet.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@
| NASNet-A (6 @ 4032) | 82.7 % | 96.2 % | 23.8 B | 88.9 |
--------------------------------------------------------------------------------
Weights obtained from the official Tensorflow repository found at
Weights obtained from the official TensorFlow repository found at
https://github.com/tensorflow/models/tree/master/research/slim/nets/nasnet
# References
Expand Down Expand Up @@ -145,7 +145,7 @@ def NASNet(input_shape=None,
backend that does not support separable convolutions.
'''
if K.backend() != 'tensorflow':
raise RuntimeError('Only Tensorflow backend is currently supported, '
raise RuntimeError('Only TensorFlow backend is currently supported, '
'as other backends do not support '
'separable convolution.')

Expand Down
2 changes: 1 addition & 1 deletion keras/backend/tensorflow_backend.py
Original file line number Diff line number Diff line change
Expand Up @@ -272,7 +272,7 @@ def _get_available_gpus():
def _has_nchw_support():
"""Check whether the current scope supports NCHW ops.
Tensorflow does not support NCHW on CPU. Therefore we check if we are not explicitly put on
TensorFlow does not support NCHW on CPU. Therefore we check if we are not explicitly put on
CPU, and have GPUs available. In this case there will be soft-placing on the GPU device.
# Returns
Expand Down
2 changes: 1 addition & 1 deletion keras/engine/training.py
Original file line number Diff line number Diff line change
Expand Up @@ -425,7 +425,7 @@ def weighted(y_true, y_pred, weights, mask=None):
# score_array has ndim >= 2
score_array = fn(y_true, y_pred)
if mask is not None:
# Cast the mask to floatX to avoid float64 upcasting in theano
# Cast the mask to floatX to avoid float64 upcasting in Theano
mask = K.cast(mask, K.floatx())
# mask should have the same shape as score_array
score_array *= mask
Expand Down
2 changes: 1 addition & 1 deletion keras/models.py
Original file line number Diff line number Diff line change
Expand Up @@ -152,7 +152,7 @@ def get_json_type(obj):
for i, (w, val) in enumerate(zip(symbolic_weights,
weight_values)):
# Default values of symbolic_weights is /variable
# for theano and cntk
# for Theano and CNTK
if K.backend() == 'theano' or K.backend() == 'cntk':
if hasattr(w, 'name'):
if w.name.split('/')[-1] == 'variable':
Expand Down

0 comments on commit abe06d5

Please sign in to comment.