DOC: Resolve errors in Ubuntu Sanity checks
This commit is contained in:
parent
8856a4f547
commit
f113e0e361
@ -104,7 +104,8 @@ def elu(x, alpha=1.0):
|
|||||||
|
|
||||||
>>> from tensorflow.keras import models, layers
|
>>> from tensorflow.keras import models, layers
|
||||||
>>> model = models.Sequential()
|
>>> model = models.Sequential()
|
||||||
>>> model.add(layers.Conv2D(32, (3, 3), activation='elu', input_shape=(28, 28, 1)))
|
>>> model.add(layers.Conv2D(32, (3, 3), activation='elu',
|
||||||
|
... input_shape=(28, 28, 1)))
|
||||||
>>> model.add(layers.MaxPooling2D((2, 2)))
|
>>> model.add(layers.MaxPooling2D((2, 2)))
|
||||||
>>> model.add(layers.Conv2D(64, (3, 3), activation='elu'))
|
>>> model.add(layers.Conv2D(64, (3, 3), activation='elu'))
|
||||||
>>> model.add(layers.MaxPooling2D((2, 2)))
|
>>> model.add(layers.MaxPooling2D((2, 2)))
|
||||||
@ -124,8 +125,8 @@ def elu(x, alpha=1.0):
|
|||||||
|
|
||||||
|
|
||||||
Reference:
|
Reference:
|
||||||
[Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs)
|
[Fast and Accurate Deep Network Learning by Exponential Linear Units
|
||||||
(Clevert et al, 2016)](https://arxiv.org/abs/1511.07289)
|
(ELUs) (Clevert et al, 2016)](https://arxiv.org/abs/1511.07289)
|
||||||
"""
|
"""
|
||||||
return K.elu(x, alpha)
|
return K.elu(x, alpha)
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user