tf.keras.layers.AlphaDropout(rate, noise_shape=None, seed=None, **kwargs)
Applies Alpha Dropout to the input.
Alpha Dropout is a
Dropout that keeps mean and variance of inputs
to their original values, in order to ensure the self-normalizing property
even after this dropout.
Alpha Dropout fits well to Scaled Exponential Linear Units
by randomly setting activations to the negative saturation value.
Dropout). The multiplicative noise will have standard deviation
sqrt(rate / (1 - rate)).
Arbitrary. Use the keyword argument
(tuple of integers, does not include the samples axis)
when using this layer as the first layer in a model.
Same shape as input.