Add Dropout layer.

This commit is contained in:
Francois Chollet 2023-04-18 19:22:13 -07:00
parent f6df67f2d2
commit 3aa1d977b1
5 changed files with 114 additions and 5 deletions

@ -5,11 +5,12 @@ class SeedGenerator:
def __init__(self, seed):
from keras_core.backend import Variable
if seed is None:
seed = make_default_seed()
if not isinstance(seed, int):
raise ValueError(
"Argument `seed` must be an integer. " f"Received: seed={seed}"
)
seed = seed or make_default_seed()
self.state = Variable([seed, 0], dtype="uint32", trainable=False)

@ -2,5 +2,4 @@ from keras_core.layers.core.dense import Dense
from keras_core.layers.core.input_layer import Input
from keras_core.layers.core.input_layer import InputLayer
from keras_core.layers.layer import Layer
# from keras_core.layers.regularization.dropout import Dropout
from keras_core.layers.regularization.dropout import Dropout

@ -0,0 +1,77 @@
from keras_core import backend
from keras_core import layers
from keras_core.api_export import keras_core_export
@keras_core_export("keras_core.layers.Dropout")
class Dropout(layers.Layer):
"""Applies dropout to the input.
The `Dropout` layer randomly sets input units to 0 with a frequency of `rate`
at each step during training time, which helps prevent overfitting.
Inputs not set to 0 are scaled up by `1 / (1 - rate)` such that the sum over
all inputs is unchanged.
Note that the `Dropout` layer only applies when `training` is set to `True`
in `call()`, such that no values are dropped during inference.
When using `model.fit`, `training` will be appropriately set to `True`
automatically. In other contexts, you can set the argument explicitly
to `True` when calling the layer.
(This is in contrast to setting `trainable=False` for a `Dropout` layer.
`trainable` does not affect the layer's behavior, as `Dropout` does
not have any variables/weights that can be frozen during training.)
Args:
rate: Float between 0 and 1. Fraction of the input units to drop.
noise_shape: 1D integer tensor representing the shape of the
binary dropout mask that will be multiplied with the input.
For instance, if your inputs have shape
`(batch_size, timesteps, features)` and
you want the dropout mask to be the same for all timesteps,
you can use `noise_shape=(batch_size, 1, features)`.
seed: A Python integer to use as random seed.
Call arguments:
inputs: Input tensor (of any rank).
training: Python boolean indicating whether the layer should behave in
training mode (adding dropout) or in inference mode (doing nothing).
"""
def __init__(
self, rate, noise_shape=None, seed=None, name=None, dtype=None
):
super().__init__(name=name, dtype=dtype)
if isinstance(rate, (int, float)) and not 0 <= rate <= 1:
raise ValueError(
f"Invalid value received for argument "
"`rate`. Expected a float value between 0 and 1. "
f"Received: rate={rate}"
)
self.rate = rate
self.seed = seed
self.noise_shape = noise_shape
self.seed_generator = backend.random.SeedGenerator(seed)
self.supports_masking = True
def call(self, inputs, training=False):
if training and self.rate > 0:
return backend.random.dropout(
inputs,
self.rate,
noise_shape=self.noise_shape,
seed=self.seed_generator,
)
return inputs
def compute_output_shape(self, input_shape):
return input_shape
def get_config(self):
base_config = super().get_config()
config = {
"rate": self.rate,
"seed": self.seed,
"noise_shape": self.noise_shape,
}
return {**base_config, **config}

@ -0,0 +1,34 @@
import numpy as np
import pytest
from keras_core import backend
from keras_core import layers
from keras_core import testing
class DropoutTest(testing.TestCase):
def test_dropout_supports_masking(self):
dropout = layers.Dropout(0.5)
self.assertEqual(True, dropout.supports_masking)
def test_dropout_rescaling(self):
inputs = np.ones((20, 500))
layer = layers.Dropout(0.5, seed=1337)
outputs = layer(inputs, training=True)
self.assertAllClose(np.mean(outputs), 1.0, atol=0.02)
self.assertAllClose(np.max(outputs), 2.0)
@pytest.mark.skipif(
backend.backend() != "tensorflow", reason="Requires dynamic shapes"
)
def test_dropout_partial_noise_shape_dynamic(self):
inputs = np.ones((20, 5, 10))
layer = layers.Dropout(0.5, noise_shape=(None, 1, None))
outputs = layer(inputs, training=True)
self.assertAllClose(outputs[:, 0, :], outputs[:, 1, :])
def test_dropout_partial_noise_shape_static(self):
inputs = np.ones((20, 5, 10))
layer = layers.Dropout(0.5, noise_shape=(20, 1, 10))
outputs = layer(inputs, training=True)
self.assertAllClose(outputs[:, 0, :], outputs[:, 1, :])

@ -1,8 +1,6 @@
import numpy as np
from keras_core import backend
from keras_core import initializers
from keras_core import operations as ops
from keras_core import regularizers
from keras_core import testing