diff --git a/keras/applications/efficientnet_v2.py b/keras/applications/efficientnet_v2.py index 16444b93c..aafd71169 100644 --- a/keras/applications/efficientnet_v2.py +++ b/keras/applications/efficientnet_v2.py @@ -34,7 +34,7 @@ from keras.engine import training from keras.utils import data_utils from keras.utils import layer_utils -BASE_WEIGHTS_PATH = "https://storage.googleapis.com/tensorflow/keras-applications/efficientnet_v2/" +BASE_WEIGHTS_PATH = "https://storage.googleapis.com/tensorflow/keras-applications/efficientnet_v2/" # noqa: E501 WEIGHTS_HASHES = { "b0": ( diff --git a/keras/callbacks_test.py b/keras/callbacks_test.py index 31adcc2a7..a9cee284c 100644 --- a/keras/callbacks_test.py +++ b/keras/callbacks_test.py @@ -1319,19 +1319,19 @@ class KerasCallbacksTest(test_combinations.TestCase): return func - test_model_checkpoint_load_weights_on_restart_true_save_weights_only_true = get_ModelCheckpoint_load_weights_on_restart_true_test.__func__( + test_model_checkpoint_load_weights_on_restart_true_save_weights_only_true = get_ModelCheckpoint_load_weights_on_restart_true_test.__func__( # noqa: E501 True ) - test_model_checkpoint_load_weights_on_restart_true_save_weights_only_false = get_ModelCheckpoint_load_weights_on_restart_true_test.__func__( + test_model_checkpoint_load_weights_on_restart_true_save_weights_only_false = get_ModelCheckpoint_load_weights_on_restart_true_test.__func__( # noqa: E501 False ) - test_model_checkpoint_load_weights_on_restart_false_save_weights_only_true = get_ModelCheckpoint_load_weights_on_restart_false_test.__func__( + test_model_checkpoint_load_weights_on_restart_false_save_weights_only_true = get_ModelCheckpoint_load_weights_on_restart_false_test.__func__( # noqa: E501 True ) - test_model_checkpoint_load_weights_on_restart_false_save_weights_only_false = get_ModelCheckpoint_load_weights_on_restart_false_test.__func__( + test_model_checkpoint_load_weights_on_restart_false_save_weights_only_false = get_ModelCheckpoint_load_weights_on_restart_false_test.__func__( # noqa: E501 False ) diff --git a/keras/datasets/boston_housing.py b/keras/datasets/boston_housing.py index 89a9bb220..22c806f47 100644 --- a/keras/datasets/boston_housing.py +++ b/keras/datasets/boston_housing.py @@ -59,7 +59,7 @@ def load_data(path="boston_housing.npz", test_split=0.2, seed=113): path = get_file( path, origin=origin_folder + "boston_housing.npz", - file_hash="f553886a1f8d56431e820c5b82552d9d95cfcb96d1e678153f8839538947dff5", + file_hash="f553886a1f8d56431e820c5b82552d9d95cfcb96d1e678153f8839538947dff5", # noqa: E501 ) with np.load( path, allow_pickle=True diff --git a/keras/datasets/cifar10.py b/keras/datasets/cifar10.py index 0225ebb84..5e5d1cd7e 100644 --- a/keras/datasets/cifar10.py +++ b/keras/datasets/cifar10.py @@ -80,7 +80,7 @@ def load_data(): dirname, origin=origin, untar=True, - file_hash="6d958be074577803d12ecdefd02955f39262c83c16fe9348329d7fe0b5c001ce", + file_hash="6d958be074577803d12ecdefd02955f39262c83c16fe9348329d7fe0b5c001ce", # noqa: E501 ) num_train_samples = 50000 diff --git a/keras/datasets/cifar100.py b/keras/datasets/cifar100.py index d0e8f9ff1..1eb5039c8 100644 --- a/keras/datasets/cifar100.py +++ b/keras/datasets/cifar100.py @@ -77,7 +77,7 @@ def load_data(label_mode="fine"): dirname, origin=origin, untar=True, - file_hash="85cd44d02ba6437773c5bbd22e183051d648de2e7d6b014e1ef29b855ba677a7", + file_hash="85cd44d02ba6437773c5bbd22e183051d648de2e7d6b014e1ef29b855ba677a7", # noqa: E501 ) fpath = os.path.join(path, "train") diff --git a/keras/datasets/imdb.py b/keras/datasets/imdb.py index 0470666fd..9dae15010 100644 --- a/keras/datasets/imdb.py +++ b/keras/datasets/imdb.py @@ -109,7 +109,7 @@ def load_data( path = get_file( path, origin=origin_folder + "imdb.npz", - file_hash="69664113be75683a8fe16e3ed0ab59fda8886cb3cd7ada244f7d9544e4676b9f", + file_hash="69664113be75683a8fe16e3ed0ab59fda8886cb3cd7ada244f7d9544e4676b9f", # noqa: E501 ) with np.load( path, allow_pickle=True diff --git a/keras/datasets/mnist.py b/keras/datasets/mnist.py index ed981de96..8d22076bd 100644 --- a/keras/datasets/mnist.py +++ b/keras/datasets/mnist.py @@ -73,7 +73,7 @@ def load_data(path="mnist.npz"): path = get_file( path, origin=origin_folder + "mnist.npz", - file_hash="731c5ac602752760c8e48fbffcf8c3b850d9dc2a2aedcf2cc48468fc17b673d1", + file_hash="731c5ac602752760c8e48fbffcf8c3b850d9dc2a2aedcf2cc48468fc17b673d1", # noqa: E501 ) with np.load( path, allow_pickle=True diff --git a/keras/datasets/reuters.py b/keras/datasets/reuters.py index 2be188a36..3e3558361 100644 --- a/keras/datasets/reuters.py +++ b/keras/datasets/reuters.py @@ -115,7 +115,7 @@ def load_data( path = get_file( path, origin=origin_folder + "reuters.npz", - file_hash="d6586e694ee56d7a4e65172e12b3e987c03096cb01eab99753921ef915959916", + file_hash="d6586e694ee56d7a4e65172e12b3e987c03096cb01eab99753921ef915959916", # noqa: E501 ) with np.load( path, allow_pickle=True diff --git a/keras/distribute/checkpointing_test.py b/keras/distribute/checkpointing_test.py index 2c378a620..f1f03dc3f 100644 --- a/keras/distribute/checkpointing_test.py +++ b/keras/distribute/checkpointing_test.py @@ -25,11 +25,11 @@ class TrainingCheckpointTests(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.tpu_strategy, - tf.__internal__.distribute.combinations.tpu_strategy_packed_var, - tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.tpu_strategy, # noqa: E501 + tf.__internal__.distribute.combinations.tpu_strategy_packed_var, # noqa: E501 + tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, # noqa: E501 ], mode=["eager"], ) @@ -87,12 +87,12 @@ class TrainingCheckpointTests(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.cloud_tpu_strategy, - tf.__internal__.distribute.combinations.tpu_strategy, - tf.__internal__.distribute.combinations.tpu_strategy_packed_var, - tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.cloud_tpu_strategy, # noqa: E501 + tf.__internal__.distribute.combinations.tpu_strategy, # noqa: E501 + tf.__internal__.distribute.combinations.tpu_strategy_packed_var, # noqa: E501 + tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, # noqa: E501 ], mode=["eager"], ) diff --git a/keras/distribute/collective_all_reduce_strategy_test.py b/keras/distribute/collective_all_reduce_strategy_test.py index f16c1894c..906272982 100644 --- a/keras/distribute/collective_all_reduce_strategy_test.py +++ b/keras/distribute/collective_all_reduce_strategy_test.py @@ -29,8 +29,8 @@ from keras.testing_infra import test_utils @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( strategy=[ - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, # noqa: E501 ], mode=["eager"], ) diff --git a/keras/distribute/ctl_correctness_test.py b/keras/distribute/ctl_correctness_test.py index af83d2216..10dbc19b8 100644 --- a/keras/distribute/ctl_correctness_test.py +++ b/keras/distribute/ctl_correctness_test.py @@ -271,7 +271,7 @@ class TestDistributionStrategyDnnCorrectness( + tf.__internal__.test.combinations.combine( distribution=[ tf.__internal__.distribute.combinations.one_device_strategy_gpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], optimizer_fn=[ optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, @@ -351,7 +351,7 @@ class TestDistributionStrategyDnnCorrectness( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], mode=["eager"], ) diff --git a/keras/distribute/custom_training_loop_optimizer_test.py b/keras/distribute/custom_training_loop_optimizer_test.py index 8fb790b6e..2b8a90815 100644 --- a/keras/distribute/custom_training_loop_optimizer_test.py +++ b/keras/distribute/custom_training_loop_optimizer_test.py @@ -68,7 +68,7 @@ class OptimizerTest(tf.test.TestCase, parameterized.TestCase): def step_fn(grads): optimizer.apply_gradients( [(grads, v)], - experimental_aggregate_gradients=experimental_aggregate_gradients, + experimental_aggregate_gradients=experimental_aggregate_gradients, # noqa: E501 ) return v.read_value() @@ -80,7 +80,7 @@ class OptimizerTest(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( - distribution=tf.__internal__.distribute.combinations.one_device_strategy, + distribution=tf.__internal__.distribute.combinations.one_device_strategy, # noqa: E501 mode=["eager"], experimental_aggregate_gradients=[True, False], ) @@ -100,7 +100,7 @@ class OptimizerTest(tf.test.TestCase, parameterized.TestCase): def step_fn(grads): optimizer.apply_gradients( [(grads, v)], - experimental_aggregate_gradients=experimental_aggregate_gradients, + experimental_aggregate_gradients=experimental_aggregate_gradients, # noqa: E501 ) return v.read_value() @@ -113,7 +113,7 @@ class OptimizerTest(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu + tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu # noqa: E501 ] ) ) diff --git a/keras/distribute/distribute_strategy_test.py b/keras/distribute/distribute_strategy_test.py index d8ede979d..4ba617eed 100644 --- a/keras/distribute/distribute_strategy_test.py +++ b/keras/distribute/distribute_strategy_test.py @@ -254,8 +254,8 @@ def all_strategy_minus_default_and_tpu_combinations(): distribution=[ tf.__internal__.distribute.combinations.one_device_strategy, tf.__internal__.distribute.combinations.one_device_strategy_gpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], mode=["graph", "eager"], ) @@ -1434,7 +1434,7 @@ class TestDistributionStrategyWithDatasets( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 tf.__internal__.distribute.combinations.one_device_strategy, ], mode=["graph", "eager"], @@ -1467,7 +1467,7 @@ class TestDistributionStrategyWithDatasets( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu # noqa: E501 ], mode=["graph", "eager"], ) @@ -1492,8 +1492,8 @@ class TestDistributionStrategyWithDatasets( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], mode=["graph", "eager"], ) @@ -2309,8 +2309,8 @@ class TestDistributionStrategyWithKerasModels( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], mode=["graph", "eager"], reduction=[ @@ -2476,8 +2476,8 @@ class TestDistributionStrategyWithKerasModels( distribution=[ tf.__internal__.distribute.combinations.one_device_strategy, tf.__internal__.distribute.combinations.one_device_strategy_gpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 ], mode=["eager"], ) @@ -3011,7 +3011,7 @@ class TestModelCapturesStrategy(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( - distribution=tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, + distribution=tf.__internal__.distribute.combinations.mirrored_strategy_with_one_cpu, # noqa: E501 mode=["eager"], ) ) diff --git a/keras/distribute/keras_dnn_correctness_test.py b/keras/distribute/keras_dnn_correctness_test.py index 8dffca153..a08b4c7c9 100644 --- a/keras/distribute/keras_dnn_correctness_test.py +++ b/keras/distribute/keras_dnn_correctness_test.py @@ -115,7 +115,7 @@ class TestDistributionStrategyDnnCorrectness( self.run_correctness_test(distribution, use_numpy, use_validation_data) @tf.__internal__.distribute.combinations.generate( - keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() # noqa: E501 + keras_correctness_test_base.multi_worker_mirrored_eager() ) def test_dnn_correctness_with_partial_last_batch_eval( @@ -129,7 +129,7 @@ class TestDistributionStrategyDnnCorrectness( ) @tf.__internal__.distribute.combinations.generate( - keras_correctness_test_base.strategy_minus_tpu_and_input_config_combinations_eager() + keras_correctness_test_base.strategy_minus_tpu_and_input_config_combinations_eager() # noqa: E501 + keras_correctness_test_base.multi_worker_mirrored_eager() ) def test_dnn_correctness_with_partial_last_batch( @@ -354,7 +354,7 @@ class TestDistributionStrategyDnnCorrectnessWithSubclassedModel( self.run_dynamic_lr_test(distribution) @tf.__internal__.distribute.combinations.generate( - keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() # noqa: E501 ) def test_dnn_correctness_with_partial_last_batch_eval( self, distribution, use_numpy, use_validation_data diff --git a/keras/distribute/keras_embedding_model_correctness_test.py b/keras/distribute/keras_embedding_model_correctness_test.py index a6d3cf368..06e7ee7c4 100644 --- a/keras/distribute/keras_embedding_model_correctness_test.py +++ b/keras/distribute/keras_embedding_model_correctness_test.py @@ -25,7 +25,7 @@ from keras.optimizers.optimizer_v2 import ( class DistributionStrategyEmbeddingModelCorrectnessTest( - keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase + keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase # noqa: E501 ): def get_model( self, @@ -83,7 +83,7 @@ class DistributionStrategyEmbeddingModelCorrectnessTest( class DistributionStrategySiameseEmbeddingModelCorrectnessTest( - keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase + keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase # noqa: E501 ): def get_model( self, diff --git a/keras/distribute/keras_image_model_correctness_test.py b/keras/distribute/keras_image_model_correctness_test.py index e8f265c41..bd096490f 100644 --- a/keras/distribute/keras_image_model_correctness_test.py +++ b/keras/distribute/keras_image_model_correctness_test.py @@ -106,7 +106,7 @@ class DistributionStrategyCnnCorrectnessTest( ): if ( distribution - == tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu + == tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu # noqa: E501 ): self.skipTest("b/183958183") self.run_correctness_test(distribution, use_numpy, use_validation_data) @@ -140,9 +140,9 @@ class DistributionStrategyCnnCorrectnessTest( ) @tf.__internal__.distribute.combinations.generate( - keras_correctness_test_base.all_strategy_and_input_config_combinations_eager() + keras_correctness_test_base.all_strategy_and_input_config_combinations_eager() # noqa: E501 + keras_correctness_test_base.multi_worker_mirrored_eager() - + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() + + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() # noqa: E501 ) def test_cnn_correctness_with_partial_last_batch_eval( self, distribution, use_numpy, use_validation_data @@ -156,9 +156,9 @@ class DistributionStrategyCnnCorrectnessTest( ) @tf.__internal__.distribute.combinations.generate( - keras_correctness_test_base.all_strategy_and_input_config_combinations_eager() + keras_correctness_test_base.all_strategy_and_input_config_combinations_eager() # noqa: E501 + keras_correctness_test_base.multi_worker_mirrored_eager() - + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() + + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() # noqa: E501 ) def test_cnn_with_batch_norm_correctness_and_partial_last_batch_eval( self, distribution, use_numpy, use_validation_data diff --git a/keras/distribute/keras_optimizer_v2_test.py b/keras/distribute/keras_optimizer_v2_test.py index afd0de071..2f28519fa 100644 --- a/keras/distribute/keras_optimizer_v2_test.py +++ b/keras/distribute/keras_optimizer_v2_test.py @@ -34,7 +34,7 @@ class MirroredStrategyOptimizerV2Test(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, # noqa: E501 ], mode=["graph", "eager"], ) @@ -96,7 +96,7 @@ class MirroredStrategyOptimizerV2Test(tf.test.TestCase, parameterized.TestCase): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, # noqa: E501 ], mode=["graph", "eager"], ) diff --git a/keras/distribute/keras_premade_models_test.py b/keras/distribute/keras_premade_models_test.py index 238f16602..8768fb372 100644 --- a/keras/distribute/keras_premade_models_test.py +++ b/keras/distribute/keras_premade_models_test.py @@ -33,14 +33,14 @@ def strategy_combinations_eager_data_fn(): tf.__internal__.distribute.combinations.default_strategy, tf.__internal__.distribute.combinations.one_device_strategy, tf.__internal__.distribute.combinations.one_device_strategy_gpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x2_gpu, - tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_cpu, - tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_1gpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, # noqa: E501 + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, # noqa: E501 + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x2_gpu, # noqa: E501 + tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_1gpu, # noqa: E501 # NOTE: TPUStrategy not tested because the models in this test are # sparse and do not work with TPUs. ], diff --git a/keras/distribute/keras_rnn_model_correctness_test.py b/keras/distribute/keras_rnn_model_correctness_test.py index 14fe31c2e..0db1c58e1 100644 --- a/keras/distribute/keras_rnn_model_correctness_test.py +++ b/keras/distribute/keras_rnn_model_correctness_test.py @@ -31,7 +31,7 @@ from keras.testing_infra import test_utils class _DistributionStrategyRnnModelCorrectnessTest( - keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase + keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase # noqa: E501 ): def _get_layer_class(self): raise NotImplementedError diff --git a/keras/distribute/keras_stateful_lstm_model_correctness_test.py b/keras/distribute/keras_stateful_lstm_model_correctness_test.py index e7ad3057d..7896a468d 100644 --- a/keras/distribute/keras_stateful_lstm_model_correctness_test.py +++ b/keras/distribute/keras_stateful_lstm_model_correctness_test.py @@ -42,7 +42,7 @@ def test_combinations_for_stateful_embedding_model(): class DistributionStrategyStatefulLstmModelCorrectnessTest( - keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase + keras_correctness_test_base.TestDistributionStrategyEmbeddingModelCorrectnessBase # noqa: E501 ): def get_model( self, @@ -97,7 +97,7 @@ class DistributionStrategyStatefulLstmModelCorrectnessTest( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.times( - keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() + keras_correctness_test_base.test_combinations_with_tpu_strategies_graph() # noqa: E501 ) ) def test_incorrectly_use_multiple_cores_for_stateful_lstm_model( diff --git a/keras/distribute/keras_utils_test.py b/keras/distribute/keras_utils_test.py index 659c5201f..8925801ea 100644 --- a/keras/distribute/keras_utils_test.py +++ b/keras/distribute/keras_utils_test.py @@ -197,7 +197,7 @@ class TestDistributionStrategyErrorCases( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph"], ) @@ -227,14 +227,14 @@ class TestDistributionStrategyErrorCases( "PerReplica:.+", ): with distribution.scope(): - distributed_training_utils_v1.validate_distributed_dataset_inputs( + distributed_training_utils_v1.validate_distributed_dataset_inputs( # noqa: E501 distribution, x, None ) @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph", "eager"], ) @@ -264,14 +264,14 @@ class TestDistributionStrategyErrorCases( "PerReplica:.+", ): with distribution.scope(): - distributed_training_utils_v1.validate_distributed_dataset_inputs( + distributed_training_utils_v1.validate_distributed_dataset_inputs( # noqa: E501 distribution, x, None ) @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph", "eager"], ) @@ -322,7 +322,7 @@ class TestDistributionStrategyErrorCases( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 tf.__internal__.distribute.combinations.one_device_strategy, ], mode=["graph", "eager"], @@ -355,7 +355,7 @@ class TestDistributionStrategyErrorCases( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 tf.__internal__.distribute.combinations.one_device_strategy, ], mode=["graph", "eager"], @@ -406,10 +406,10 @@ class TestDistributionStrategyWithLossMasking( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph", "eager"], - optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, + optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, # noqa: E501 ) ) def test_masking(self, distribution, optimizer): @@ -443,7 +443,7 @@ class TestDistributionStrategyWithNormalizationLayer( keras_test_lib.all_strategy_combinations(), tf.__internal__.test.combinations.combine( fused=[True, False], - optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, + optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, # noqa: E501 ), ) ) @@ -489,7 +489,7 @@ class TestDistributionStrategyWithNormalizationLayer( tf.__internal__.test.combinations.times( keras_test_lib.tpu_strategy_combinations(), tf.__internal__.test.combinations.combine( - optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn + optimizer=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn # noqa: E501 ), ) ) @@ -653,7 +653,7 @@ class TestDistributionStrategyWithStaticShapes( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph", "eager"], ) @@ -670,7 +670,7 @@ class TestDistributionStrategyWithStaticShapes( @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["graph", "eager"], ) diff --git a/keras/distribute/minimize_loss_test.py b/keras/distribute/minimize_loss_test.py index 7f6ee3538..3f5b087a1 100644 --- a/keras/distribute/minimize_loss_test.py +++ b/keras/distribute/minimize_loss_test.py @@ -388,15 +388,15 @@ class MinimizeLossStepTest(tf.test.TestCase, parameterized.TestCase): tf.__internal__.test.combinations.times( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.one_device_strategy, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, + tf.__internal__.distribute.combinations.one_device_strategy, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, # noqa: E501 ] ), tf.__internal__.test.combinations.times( tf.__internal__.test.combinations.combine( - optimizer_fn=optimizer_combinations.gradient_descent_optimizer_v1_fn + optimizer_fn=optimizer_combinations.gradient_descent_optimizer_v1_fn # noqa: E501 ), tf.__internal__.test.combinations.combine( mode=["graph"], use_callable_loss=[True, False] @@ -407,7 +407,7 @@ class MinimizeLossStepTest(tf.test.TestCase, parameterized.TestCase): ) + tf.__internal__.test.combinations.times( tf.__internal__.test.combinations.combine( - optimizer_fn=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn + optimizer_fn=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn # noqa: E501 ), tf.__internal__.test.combinations.combine( mode=["graph", "eager"], use_callable_loss=[True] @@ -418,7 +418,7 @@ class MinimizeLossStepTest(tf.test.TestCase, parameterized.TestCase): distribution=[ tf.__internal__.distribute.combinations.tpu_strategy ], - optimizer_fn=optimizer_combinations.gradient_descent_optimizer_v1_fn, + optimizer_fn=optimizer_combinations.gradient_descent_optimizer_v1_fn, # noqa: E501 mode=["graph"], use_callable_loss=[True, False], ) @@ -426,7 +426,7 @@ class MinimizeLossStepTest(tf.test.TestCase, parameterized.TestCase): distribution=[ tf.__internal__.distribute.combinations.tpu_strategy ], - optimizer_fn=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, + optimizer_fn=optimizer_combinations.gradient_descent_optimizer_keras_v2_fn, # noqa: E501 mode=["graph"], use_callable_loss=[True], ), diff --git a/keras/distribute/mirrored_strategy_test.py b/keras/distribute/mirrored_strategy_test.py index 39b61f592..f4476c309 100644 --- a/keras/distribute/mirrored_strategy_test.py +++ b/keras/distribute/mirrored_strategy_test.py @@ -49,7 +49,7 @@ class MiniModel(keras_training.Model): @tf.__internal__.distribute.combinations.generate( tf.__internal__.test.combinations.combine( distribution=[ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 ], mode=["eager"], ) diff --git a/keras/distribute/mirrored_variable_test.py b/keras/distribute/mirrored_variable_test.py index 192f18b06..e6a198f8b 100644 --- a/keras/distribute/mirrored_variable_test.py +++ b/keras/distribute/mirrored_variable_test.py @@ -51,7 +51,7 @@ def get_strategy_with_mimicing_cpus(): filter( None.__ne__, [ - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 get_strategy_with_mimicing_cpus(), ], ) diff --git a/keras/distribute/multi_worker_callback_tf2_test.py b/keras/distribute/multi_worker_callback_tf2_test.py index 21ec37b5a..d107d9b5b 100644 --- a/keras/distribute/multi_worker_callback_tf2_test.py +++ b/keras/distribute/multi_worker_callback_tf2_test.py @@ -159,7 +159,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_model_checkpoint_saves_on_chief_but_not_otherwise, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self, file_format), @@ -192,7 +192,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_model_checkpoint_works_with_same_file_path, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self, saving_filepath), @@ -263,7 +263,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_model_checkpoint_works_with_same_file_path, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self, saving_filepath), @@ -306,7 +306,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_profiler_saves_on_both_chief_and_non_chief, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self,), @@ -357,7 +357,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_tensorboard_saves_on_chief_but_not_otherwise, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self,), @@ -395,7 +395,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_tensorboard_can_still_save_to_temp_even_if_it_exists, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self,), @@ -432,7 +432,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_tensorboard_works_with_same_file_path, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self, saving_filepath), @@ -466,7 +466,7 @@ class KerasCallbackMultiProcessTest(parameterized.TestCase, tf.test.TestCase): tf.__internal__.distribute.multi_process_runner.run( proc_early_stopping, - cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + cluster_spec=tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=2 ), args=(self,), diff --git a/keras/distribute/multi_worker_test.py b/keras/distribute/multi_worker_test.py index 571294757..8bdd6782e 100644 --- a/keras/distribute/multi_worker_test.py +++ b/keras/distribute/multi_worker_test.py @@ -194,8 +194,8 @@ class KerasMultiWorkerTestIndependentWorker( tf.__internal__.test.combinations.combine( mode=["eager"], strategy=[ - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, # noqa: E501 ], ) ) @@ -236,7 +236,7 @@ class KPLMultiWorkerTest(tf.test.TestCase, parameterized.TestCase): mode=["eager"], use_adapt=[False], # TODO(b/180742437): Add tests for using adapt. strategy=[ - tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, + tf.__internal__.distribute.combinations.multi_worker_mirrored_2x1_gpu, # noqa: E501 # TODO(b/183956672): Re-enable # strategy_combinations.multi_worker_mirrored_2x2_gpu, ], diff --git a/keras/distribute/optimizer_combinations.py b/keras/distribute/optimizer_combinations.py index 30005886a..8f8390448 100644 --- a/keras/distribute/optimizer_combinations.py +++ b/keras/distribute/optimizer_combinations.py @@ -100,9 +100,9 @@ def distributions_and_v1_optimizers(): return tf.__internal__.test.combinations.combine( distribution=[ tf.__internal__.distribute.combinations.one_device_strategy, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, # noqa: E501 ], optimizer_fn=optimizers_v1, ) @@ -114,9 +114,9 @@ def distributions_and_v2_optimizers(): return tf.__internal__.test.combinations.combine( distribution=[ tf.__internal__.distribute.combinations.one_device_strategy, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, # noqa: E501 ], optimizer_fn=optimizers_v2, ) @@ -128,9 +128,9 @@ def distributions_and_v1_and_v2_optimizers(): return tf.__internal__.test.combinations.combine( distribution=[ tf.__internal__.distribute.combinations.one_device_strategy, - tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, + tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, # noqa: E501 + tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus_no_merge_call, # noqa: E501 ], optimizer_fn=optimizers_v1_and_v2, ) diff --git a/keras/distribute/saved_model_test_base.py b/keras/distribute/saved_model_test_base.py index c61ca361a..09e8e5aff 100644 --- a/keras/distribute/saved_model_test_base.py +++ b/keras/distribute/saved_model_test_base.py @@ -49,7 +49,7 @@ strategies = [ tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, tf.__internal__.distribute.combinations.tpu_strategy, tf.__internal__.distribute.combinations.tpu_strategy_packed_var, - tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, + tf.__internal__.distribute.combinations.central_storage_strategy_with_two_gpus, # noqa: E501 ] diff --git a/keras/distribute/sharded_variable_test.py b/keras/distribute/sharded_variable_test.py index 11d29b8b1..bcd1250c1 100644 --- a/keras/distribute/sharded_variable_test.py +++ b/keras/distribute/sharded_variable_test.py @@ -30,7 +30,7 @@ class ShardedVariableTest(tf.test.TestCase, parameterized.TestCase): super().setUpClass() cls.strategy = tf.distribute.experimental.ParameterServerStrategy( multi_worker_testing_utils.make_parameter_server_cluster(3, 2), - variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( + variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( # noqa: E501 2 ), ) @@ -184,7 +184,7 @@ class ShardedVariableTest(tf.test.TestCase, parameterized.TestCase): if shard_config[0] > 2: strategy = tf.distribute.experimental.ParameterServerStrategy( multi_worker_testing_utils.make_parameter_server_cluster(3, 3), - variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( + variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( # noqa: E501 shard_config[0] ), ) @@ -217,7 +217,7 @@ class ShardedVariableTest(tf.test.TestCase, parameterized.TestCase): if shard_config[1] > 2: strategy2 = tf.distribute.experimental.ParameterServerStrategy( multi_worker_testing_utils.make_parameter_server_cluster(3, 3), - variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( + variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( # noqa: E501 shard_config[1] ), ) @@ -384,7 +384,7 @@ class ShardedVariableTest(tf.test.TestCase, parameterized.TestCase): # Create new strategy with different number of shards strategy2 = tf.distribute.experimental.ParameterServerStrategy( multi_worker_testing_utils.make_parameter_server_cluster(3, 2), - variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( + variable_partitioner=tf.distribute.experimental.partitioners.FixedShardsPartitioner( # noqa: E501 3 ), ) diff --git a/keras/distribute/strategy_combinations.py b/keras/distribute/strategy_combinations.py index ea7c0016a..8261e2386 100644 --- a/keras/distribute/strategy_combinations.py +++ b/keras/distribute/strategy_combinations.py @@ -33,7 +33,7 @@ strategies_minus_default_minus_tpu = [ tf.__internal__.distribute.combinations.one_device_strategy_gpu, tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu, # noqa: E501 ] strategies_minus_tpu = [ @@ -42,7 +42,7 @@ strategies_minus_tpu = [ tf.__internal__.distribute.combinations.one_device_strategy_gpu, tf.__internal__.distribute.combinations.mirrored_strategy_with_gpu_and_cpu, tf.__internal__.distribute.combinations.mirrored_strategy_with_two_gpus, - tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu, + tf.__internal__.distribute.combinations.central_storage_strategy_with_gpu_and_cpu, # noqa: E501 ] multi_worker_mirrored_strategies = [ @@ -56,13 +56,13 @@ tpu_strategies = [ ] parameter_server_strategies_single_worker = [ - tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_cpu, - tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_1gpu, + tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.parameter_server_strategy_1worker_2ps_1gpu, # noqa: E501 ] parameter_server_strategies_multi_worker = [ - tf.__internal__.distribute.combinations.parameter_server_strategy_3worker_2ps_cpu, - tf.__internal__.distribute.combinations.parameter_server_strategy_3worker_2ps_1gpu, + tf.__internal__.distribute.combinations.parameter_server_strategy_3worker_2ps_cpu, # noqa: E501 + tf.__internal__.distribute.combinations.parameter_server_strategy_3worker_2ps_1gpu, # noqa: E501 ] all_strategies = strategies_minus_tpu + tpu_strategies diff --git a/keras/dtensor/optimizers.py b/keras/dtensor/optimizers.py index 79c5cb9de..a6af0f7c8 100644 --- a/keras/dtensor/optimizers.py +++ b/keras/dtensor/optimizers.py @@ -153,7 +153,7 @@ class Optimizer(optimizer_lib._BaseOptimizer): def _overwrite_model_variables_with_average_value_helper(self, var_list): """Helper function to _overwrite_model_variables_with_average_value.""" ( - optimizer_lib._BaseOptimizer._overwrite_model_variables_with_average_value_helper( + optimizer_lib._BaseOptimizer._overwrite_model_variables_with_average_value_helper( # noqa: E501 self, var_list ) ) diff --git a/keras/engine/training.py b/keras/engine/training.py index bca7fafaa..b627d0eaf 100644 --- a/keras/engine/training.py +++ b/keras/engine/training.py @@ -1498,7 +1498,7 @@ class Model(base_layer.Layer, version_utils.ModelVersionSelector): ) ) - with self.distribute_strategy.scope(), training_utils.RespectCompiledTrainableState( + with self.distribute_strategy.scope(), training_utils.RespectCompiledTrainableState( # noqa: E501 self ): # Creates a `tf.data.Dataset` and handles batch and epoch iteration. @@ -2377,7 +2377,7 @@ class Model(base_layer.Layer, version_utils.ModelVersionSelector): _disallow_inside_tf_function("train_on_batch") if reset_metrics: self.reset_metrics() - with self.distribute_strategy.scope(), training_utils.RespectCompiledTrainableState( + with self.distribute_strategy.scope(), training_utils.RespectCompiledTrainableState( # noqa: E501 self ): iterator = data_adapter.single_batch_iterator( diff --git a/keras/engine/training_arrays_v1.py b/keras/engine/training_arrays_v1.py index 298714c9c..f44bdc483 100644 --- a/keras/engine/training_arrays_v1.py +++ b/keras/engine/training_arrays_v1.py @@ -306,7 +306,7 @@ def model_iteration( # case. if not callable(ins) or ( model._distribution_strategy - and not distributed_training_utils_v1.is_distributing_by_cloning( + and not distributed_training_utils_v1.is_distributing_by_cloning( # noqa: E501 model ) ): @@ -353,7 +353,7 @@ def model_iteration( batch_outs = [batch_outs] if model._distribution_strategy: - batch_outs = distributed_training_utils_v1._per_replica_aggregate_batch( + batch_outs = distributed_training_utils_v1._per_replica_aggregate_batch( # noqa: E501 model._distribution_strategy, batch_outs, model, mode ) diff --git a/keras/engine/training_dataset_test.py b/keras/engine/training_dataset_test.py index 4aab91231..500c48d58 100644 --- a/keras/engine/training_dataset_test.py +++ b/keras/engine/training_dataset_test.py @@ -346,7 +346,7 @@ class TestTrainingWithDataset(test_combinations.TestCase): ) def test_dataset_input_shape_validation(self): - with tf.compat.v1.get_default_graph().as_default(), self.cached_session(): + with tf.compat.v1.get_default_graph().as_default(), self.cached_session(): # noqa: E501 model = test_utils.get_small_functional_mlp(1, 4, input_dim=3) model.compile(optimizer="rmsprop", loss="mse") diff --git a/keras/engine/training_gpu_test.py b/keras/engine/training_gpu_test.py index 1e99035fc..602b871e3 100644 --- a/keras/engine/training_gpu_test.py +++ b/keras/engine/training_gpu_test.py @@ -45,7 +45,7 @@ class TrainingGPUTest(tf.test.TestCase, parameterized.TestCase): num_channels = None activation = None if loss_name == "sparse_categorical_crossentropy": - loss = lambda y_true, y_pred: backend.sparse_categorical_crossentropy( + loss = lambda y_true, y_pred: backend.sparse_categorical_crossentropy( # noqa: E501 y_true, y_pred, axis=axis ) num_channels = int(np.amax(target) + 1) diff --git a/keras/engine/training_v1.py b/keras/engine/training_v1.py index 37e23962a..918a8829e 100644 --- a/keras/engine/training_v1.py +++ b/keras/engine/training_v1.py @@ -644,12 +644,12 @@ class Model(training_lib.Model): # Case 1: distribution strategy. if self._distribution_strategy: if self._in_multi_worker_mode(): - return training_distributed_v1.DistributionMultiWorkerTrainingLoop( - training_distributed_v1.DistributionSingleWorkerTrainingLoop() + return training_distributed_v1.DistributionMultiWorkerTrainingLoop( # noqa: E501 + training_distributed_v1.DistributionSingleWorkerTrainingLoop() # noqa: E501 ) else: return ( - training_distributed_v1.DistributionSingleWorkerTrainingLoop() + training_distributed_v1.DistributionSingleWorkerTrainingLoop() # noqa: E501 ) # Case 2: generator-like. Input is Python generator, or Sequence object, diff --git a/keras/feature_column/sequence_feature_column.py b/keras/feature_column/sequence_feature_column.py index e96dd037b..cc9b96959 100644 --- a/keras/feature_column/sequence_feature_column.py +++ b/keras/feature_column/sequence_feature_column.py @@ -101,7 +101,7 @@ class SequenceFeatures(kfc._BaseFeaturesLayer): feature_columns=feature_columns, trainable=trainable, name=name, - expected_column_type=tf.__internal__.feature_column.SequenceDenseColumn, + expected_column_type=tf.__internal__.feature_column.SequenceDenseColumn, # noqa: E501 **kwargs ) diff --git a/keras/feature_column/sequence_feature_column_test.py b/keras/feature_column/sequence_feature_column_test.py index 80d441138..0cc46cc7d 100644 --- a/keras/feature_column/sequence_feature_column_test.py +++ b/keras/feature_column/sequence_feature_column_test.py @@ -926,7 +926,7 @@ class SequenceFeaturesSavingTest(tf.test.TestCase, parameterized.TestCase): cols = [ tf.feature_column.sequence_numeric_column("a"), tf.feature_column.indicator_column( - tf.feature_column.sequence_categorical_column_with_vocabulary_list( + tf.feature_column.sequence_categorical_column_with_vocabulary_list( # noqa: E501 "b", ["one", "two"] ) ), diff --git a/keras/integration_test/multi_worker_tutorial_test.py b/keras/integration_test/multi_worker_tutorial_test.py index 9134d12b2..89df14576 100644 --- a/keras/integration_test/multi_worker_tutorial_test.py +++ b/keras/integration_test/multi_worker_tutorial_test.py @@ -242,7 +242,7 @@ class MultiWorkerTutorialTest(parameterized.TestCase, tf.test.TestCase): try: mpr_result = tf.__internal__.distribute.multi_process_runner.run( fn, - tf.__internal__.distribute.multi_process_runner.create_cluster_spec( + tf.__internal__.distribute.multi_process_runner.create_cluster_spec( # noqa: E501 num_workers=NUM_WORKERS ), args=(model_path, checkpoint_dir), diff --git a/keras/layers/core/core_test.py b/keras/layers/core/core_test.py index 6cecb3581..44f4d866f 100644 --- a/keras/layers/core/core_test.py +++ b/keras/layers/core/core_test.py @@ -352,11 +352,12 @@ class TestStatefulLambda(test_combinations.TestCase): expected_error = textwrap.dedent( r""" - ( )?The following Variables were created within a Lambda layer \(shift_and_scale\) - ( )?but are not tracked by said layer: - ( )?