diff --git a/tests/integration/gradient_accumulation_test.py b/tests/integration/gradient_accumulation_test.py index 468c7aced8..5b464147c5 100644 --- a/tests/integration/gradient_accumulation_test.py +++ b/tests/integration/gradient_accumulation_test.py @@ -62,6 +62,7 @@ def test_grad_accumulate_same_loss(self): get_test_config_path(), f"base_output_directory={self.base_output_directory}", f"dataset_path={self.dataset_path}", + "dataset_type=synthetic", "gradient_clipping_threshold=0", # Ensures we are testing raw scales of gradients (clipping off) "enable_checkpointing=False", "enable_goodput_recording=False",