From d9c374c69b71b73f453ef213d100fe299eec2cf9 Mon Sep 17 00:00:00 2001 From: Mohammed Yasin <32206511+Y-T-G@users.noreply.github.com> Date: Fri, 17 Jan 2025 05:45:49 +0800 Subject: [PATCH] Fix automatic optimizer LR with DDP training (#18641) Co-authored-by: Glenn Jocher Co-authored-by: Ultralytics Assistant <135830346+UltralyticsAssistant@users.noreply.github.com> --- ultralytics/engine/trainer.py | 1 + 1 file changed, 1 insertion(+) diff --git a/ultralytics/engine/trainer.py b/ultralytics/engine/trainer.py index 61c7d4ed..a373cd82 100644 --- a/ultralytics/engine/trainer.py +++ b/ultralytics/engine/trainer.py @@ -271,6 +271,7 @@ class BaseTrainer: ) if world_size > 1: self.model = nn.parallel.DistributedDataParallel(self.model, device_ids=[RANK], find_unused_parameters=True) + self.set_model_attributes() # set again after DDP wrapper # Check imgsz gs = max(int(self.model.stride.max() if hasattr(self.model, "stride") else 32), 32) # grid size (max stride)