From 463ca1a804adbb722178cb7879ece73674101b1b Mon Sep 17 00:00:00 2001 From: Jason Guo <131695330+JasonG98@users.noreply.github.com> Date: Thu, 12 Sep 2024 00:37:40 +0800 Subject: [PATCH] Update TwoWayTransformer Docs. (#16161) Co-authored-by: Glenn Jocher Co-authored-by: UltralyticsAssistant Co-authored-by: Ultralytics Assistant <135830346+UltralyticsAssistant@users.noreply.github.com> --- ultralytics/models/sam/modules/transformer.py | 1 - 1 file changed, 1 deletion(-) diff --git a/ultralytics/models/sam/modules/transformer.py b/ultralytics/models/sam/modules/transformer.py index df0b3ce2..5cc0b02f 100644 --- a/ultralytics/models/sam/modules/transformer.py +++ b/ultralytics/models/sam/modules/transformer.py @@ -61,7 +61,6 @@ class TwoWayTransformer(nn.Module): Attributes: depth (int): Number of layers in the transformer. embedding_dim (int): Channel dimension for input embeddings. - embedding_dim (int): Channel dimension for input embeddings. num_heads (int): Number of heads for multihead attention. mlp_dim (int): Internal channel dimension for the MLP block. layers (nn.ModuleList): List of TwoWayAttentionBlock layers.