Apply Ruff 0.9.0 (#18622)
Signed-off-by: Glenn Jocher <glenn.jocher@ultralytics.com> Co-authored-by: UltralyticsAssistant <web@ultralytics.com>
This commit is contained in:
parent
cc1e77138c
commit
3902e740cf
22 changed files with 69 additions and 65 deletions
|
|
@ -479,9 +479,9 @@ class ImageEncoder(nn.Module):
|
|||
self.trunk = trunk
|
||||
self.neck = neck
|
||||
self.scalp = scalp
|
||||
assert (
|
||||
self.trunk.channel_list == self.neck.backbone_channel_list
|
||||
), f"Channel dims of trunk {self.trunk.channel_list} and neck {self.neck.backbone_channel_list} do not match."
|
||||
assert self.trunk.channel_list == self.neck.backbone_channel_list, (
|
||||
f"Channel dims of trunk {self.trunk.channel_list} and neck {self.neck.backbone_channel_list} do not match."
|
||||
)
|
||||
|
||||
def forward(self, sample: torch.Tensor):
|
||||
"""Encodes input through patch embedding, positional embedding, transformer blocks, and neck module."""
|
||||
|
|
|
|||
|
|
@ -279,9 +279,9 @@ class Predictor(BasePredictor):
|
|||
if labels is None:
|
||||
labels = np.ones(points.shape[:-1])
|
||||
labels = torch.as_tensor(labels, dtype=torch.int32, device=self.device)
|
||||
assert (
|
||||
points.shape[-2] == labels.shape[-1]
|
||||
), f"Number of points {points.shape[-2]} should match number of labels {labels.shape[-1]}."
|
||||
assert points.shape[-2] == labels.shape[-1], (
|
||||
f"Number of points {points.shape[-2]} should match number of labels {labels.shape[-1]}."
|
||||
)
|
||||
points *= r
|
||||
if points.ndim == 2:
|
||||
# (N, 2) --> (N, 1, 2), (N, ) --> (N, 1)
|
||||
|
|
@ -552,9 +552,9 @@ class Predictor(BasePredictor):
|
|||
|
||||
def get_im_features(self, im):
|
||||
"""Extracts image features using the SAM model's image encoder for subsequent mask prediction."""
|
||||
assert (
|
||||
isinstance(self.imgsz, (tuple, list)) and self.imgsz[0] == self.imgsz[1]
|
||||
), f"SAM models only support square image size, but got {self.imgsz}."
|
||||
assert isinstance(self.imgsz, (tuple, list)) and self.imgsz[0] == self.imgsz[1], (
|
||||
f"SAM models only support square image size, but got {self.imgsz}."
|
||||
)
|
||||
self.model.set_imgsz(self.imgsz)
|
||||
return self.model.image_encoder(im)
|
||||
|
||||
|
|
@ -795,9 +795,9 @@ class SAM2Predictor(Predictor):
|
|||
|
||||
def get_im_features(self, im):
|
||||
"""Extracts image features from the SAM image encoder for subsequent processing."""
|
||||
assert (
|
||||
isinstance(self.imgsz, (tuple, list)) and self.imgsz[0] == self.imgsz[1]
|
||||
), f"SAM 2 models only support square image size, but got {self.imgsz}."
|
||||
assert isinstance(self.imgsz, (tuple, list)) and self.imgsz[0] == self.imgsz[1], (
|
||||
f"SAM 2 models only support square image size, but got {self.imgsz}."
|
||||
)
|
||||
self.model.set_imgsz(self.imgsz)
|
||||
self._bb_feat_sizes = [[x // (4 * i) for x in self.imgsz] for i in [1, 2, 4]]
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue