Skip to content

Commit

Permalink
Fix pyre errors
Browse files Browse the repository at this point in the history
  • Loading branch information
jesszzzz committed Dec 6, 2024
1 parent 648913c commit e390161
Show file tree
Hide file tree
Showing 8 changed files with 18 additions and 12 deletions.
5 changes: 5 additions & 0 deletions .github/workflows/pyre.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,3 +25,8 @@ jobs:
lintrunner init
- name: Run Pyre
run: scripts/pyre.sh
continue-on-error: true
- name: log
run: cat /home/runner/work/torchx/torchx/.pyre/*
- name: Rage
run: pyre rage
4 changes: 2 additions & 2 deletions scripts/pyre.sh
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.

set -eux

pyre --version
pyre --noninteractive check
pyre --noninteractive --show-error-traces check
pyre rage
9 changes: 3 additions & 6 deletions torchx/examples/apps/lightning/data.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,18 +64,15 @@ def __len__(self) -> int:
# our trainer and other components that need to load data.


# pyre-fixme[13]: Attribute `test_ds` is never initialized.
# pyre-fixme[13]: Attribute `train_ds` is never initialized.
# pyre-fixme[13]: Attribute `val_ds` is never initialized.
class TinyImageNetDataModule(pl.LightningDataModule):
"""
TinyImageNetDataModule is a pytorch LightningDataModule for the tiny
imagenet dataset.
"""

train_ds: ImageFolderSamplesDataset
val_ds: ImageFolderSamplesDataset
test_ds: ImageFolderSamplesDataset
train_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `train_ds` is never initialized.
val_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `val_ds` is never initialized.
test_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `test_ds` is never initialized.

def __init__(
self, data_dir: str, batch_size: int = 16, num_samples: Optional[int] = None
Expand Down
1 change: 1 addition & 0 deletions torchx/examples/apps/tracker/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -99,6 +99,7 @@ def test(
for data, target in test_loader:
data, target = data.to(device), target.to(device)
output = model(data)
# pyre-fixme[58] Assuming F.nll_loss(...).item() is a number
test_loss += F.nll_loss(
output, target, reduction="sum"
).item() # sum up batch loss
Expand Down
2 changes: 1 addition & 1 deletion torchx/pipelines/kfp/adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@ def component_spec_from_app(app: api.AppDef) -> Tuple[str, api.Role]:
role = app.roles[0]
assert (
role.num_replicas == 1
), f"KFP adapter only supports one replica, got {app.num_replicas}"
), f"KFP adapter only supports one replica, got {app.num_replicas}" # pyre-fixme[16] Assume num_replicas is available on app

command = [role.entrypoint, *role.args]

Expand Down
1 change: 1 addition & 0 deletions torchx/schedulers/aws_batch_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -809,6 +809,7 @@ def _stream_events(
startFromHead=True,
**args,
)
# pyre-fixme[66] Assume this ResourceNotFoundException extends BaseException
except self._log_client.exceptions.ResourceNotFoundException:
return [] # noqa: B901
if response["nextForwardToken"] == next_token:
Expand Down
4 changes: 2 additions & 2 deletions torchx/schedulers/aws_sagemaker_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -261,9 +261,9 @@ def _submit_dryrun(
raise ValueError(
f"{key} is controlled by aws_sagemaker_scheduler and is set to {job_def[key]}"
)
value = cfg.get(key) # pyre-ignore[26]
value = cfg.get(key)
if value is not None:
job_def[key] = value
job_def[key] = value # pyre-ignore[6]

req = AWSSageMakerJob(
job_name=job_name,
Expand Down
4 changes: 3 additions & 1 deletion torchx/schedulers/ray/ray_driver.py
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,9 @@ def load_actor_json(filename: str) -> List[RayActor]:
return actors


def create_placement_group_async(replicas: List[RayActor]) -> PlacementGroup:
def create_placement_group_async(
replicas: List[RayActor],
) -> PlacementGroup: # pyre-ignore[11]
"""return a placement group reference, the corresponding placement group could be scheduled or pending"""
bundles = []
for replica in replicas:
Expand Down

0 comments on commit e390161

Please sign in to comment.