Skip to content

Fix pyre errors #981

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 5 additions & 0 deletions .github/workflows/pyre.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,3 +25,8 @@ jobs:
lintrunner init
- name: Run Pyre
run: scripts/pyre.sh
continue-on-error: true
- name: log
run: cat /home/runner/work/torchx/torchx/.pyre/*
- name: Rage
run: pyre rage
4 changes: 2 additions & 2 deletions scripts/pyre.sh
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@
# This source code is licensed under the BSD-style license found in the
# LICENSE file in the root directory of this source tree.

set -eux

pyre --version
pyre --noninteractive check
pyre --noninteractive --show-error-traces check
pyre rage
9 changes: 3 additions & 6 deletions torchx/examples/apps/lightning/data.py
Original file line number Diff line number Diff line change
Expand Up @@ -64,18 +64,15 @@ def __len__(self) -> int:
# our trainer and other components that need to load data.


# pyre-fixme[13]: Attribute `test_ds` is never initialized.
# pyre-fixme[13]: Attribute `train_ds` is never initialized.
# pyre-fixme[13]: Attribute `val_ds` is never initialized.
class TinyImageNetDataModule(pl.LightningDataModule):
"""
TinyImageNetDataModule is a pytorch LightningDataModule for the tiny
imagenet dataset.
"""

train_ds: ImageFolderSamplesDataset
val_ds: ImageFolderSamplesDataset
test_ds: ImageFolderSamplesDataset
train_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `train_ds` is never initialized.
val_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `val_ds` is never initialized.
test_ds: ImageFolderSamplesDataset # pyre-fixme[13]: Attribute `test_ds` is never initialized.

def __init__(
self, data_dir: str, batch_size: int = 16, num_samples: Optional[int] = None
Expand Down
1 change: 1 addition & 0 deletions torchx/examples/apps/tracker/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -99,6 +99,7 @@ def test(
for data, target in test_loader:
data, target = data.to(device), target.to(device)
output = model(data)
# pyre-fixme[58] Assuming F.nll_loss(...).item() is a number
test_loss += F.nll_loss(
output, target, reduction="sum"
).item() # sum up batch loss
Expand Down
2 changes: 1 addition & 1 deletion torchx/pipelines/kfp/adapter.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@ def component_spec_from_app(app: api.AppDef) -> Tuple[str, api.Role]:
role = app.roles[0]
assert (
role.num_replicas == 1
), f"KFP adapter only supports one replica, got {app.num_replicas}"
), f"KFP adapter only supports one replica, got {app.num_replicas}" # pyre-fixme[16] Assume num_replicas is available on app

command = [role.entrypoint, *role.args]

Expand Down
1 change: 1 addition & 0 deletions torchx/schedulers/aws_batch_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -809,6 +809,7 @@ def _stream_events(
startFromHead=True,
**args,
)
# pyre-fixme[66] Assume this ResourceNotFoundException extends BaseException
except self._log_client.exceptions.ResourceNotFoundException:
return [] # noqa: B901
if response["nextForwardToken"] == next_token:
Expand Down
4 changes: 2 additions & 2 deletions torchx/schedulers/aws_sagemaker_scheduler.py
Original file line number Diff line number Diff line change
Expand Up @@ -261,9 +261,9 @@ def _submit_dryrun(
raise ValueError(
f"{key} is controlled by aws_sagemaker_scheduler and is set to {job_def[key]}"
)
value = cfg.get(key) # pyre-ignore[26]
value = cfg.get(key)
if value is not None:
job_def[key] = value
job_def[key] = value # pyre-ignore[6]

req = AWSSageMakerJob(
job_name=job_name,
Expand Down
4 changes: 3 additions & 1 deletion torchx/schedulers/ray/ray_driver.py
Original file line number Diff line number Diff line change
Expand Up @@ -116,7 +116,9 @@ def load_actor_json(filename: str) -> List[RayActor]:
return actors


def create_placement_group_async(replicas: List[RayActor]) -> PlacementGroup:
def create_placement_group_async(
replicas: List[RayActor],
) -> PlacementGroup: # pyre-ignore[11]
"""return a placement group reference, the corresponding placement group could be scheduled or pending"""
bundles = []
for replica in replicas:
Expand Down
Loading