mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 21:14:14 +08:00
This is one of a series of PRs to update us to PEP585 (changing Dict -> dict, List -> list, etc). Most of the PRs were completely automated with RUFF as follows: Since RUFF UP006 is considered an "unsafe" fix first we need to enable unsafe fixes: ``` --- a/tools/linter/adapters/ruff_linter.py +++ b/tools/linter/adapters/ruff_linter.py @@ -313,6 +313,7 @@ "ruff", "check", "--fix-only", + "--unsafe-fixes", "--exit-zero", *([f"--config={config}"] if config else []), "--stdin-filename", ``` Then we need to tell RUFF to allow UP006 (as a final PR once all of these have landed this will be made permanent): ``` --- a/pyproject.toml +++ b/pyproject.toml @@ -40,7 +40,7 @@ [tool.ruff] -target-version = "py38" +target-version = "py39" line-length = 88 src = ["caffe2", "torch", "torchgen", "functorch", "test"] @@ -87,7 +87,6 @@ "SIM116", # Disable Use a dictionary instead of consecutive `if` statements "SIM117", "SIM118", - "UP006", # keep-runtime-typing "UP007", # keep-runtime-typing ] select = [ ``` Finally running `lintrunner -a --take RUFF` will fix up the deprecated uses. Pull Request resolved: https://github.com/pytorch/pytorch/pull/145101 Approved by: https://github.com/bobrenjc93
75 lines
2.1 KiB
Python
75 lines
2.1 KiB
Python
import sys
|
|
import time
|
|
from functools import cache
|
|
from typing import Any
|
|
|
|
from tools.stats.test_dashboard import upload_additional_info
|
|
from tools.stats.upload_stats_lib import get_s3_resource
|
|
from tools.stats.upload_test_stats import get_tests
|
|
|
|
|
|
BUCKET_PREFIX = "workflows_failing_pending_upload"
|
|
|
|
|
|
@cache
|
|
def get_bucket() -> Any:
|
|
return get_s3_resource().Bucket("gha-artifacts")
|
|
|
|
|
|
def delete_obj(key: str) -> None:
|
|
# Does not raise error if key does not exist
|
|
get_bucket().delete_objects(
|
|
Delete={
|
|
"Objects": [{"Key": key}],
|
|
"Quiet": True,
|
|
}
|
|
)
|
|
|
|
|
|
def put_object(key: str) -> None:
|
|
get_bucket().put_object(
|
|
Key=key,
|
|
Body=b"",
|
|
)
|
|
|
|
|
|
def do_upload(workflow_id: int) -> None:
|
|
workflow_attempt = 1
|
|
test_cases = get_tests(workflow_id, workflow_attempt)
|
|
# Flush stdout so that any errors in upload show up last in the logs.
|
|
sys.stdout.flush()
|
|
upload_additional_info(workflow_id, workflow_attempt, test_cases)
|
|
|
|
|
|
def get_workflow_ids(pending: bool = False) -> list[int]:
|
|
prefix = f"{BUCKET_PREFIX}/{'pending/' if pending else ''}"
|
|
objs = get_bucket().objects.filter(Prefix=prefix)
|
|
return [int(obj.key.split("/")[-1].split(".")[0]) for obj in objs]
|
|
|
|
|
|
def read_s3(pending: bool = False) -> None:
|
|
while True:
|
|
workflows = get_workflow_ids(pending)
|
|
if not workflows:
|
|
if pending:
|
|
break
|
|
# Wait for more stuff to show up
|
|
print("Sleeping for 60 seconds")
|
|
time.sleep(60)
|
|
for workflow_id in workflows:
|
|
print(f"Processing {workflow_id}")
|
|
put_object(f"{BUCKET_PREFIX}/pending/{workflow_id}.txt")
|
|
delete_obj(f"{BUCKET_PREFIX}/{workflow_id}.txt")
|
|
try:
|
|
do_upload(workflow_id)
|
|
except Exception as e:
|
|
print(f"Failed to upload {workflow_id}: {e}")
|
|
delete_obj(f"{BUCKET_PREFIX}/pending/{workflow_id}.txt")
|
|
|
|
|
|
if __name__ == "__main__":
|
|
# Workflows in the pending folder were previously in progress of uploading
|
|
# but failed to complete, so we need to retry them.
|
|
read_s3(pending=True)
|
|
read_s3()
|