mirror of
https://github.com/huggingface/transformers.git
synced 2025-10-21 17:48:57 +08:00
Compare commits
7 Commits
v4.51.3-Bi
...
simplify
Author | SHA1 | Date | |
---|---|---|---|
7ff8c111c2 | |||
bcb1c99b5a | |||
668a3d725b | |||
14d1f06e9b | |||
817f3c78f0 | |||
7032ced604 | |||
0e1d8e2a5d |
@ -106,7 +106,7 @@ jobs:
|
|||||||
} >> "$GITHUB_ENV"
|
} >> "$GITHUB_ENV"
|
||||||
|
|
||||||
- name: Send processed report
|
- name: Send processed report
|
||||||
if: ${{ env.REPORT_TEXT != '' }}
|
if: ${{ !endsWith(env.REPORT_TEXT, '{}') }}
|
||||||
uses: slackapi/slack-github-action@6c661ce58804a1a20f6dc5fbee7f0381b469e001
|
uses: slackapi/slack-github-action@6c661ce58804a1a20f6dc5fbee7f0381b469e001
|
||||||
with:
|
with:
|
||||||
# Slack channel id, channel name, or user id to post message.
|
# Slack channel id, channel name, or user id to post message.
|
||||||
|
58
.github/workflows/self-scheduled-caller.yml
vendored
58
.github/workflows/self-scheduled-caller.yml
vendored
@ -7,7 +7,7 @@ on:
|
|||||||
- cron: "17 2 * * *"
|
- cron: "17 2 * * *"
|
||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
- run_scheduled_ci*
|
- simplify
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
model-ci:
|
model-ci:
|
||||||
@ -20,59 +20,3 @@ jobs:
|
|||||||
docker: huggingface/transformers-all-latest-gpu
|
docker: huggingface/transformers-all-latest-gpu
|
||||||
ci_event: Daily CI
|
ci_event: Daily CI
|
||||||
secrets: inherit
|
secrets: inherit
|
||||||
|
|
||||||
torch-pipeline:
|
|
||||||
name: Torch pipeline CI
|
|
||||||
uses: ./.github/workflows/self-scheduled.yml
|
|
||||||
with:
|
|
||||||
job: run_pipelines_torch_gpu
|
|
||||||
slack_report_channel: "#transformers-ci-daily-pipeline-torch"
|
|
||||||
runner: daily-ci
|
|
||||||
docker: huggingface/transformers-pytorch-gpu
|
|
||||||
ci_event: Daily CI
|
|
||||||
secrets: inherit
|
|
||||||
|
|
||||||
tf-pipeline:
|
|
||||||
name: TF pipeline CI
|
|
||||||
uses: ./.github/workflows/self-scheduled.yml
|
|
||||||
with:
|
|
||||||
job: run_pipelines_tf_gpu
|
|
||||||
slack_report_channel: "#transformers-ci-daily-pipeline-tf"
|
|
||||||
runner: daily-ci
|
|
||||||
docker: huggingface/transformers-tensorflow-gpu
|
|
||||||
ci_event: Daily CI
|
|
||||||
secrets: inherit
|
|
||||||
|
|
||||||
example-ci:
|
|
||||||
name: Example CI
|
|
||||||
uses: ./.github/workflows/self-scheduled.yml
|
|
||||||
with:
|
|
||||||
job: run_examples_gpu
|
|
||||||
slack_report_channel: "#transformers-ci-daily-examples"
|
|
||||||
runner: daily-ci
|
|
||||||
docker: huggingface/transformers-all-latest-gpu
|
|
||||||
ci_event: Daily CI
|
|
||||||
secrets: inherit
|
|
||||||
|
|
||||||
deepspeed-ci:
|
|
||||||
name: DeepSpeed CI
|
|
||||||
uses: ./.github/workflows/self-scheduled.yml
|
|
||||||
with:
|
|
||||||
job: run_torch_cuda_extensions_gpu
|
|
||||||
slack_report_channel: "#transformers-ci-daily-deepspeed"
|
|
||||||
runner: daily-ci
|
|
||||||
docker: huggingface/transformers-pytorch-deepspeed-latest-gpu
|
|
||||||
ci_event: Daily CI
|
|
||||||
working-directory-prefix: /workspace
|
|
||||||
secrets: inherit
|
|
||||||
|
|
||||||
quantization-ci:
|
|
||||||
name: Quantization CI
|
|
||||||
uses: ./.github/workflows/self-scheduled.yml
|
|
||||||
with:
|
|
||||||
job: run_quantization_torch_gpu
|
|
||||||
slack_report_channel: "#transformers-ci-daily-quantization"
|
|
||||||
runner: daily-ci
|
|
||||||
docker: huggingface/transformers-quantization-latest-gpu
|
|
||||||
ci_event: Daily CI
|
|
||||||
secrets: inherit
|
|
||||||
|
@ -16,6 +16,8 @@
|
|||||||
|
|
||||||
import unittest
|
import unittest
|
||||||
|
|
||||||
|
from accelerate.test_utils import require_multi_gpu
|
||||||
|
|
||||||
from transformers import ViTConfig
|
from transformers import ViTConfig
|
||||||
from transformers.testing_utils import (
|
from transformers.testing_utils import (
|
||||||
require_accelerate,
|
require_accelerate,
|
||||||
@ -208,6 +210,9 @@ class ViTModelTest(ModelTesterMixin, PipelineTesterMixin, unittest.TestCase):
|
|||||||
test_resize_embeddings = False
|
test_resize_embeddings = False
|
||||||
test_head_masking = False
|
test_head_masking = False
|
||||||
|
|
||||||
|
def test_foo(self):
|
||||||
|
assert 1 == 2
|
||||||
|
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
self.model_tester = ViTModelTester(self)
|
self.model_tester = ViTModelTester(self)
|
||||||
self.config_tester = ConfigTester(self, config_class=ViTConfig, has_text_modality=False, hidden_size=37)
|
self.config_tester = ConfigTester(self, config_class=ViTConfig, has_text_modality=False, hidden_size=37)
|
||||||
|
@ -182,7 +182,15 @@ if __name__ == "__main__":
|
|||||||
info = {"test": test, "commit": commit}
|
info = {"test": test, "commit": commit}
|
||||||
info.update(get_commit_info(commit))
|
info.update(get_commit_info(commit))
|
||||||
failed_tests_with_bad_commits.append(info)
|
failed_tests_with_bad_commits.append(info)
|
||||||
|
|
||||||
|
# If no single-gpu test failures, remove the key
|
||||||
|
if len(failed_tests_with_bad_commits) > 0:
|
||||||
reports[model]["single-gpu"] = failed_tests_with_bad_commits
|
reports[model]["single-gpu"] = failed_tests_with_bad_commits
|
||||||
|
else:
|
||||||
|
reports[model].pop("single-gpu", None)
|
||||||
|
|
||||||
|
# remove the models without any test failure
|
||||||
|
reports = {k: v for k, v in reports.items() if len(v) > 0}
|
||||||
|
|
||||||
with open(args.output_file, "w", encoding="UTF-8") as fp:
|
with open(args.output_file, "w", encoding="UTF-8") as fp:
|
||||||
json.dump(reports, fp, ensure_ascii=False, indent=4)
|
json.dump(reports, fp, ensure_ascii=False, indent=4)
|
||||||
|
@ -530,14 +530,13 @@ class Message:
|
|||||||
|
|
||||||
# upload results to Hub dataset
|
# upload results to Hub dataset
|
||||||
file_path = os.path.join(os.getcwd(), f"ci_results_{job_name}/new_model_failures.txt")
|
file_path = os.path.join(os.getcwd(), f"ci_results_{job_name}/new_model_failures.txt")
|
||||||
commit_info = api.upload_file(
|
# commit_info = api.upload_file(
|
||||||
path_or_fileobj=file_path,
|
# path_or_fileobj=file_path,
|
||||||
path_in_repo=f"{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.txt",
|
# path_in_repo=f"{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.txt",
|
||||||
repo_id="hf-internal-testing/transformers_daily_ci",
|
# repo_id="hf-internal-testing/transformers_daily_ci",
|
||||||
repo_type="dataset",
|
# repo_type="dataset",
|
||||||
token=os.environ.get("TRANSFORMERS_CI_RESULTS_UPLOAD_TOKEN", None),
|
# token=os.environ.get("TRANSFORMERS_CI_RESULTS_UPLOAD_TOKEN", None),
|
||||||
)
|
# )
|
||||||
url = f"https://huggingface.co/datasets/hf-internal-testing/transformers_daily_ci/raw/{commit_info.oid}/{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.txt"
|
|
||||||
|
|
||||||
# extra processing to save to json format
|
# extra processing to save to json format
|
||||||
new_failed_tests = {}
|
new_failed_tests = {}
|
||||||
@ -556,13 +555,14 @@ class Message:
|
|||||||
|
|
||||||
# upload results to Hub dataset
|
# upload results to Hub dataset
|
||||||
file_path = os.path.join(os.getcwd(), f"ci_results_{job_name}/new_model_failures.json")
|
file_path = os.path.join(os.getcwd(), f"ci_results_{job_name}/new_model_failures.json")
|
||||||
_ = api.upload_file(
|
commit_info = api.upload_file(
|
||||||
path_or_fileobj=file_path,
|
path_or_fileobj=file_path,
|
||||||
path_in_repo=f"{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.json",
|
path_in_repo=f"{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.json",
|
||||||
repo_id="hf-internal-testing/transformers_daily_ci",
|
repo_id="hf-internal-testing/transformers_daily_ci",
|
||||||
repo_type="dataset",
|
repo_type="dataset",
|
||||||
token=os.environ.get("TRANSFORMERS_CI_RESULTS_UPLOAD_TOKEN", None),
|
token=os.environ.get("TRANSFORMERS_CI_RESULTS_UPLOAD_TOKEN", None),
|
||||||
)
|
)
|
||||||
|
url = f"https://huggingface.co/datasets/hf-internal-testing/transformers_daily_ci/raw/{commit_info.oid}/{datetime.datetime.today().strftime('%Y-%m-%d')}/ci_results_{job_name}/new_model_failures.json"
|
||||||
|
|
||||||
block = {
|
block = {
|
||||||
"type": "section",
|
"type": "section",
|
||||||
@ -1250,7 +1250,7 @@ if __name__ == "__main__":
|
|||||||
)
|
)
|
||||||
|
|
||||||
prev_ci_artifacts = None
|
prev_ci_artifacts = None
|
||||||
if is_scheduled_ci_run:
|
if True:
|
||||||
if job_name == "run_models_gpu":
|
if job_name == "run_models_gpu":
|
||||||
# Get the last previously completed CI's failure tables
|
# Get the last previously completed CI's failure tables
|
||||||
artifact_names = [f"ci_results_{job_name}"]
|
artifact_names = [f"ci_results_{job_name}"]
|
||||||
|
@ -28,7 +28,7 @@ if __name__ == "__main__":
|
|||||||
data = json.load(fp)
|
data = json.load(fp)
|
||||||
|
|
||||||
# TODO: extend
|
# TODO: extend
|
||||||
team_members = ["ydshieh", "zucchini-nlp", "ArthurZucker", "gante", "LysandreJik", "molbap", "qubvel"]
|
team_members = ["ydshieh", "zucchini-nlp", "ArthurZucker", "gante", "LysandreJik", "molbap", "qubvel", "Rocketknight1", "muellerzr", "SunMarc"]
|
||||||
|
|
||||||
# Counting the number of failures grouped by authors
|
# Counting the number of failures grouped by authors
|
||||||
new_data = {}
|
new_data = {}
|
||||||
|
@ -62,4 +62,5 @@ if __name__ == "__main__":
|
|||||||
start = end
|
start = end
|
||||||
end = start + num_jobs_per_splits + (1 if idx < num_jobs % args.num_splits else 0)
|
end = start + num_jobs_per_splits + (1 if idx < num_jobs % args.num_splits else 0)
|
||||||
model_splits.append(d[start:end])
|
model_splits.append(d[start:end])
|
||||||
|
model_splits = [["models/vit"]]
|
||||||
print(model_splits)
|
print(model_splits)
|
||||||
|
Reference in New Issue
Block a user