mirror of
https://github.com/pytorch/pytorch.git
synced 2025-10-20 21:14:14 +08:00
Land code coverage tool for OSS (#43778)
Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/43778 Move code_coverage_tool from experimental folder to caffe2/tools folder. Delete `TODO` and fb-related code. Test Plan: Test locally Reviewed By: malfet Differential Revision: D23399983 fbshipit-source-id: 92316fd3cc88409d087d2dc6ed0be674155b3762
This commit is contained in:
committed by
Facebook GitHub Bot
parent
89e2a3591e
commit
0564d7a652
174
tools/code_coverage/package/tool/summarize_jsons.py
Normal file
174
tools/code_coverage/package/tool/summarize_jsons.py
Normal file
@ -0,0 +1,174 @@
|
||||
import json
|
||||
import os
|
||||
import time
|
||||
from typing import Any, Dict, List, Optional, Set, Tuple
|
||||
|
||||
from ..util.setting import JSON_FOLDER_BASE_DIR, TestList, TestStatusType
|
||||
from ..util.utils import (
|
||||
check_compiler_type,
|
||||
get_cov_type,
|
||||
print_error,
|
||||
print_time,
|
||||
related_to_test_list,
|
||||
)
|
||||
from .parser.coverage_record import CoverageRecord
|
||||
from .parser.gcov_coverage_parser import GcovCoverageParser
|
||||
from .parser.llvm_coverage_parser import LlvmCoverageParser
|
||||
from .print_report import file_oriented_report, line_oriented_report
|
||||
|
||||
|
||||
# coverage_records: Dict[str, LineInfo] = dict()
|
||||
covered_lines: Dict[str, Set[int]] = {}
|
||||
uncovered_lines: Dict[str, Set[int]] = {}
|
||||
tests_type: TestStatusType = {"success": set(), "partial": set(), "fail": set()}
|
||||
|
||||
|
||||
def transform_file_name(file_path: str, interested_folders: List[str]) -> Optional[str]:
|
||||
remove_patterns: Set[str] = {".DEFAULT.cpp", ".AVX.cpp", ".AVX2.cpp"}
|
||||
for pattern in remove_patterns:
|
||||
file_path = file_path.replace(pattern, "")
|
||||
# if have interested folder
|
||||
for folder in interested_folders:
|
||||
if folder in file_path:
|
||||
return file_path[file_path.find(folder) :]
|
||||
return ""
|
||||
|
||||
|
||||
def is_intrested_file(file_path: str, interested_folders: List[str]):
|
||||
ignored_patterns = ["cuda", "aten/gen_aten", "aten/aten_", "build/"]
|
||||
if any([pattern in file_path for pattern in ignored_patterns]):
|
||||
return False
|
||||
|
||||
for folder in interested_folders:
|
||||
i_folder = folder if folder.endswith("/") else f"{folder}/"
|
||||
if i_folder in file_path:
|
||||
return True
|
||||
return False
|
||||
|
||||
|
||||
def get_json_obj(json_file: str) -> Tuple[Any, int]:
|
||||
"""
|
||||
Sometimes at the start of file llvm/gcov will complains "fail to find coverage data",
|
||||
then we need to skip these lines
|
||||
-- success read: 0 - this json file have the full json coverage information
|
||||
-- partial success: 1 - this json file starts with some error prompt, but still have the coverage information
|
||||
-- fail to read: 2 - this json file doesn't have any coverage information
|
||||
"""
|
||||
read_status = -1
|
||||
with open(json_file) as f:
|
||||
lines = f.readlines()
|
||||
for line in lines:
|
||||
try:
|
||||
json_obj = json.loads(line)
|
||||
except json.JSONDecodeError:
|
||||
read_status = 1
|
||||
continue
|
||||
else:
|
||||
if read_status == -1:
|
||||
# not meet jsonDecoderError before, return success
|
||||
read_status = 0
|
||||
return (json_obj, read_status)
|
||||
return None, 2
|
||||
|
||||
|
||||
def parse_json(json_file: str) -> List[CoverageRecord]:
|
||||
print("start parse:", json_file)
|
||||
json_obj, read_status = get_json_obj(json_file)
|
||||
if read_status == 0:
|
||||
tests_type["success"].add(json_file)
|
||||
elif read_status == 1:
|
||||
tests_type["partial"].add(json_file)
|
||||
else:
|
||||
tests_type["fail"].add(json_file)
|
||||
raise RuntimeError(
|
||||
"Fail to do code coverage! Fail to load json file: ", json_file
|
||||
)
|
||||
cov_type = get_cov_type()
|
||||
check_compiler_type(cov_type)
|
||||
if cov_type == "CLANG":
|
||||
coverage_records = LlvmCoverageParser(json_obj).parse("fbcode")
|
||||
# print(coverage_records)
|
||||
elif cov_type == "GCC":
|
||||
coverage_records = GcovCoverageParser(json_obj).parse()
|
||||
|
||||
return coverage_records
|
||||
|
||||
|
||||
def parse_jsons(test_list: TestList, interested_folders: List[str]) -> None:
|
||||
g = os.walk(JSON_FOLDER_BASE_DIR)
|
||||
|
||||
for path, _, file_list in g:
|
||||
for file_name in file_list:
|
||||
if file_name.endswith(".json"):
|
||||
if not related_to_test_list(file_name, test_list):
|
||||
continue
|
||||
json_file = os.path.join(path, file_name)
|
||||
try:
|
||||
coverage_records = parse_json(json_file)
|
||||
except RuntimeError:
|
||||
print_error("Fail to load json file: ", json_file)
|
||||
continue
|
||||
# collect information from each target's export file and merge them together:
|
||||
update_coverage(coverage_records, interested_folders)
|
||||
|
||||
|
||||
def update_coverage(
|
||||
coverage_records: List[CoverageRecord], interested_folders: List[str]
|
||||
) -> None:
|
||||
for item in coverage_records:
|
||||
# extract information for the record
|
||||
record = item.to_dict()
|
||||
file_path = record["filepath"]
|
||||
if not is_intrested_file(file_path, interested_folders):
|
||||
continue
|
||||
covered_range = record["covered_lines"]
|
||||
uncovered_range = record["uncovered_lines"]
|
||||
# transform file name: remote/13223/caffe2/aten -> caffe2/aten
|
||||
file_path = transform_file_name(file_path, interested_folders)
|
||||
if file_path is None:
|
||||
continue
|
||||
# if file not exists, add it into dictionary
|
||||
if file_path not in covered_lines:
|
||||
covered_lines[file_path] = set()
|
||||
if file_path not in uncovered_lines:
|
||||
uncovered_lines[file_path] = set()
|
||||
# update this file's covered and uncovered lines
|
||||
if covered_range is not None:
|
||||
covered_lines[file_path].update(covered_range)
|
||||
if uncovered_range is not None:
|
||||
uncovered_lines[file_path].update(uncovered_range)
|
||||
|
||||
|
||||
def update_set() -> None:
|
||||
for file_name in covered_lines:
|
||||
# difference_update
|
||||
uncovered_lines[file_name].difference_update(covered_lines[file_name])
|
||||
|
||||
|
||||
def summarize_jsons(
|
||||
test_list: TestList,
|
||||
interested_folders: List[str],
|
||||
coverage_only: List[str],
|
||||
program_start_time: float,
|
||||
) -> None:
|
||||
start_time = time.time()
|
||||
parse_jsons(test_list, interested_folders)
|
||||
update_set()
|
||||
line_oriented_report(
|
||||
test_list,
|
||||
tests_type,
|
||||
interested_folders,
|
||||
coverage_only,
|
||||
covered_lines,
|
||||
uncovered_lines,
|
||||
)
|
||||
file_oriented_report(
|
||||
test_list,
|
||||
tests_type,
|
||||
interested_folders,
|
||||
coverage_only,
|
||||
program_start_time,
|
||||
covered_lines,
|
||||
uncovered_lines,
|
||||
)
|
||||
print_time("summary jsons take time: ", start_time)
|
Reference in New Issue
Block a user