Files
verl/tests/utils/test_fs_on_cpu.py

95 lines
3.2 KiB
Python

# Copyright 2024 Bytedance Ltd. and/or its affiliates
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import os
from pathlib import Path
import verl.utils.fs as fs
def test_record_and_check_directory_structure(tmp_path):
# Create test directory structure
test_dir = tmp_path / "test_dir"
test_dir.mkdir()
(test_dir / "file1.txt").write_text("test")
(test_dir / "subdir").mkdir()
(test_dir / "subdir" / "file2.txt").write_text("test")
# Create structure record
record_file = fs._record_directory_structure(test_dir)
# Verify record file exists
assert os.path.exists(record_file)
# Initial check should pass
assert fs._check_directory_structure(test_dir, record_file) is True
# Modify structure and verify check fails
(test_dir / "new_file.txt").write_text("test")
assert fs._check_directory_structure(test_dir, record_file) is False
def test_copy_from_hdfs_with_mocks(tmp_path, monkeypatch):
# Mock HDFS dependencies
monkeypatch.setattr(fs, "is_non_local", lambda path: True)
# side_effect will simulate the copy by creating parent dirs + empty file
def fake_copy(src: str, dst: str, *args, **kwargs):
dst_path = Path(dst)
dst_path.parent.mkdir(parents=True, exist_ok=True)
dst_path.write_bytes(b"") # touch an empty file
monkeypatch.setattr(fs, "copy", fake_copy) # Mock actual HDFS copy
# Test parameters
test_cache = tmp_path / "cache"
hdfs_path = "hdfs://test/path/file.txt"
# Test initial copy
local_path = fs.copy_to_local(hdfs_path, cache_dir=test_cache)
expected_path = os.path.join(test_cache, fs.md5_encode(hdfs_path), os.path.basename(hdfs_path))
assert local_path == expected_path
assert os.path.exists(local_path)
def test_always_recopy_flag(tmp_path, monkeypatch):
# Mock HDFS dependencies
monkeypatch.setattr(fs, "is_non_local", lambda path: True)
copy_call_count = 0
def fake_copy(src: str, dst: str, *args, **kwargs):
nonlocal copy_call_count
copy_call_count += 1
dst_path = Path(dst)
dst_path.parent.mkdir(parents=True, exist_ok=True)
dst_path.write_bytes(b"")
monkeypatch.setattr(fs, "copy", fake_copy) # Mock actual HDFS copy
test_cache = tmp_path / "cache"
hdfs_path = "hdfs://test/path/file.txt"
# Initial copy (always_recopy=False)
fs.copy_to_local(hdfs_path, cache_dir=test_cache)
assert copy_call_count == 1
# Force recopy (always_recopy=True)
fs.copy_to_local(hdfs_path, cache_dir=test_cache, always_recopy=True)
assert copy_call_count == 2
# Subsequent normal call (always_recopy=False)
fs.copy_to_local(hdfs_path, cache_dir=test_cache)
assert copy_call_count == 2 # Should not increment