Use azpysdk next-* and ruff checks in CI #44679
Merged
Azure Pipelines / python - storage - tests-weekly
failed
Mar 2, 2026 in 3h 2m 59s
Build #20260302.1 had test failures
Details
- Failed: 1 (0.01%, 1 new, 0 recurring)
- Passed: 19,055 (97.45%)
- Other: 498 (2.55%)
- Total: 19,554
Annotations
Check failure on line 5415 in Build log
azure-pipelines / python - storage - tests-weekly
Build log #L5415
The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30
Check failure on line 5416 in Build log
azure-pipelines / python - storage - tests-weekly
Build log #L5416
The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30
Check failure on line 5414 in Build log
azure-pipelines / python - storage - tests-weekly
Build log #L5414
The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30
Check failure on line 5415 in Build log
azure-pipelines / python - storage - tests-weekly
Build log #L5415
The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30
Check failure on line 1 in test_get_deleted_paths
azure-pipelines / python - storage - tests-weekly
test_get_deleted_paths
AssertionError: assert 2 == 4
+ where 2 = len([{'name': 'dir3/file_in_dir3', 'deleted_time': datetime.datetime(2026, 3, 2, 18, 30, 15, tzinfo=datetime.timezone.utc), 'remaining_retention_days': 0, 'deletion_id': '134169498159844792', 'file_system': 'filesystem9a1a2b29'}, {'name': 'dir3/subdir/file_in_subdir', 'deleted_time': datetime.datetime(2026, 3, 2, 18, 30, 16, tzinfo=datetime.timezone.utc), 'remaining_retention_days': 0, 'deletion_id': '134169498163155348', 'file_system': 'filesystem9a1a2b29'}])
Raw output
args = (<test_file_system.TestFileSystem object at 0x7f4b165e16f0>,)
kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvaria...ke_soft_delete_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>, ...}
trimmed_kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvaria...ke_soft_delete_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>, ...}
def record_wrap(*args, **kwargs):
# ---- your existing trimming/early-exit logic ----
trimmed_kwargs = {k: v for k, v in kwargs.items()}
trim_kwargs_from_test_function(test_func, trimmed_kwargs)
if is_live_and_not_recording():
> return test_func(*args, **trimmed_kwargs)
../../../.venv/azure-storage-file-datalake/.venv_whl/lib/python3.10/site-packages/devtools_testutils/proxy_testcase.py:306:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <test_file_system.TestFileSystem object at 0x7f4b165e16f0>
kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523bac0>, 'datalake_storage_account_name': 'dtlkt6cc2793721e94c2c'}
datalake_storage_account_name = 'dtlksdt6cc2793721e94c2c'
datalake_storage_account_key = <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>
file_system = <azure.storage.filedatalake._file_system_client.FileSystemClient object at 0x7f4b1524f6d0>
file0 = <azure.storage.filedatalake._data_lake_file_client.DataLakeFileClient object at 0x7f4b1524f580>
file1 = <azure.storage.filedatalake._data_lake_file_client.DataLakeFileClient object at 0x7f4b1525d9c0>
dir1 = <azure.storage.filedatalake._data_lake_directory_client.DataLakeDirectoryClient object at 0x7f4b1525e800>
dir2 = <azure.storage.filedatalake._data_lake_directory_client.DataLakeDirectoryClient object at 0x7f4b1525fc40>
@DataLakePreparer()
@recorded_by_proxy
def test_get_deleted_paths(self, **kwargs):
datalake_storage_account_name = kwargs.pop("storage_data_lake_soft_delete_account_name")
datalake_storage_account_key = kwargs.pop("storage_data_lake_soft_delete_account_key")
self._setUp(datalake_storage_account_name, datalake_storage_account_key)
# Arrange
file_system = self._create_file_system()
file0 = file_system.create_file("file0")
file1 = file_system.create_file("file1")
dir1 = file_system.create_directory("dir1")
dir2 = file_system.create_directory("dir2")
dir3 = file_system.create_directory("dir3")
file_in_dir3 = dir3.create_file("file_in_dir3")
file_in_subdir = dir3.create_file("subdir/file_in_subdir")
file0.delete_file()
file1.delete_file()
dir1.delete_directory()
dir2.delete_directory()
file_in_dir3.delete_file()
file_in_subdir.delete_file()
deleted_paths = list(file_system.list_deleted_paths())
dir3_paths = list(file_system.list_deleted_paths(path_prefix="dir3/"))
# Assert
assert len(deleted_paths) == 6
assert len(dir3_paths) == 2
assert dir3_paths[0].deletion_id is not None
assert dir3_paths[1].deletion_id is not None
assert dir3_paths[0].name == 'dir3/file_in_dir3'
assert dir3_paths[1].name == 'dir3/subdir/file_in_subdir'
paths_generator1 = file_system.list_deleted_paths(results_per_page=2).by_page()
paths1 = list(next(paths_generator1))
paths_generator2 = file_system.list_deleted_paths(results_per_page=4).by_page(
continuation_token=paths_generator1.continuation_token)
paths2 = list(next(paths_generator2))
# Assert
assert
Loading