Skip to content

refactor

0f28a75
Select commit
Loading
Failed to load commit list.
Merged

Use azpysdk next-* and ruff checks in CI #44679

refactor
0f28a75
Select commit
Loading
Failed to load commit list.
Azure Pipelines / python - storage - tests-weekly failed Mar 2, 2026 in 3h 2m 59s

Build #20260302.1 had test failures

Details

Tests

  • Failed: 1 (0.01%, 1 new, 0 recurring)
  • Passed: 19,055 (97.45%)
  • Other: 498 (2.55%)
  • Total: 19,554

Annotations

Check failure on line 5415 in Build log

See this annotation in the file changed.

@azure-pipelines azure-pipelines / python - storage - tests-weekly

Build log #L5415

The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30

Check failure on line 5416 in Build log

See this annotation in the file changed.

@azure-pipelines azure-pipelines / python - storage - tests-weekly

Build log #L5416

The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30

Check failure on line 5414 in Build log

See this annotation in the file changed.

@azure-pipelines azure-pipelines / python - storage - tests-weekly

Build log #L5414

The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30

Check failure on line 5415 in Build log

See this annotation in the file changed.

@azure-pipelines azure-pipelines / python - storage - tests-weekly

Build log #L5415

The process '/opt/hostedtoolcache/Python/3.10.19/x64/bin/python' failed with exit code 30

Check failure on line 1 in test_get_deleted_paths

See this annotation in the file changed.

@azure-pipelines azure-pipelines / python - storage - tests-weekly

test_get_deleted_paths

AssertionError: assert 2 == 4
 +  where 2 = len([{'name': 'dir3/file_in_dir3', 'deleted_time': datetime.datetime(2026, 3, 2, 18, 30, 15, tzinfo=datetime.timezone.utc), 'remaining_retention_days': 0, 'deletion_id': '134169498159844792', 'file_system': 'filesystem9a1a2b29'}, {'name': 'dir3/subdir/file_in_subdir', 'deleted_time': datetime.datetime(2026, 3, 2, 18, 30, 16, tzinfo=datetime.timezone.utc), 'remaining_retention_days': 0, 'deletion_id': '134169498163155348', 'file_system': 'filesystem9a1a2b29'}])
Raw output
args = (<test_file_system.TestFileSystem object at 0x7f4b165e16f0>,)
kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvaria...ke_soft_delete_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>, ...}
trimmed_kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvaria...ke_soft_delete_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>, ...}

    def record_wrap(*args, **kwargs):
        # ---- your existing trimming/early-exit logic ----
        trimmed_kwargs = {k: v for k, v in kwargs.items()}
        trim_kwargs_from_test_function(test_func, trimmed_kwargs)
    
        if is_live_and_not_recording():
>           return test_func(*args, **trimmed_kwargs)

../../../.venv/azure-storage-file-datalake/.venv_whl/lib/python3.10/site-packages/devtools_testutils/proxy_testcase.py:306: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <test_file_system.TestFileSystem object at 0x7f4b165e16f0>
kwargs = {'__aggregate_cache_key': ('EnvironmentVariableLoader',), 'datalake_storage_account_key': <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523bac0>, 'datalake_storage_account_name': 'dtlkt6cc2793721e94c2c'}
datalake_storage_account_name = 'dtlksdt6cc2793721e94c2c'
datalake_storage_account_key = <devtools_testutils.envvariable_loader.EnvironmentVariable object at 0x7f4b1523b6d0>
file_system = <azure.storage.filedatalake._file_system_client.FileSystemClient object at 0x7f4b1524f6d0>
file0 = <azure.storage.filedatalake._data_lake_file_client.DataLakeFileClient object at 0x7f4b1524f580>
file1 = <azure.storage.filedatalake._data_lake_file_client.DataLakeFileClient object at 0x7f4b1525d9c0>
dir1 = <azure.storage.filedatalake._data_lake_directory_client.DataLakeDirectoryClient object at 0x7f4b1525e800>
dir2 = <azure.storage.filedatalake._data_lake_directory_client.DataLakeDirectoryClient object at 0x7f4b1525fc40>

    @DataLakePreparer()
    @recorded_by_proxy
    def test_get_deleted_paths(self, **kwargs):
        datalake_storage_account_name = kwargs.pop("storage_data_lake_soft_delete_account_name")
        datalake_storage_account_key = kwargs.pop("storage_data_lake_soft_delete_account_key")
    
        self._setUp(datalake_storage_account_name, datalake_storage_account_key)
        # Arrange
        file_system = self._create_file_system()
        file0 = file_system.create_file("file0")
        file1 = file_system.create_file("file1")
    
        dir1 = file_system.create_directory("dir1")
        dir2 = file_system.create_directory("dir2")
        dir3 = file_system.create_directory("dir3")
        file_in_dir3 = dir3.create_file("file_in_dir3")
        file_in_subdir = dir3.create_file("subdir/file_in_subdir")
    
        file0.delete_file()
        file1.delete_file()
        dir1.delete_directory()
        dir2.delete_directory()
        file_in_dir3.delete_file()
        file_in_subdir.delete_file()
        deleted_paths = list(file_system.list_deleted_paths())
        dir3_paths = list(file_system.list_deleted_paths(path_prefix="dir3/"))
    
        # Assert
        assert len(deleted_paths) == 6
        assert len(dir3_paths) == 2
        assert dir3_paths[0].deletion_id is not None
        assert dir3_paths[1].deletion_id is not None
        assert dir3_paths[0].name == 'dir3/file_in_dir3'
        assert dir3_paths[1].name == 'dir3/subdir/file_in_subdir'
    
        paths_generator1 = file_system.list_deleted_paths(results_per_page=2).by_page()
        paths1 = list(next(paths_generator1))
    
        paths_generator2 = file_system.list_deleted_paths(results_per_page=4).by_page(
            continuation_token=paths_generator1.continuation_token)
        paths2 = list(next(paths_generator2))
    
        # Assert
        assert