galtimur commited on
Commit
e10d948
1 Parent(s): 0107730

Delete data/python

Browse files

removed files before adding new filtered

This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. data/python/028ad1e.json +0 -1
  2. data/python/02d7747.json +0 -0
  3. data/python/034b195.json +0 -1
  4. data/python/03669a5.json +0 -1
  5. data/python/06bc6b6.json +0 -1
  6. data/python/077f6aa.json +0 -1
  7. data/python/07bc10c.json +0 -1
  8. data/python/0828c8d.json +0 -0
  9. data/python/08c47c4.json +0 -1
  10. data/python/0b08b8e.json +0 -0
  11. data/python/0b5c0d2.json +0 -0
  12. data/python/0b93d2d.json +0 -0
  13. data/python/0d26cc1.json +0 -1
  14. data/python/0d367ce.json +0 -0
  15. data/python/0f71221.json +0 -1
  16. data/python/102f918.json +0 -1
  17. data/python/1168b92.json +0 -1
  18. data/python/124065b.json +0 -1
  19. data/python/16a0c04.json +0 -1
  20. data/python/1afe2c9.json +0 -1
  21. data/python/1b22eb7.json +0 -1
  22. data/python/1e48195.json +0 -1
  23. data/python/212ff27.json +0 -1
  24. data/python/2201be2.json +0 -1
  25. data/python/2327ca6.json +0 -0
  26. data/python/2a104bf.json +0 -1
  27. data/python/2a59b55.json +0 -1
  28. data/python/2ab9e84.json +0 -1
  29. data/python/2c06ffa.json +0 -1
  30. data/python/2ccb7d3.json +0 -1
  31. data/python/2e41e78.json +0 -0
  32. data/python/2f0605c.json +0 -1
  33. data/python/3054659.json +0 -0
  34. data/python/3137ef6.json +0 -1
  35. data/python/31c3e1d.json +0 -1
  36. data/python/385c14d.json +0 -1
  37. data/python/386bf6f.json +0 -1
  38. data/python/38a3024.json +0 -1
  39. data/python/3dd8e44.json +0 -1
  40. data/python/3ed7a88.json +0 -0
  41. data/python/3fcfe29.json +0 -1
  42. data/python/434321a.json +0 -0
  43. data/python/43dd59c.json +0 -1
  44. data/python/4410203.json +0 -1
  45. data/python/44b56e0.json +0 -1
  46. data/python/4541644.json +0 -1
  47. data/python/483a1db.json +0 -1
  48. data/python/4921af9.json +0 -0
  49. data/python/4b7e472.json +0 -1
  50. data/python/4d5898b.json +0 -1
data/python/028ad1e.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 129, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "AjayP13-patch-1", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "AjayP13", "sha_fail": "028ad1efee2c41691d78e5a4de90ebd6f8236cad", "sha_success": "9ff59024aee19e24948401eb8cc7057602592b49", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\nsrc/accelerate/utils/__init__.py:1:1: I001 [*] Import block is un-sorted or un-formatted\nsrc/accelerate/utils/modeling.py:15:1: I001 [*] Import block is un-sorted or un-formatted\nFound 2 errors.\n[*] 2 fixable with the `--fix` option.\nmake: *** [Makefile:16: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/src/accelerate/utils/__init__.py b/src/accelerate/utils/__init__.py\nindex ddf794a..7179afb 100644\n--- a/src/accelerate/utils/__init__.py\n+++ b/src/accelerate/utils/__init__.py\n@@ -59,7 +59,6 @@ from .imports import (\n is_comet_ml_available,\n is_cuda_available,\n is_datasets_available,\n- is_peft_available,\n is_deepspeed_available,\n is_dvclive_available,\n is_fp8_available,\n@@ -70,6 +69,7 @@ from .imports import (\n is_msamp_available,\n is_npu_available,\n is_pandas_available,\n+ is_peft_available,\n is_rich_available,\n is_sagemaker_available,\n is_tensorboard_available,\n@@ -81,7 +81,6 @@ from .imports import (\n is_xpu_available,\n )\n from .modeling import (\n- is_peft_model,\n calculate_maximum_sizes,\n check_device_map,\n check_tied_parameters_in_config,\n@@ -96,6 +95,7 @@ from .modeling import (\n get_mixed_precision_context_manager,\n id_tensor_storage,\n infer_auto_device_map,\n+ is_peft_model,\n load_checkpoint_in_model,\n load_offloaded_weights,\n load_state_dict,\ndiff --git a/src/accelerate/utils/modeling.py b/src/accelerate/utils/modeling.py\nindex 802b13c..03d3a39 100644\n--- a/src/accelerate/utils/modeling.py\n+++ b/src/accelerate/utils/modeling.py\n@@ -30,7 +30,7 @@ import torch.nn as nn\n from ..state import AcceleratorState\n from .constants import SAFE_WEIGHTS_NAME, WEIGHTS_NAME\n from .dataclasses import AutocastKwargs, CustomDtype, DistributedType\n-from .imports import is_mps_available, is_npu_available, is_xpu_available, is_peft_available\n+from .imports import is_mps_available, is_npu_available, is_peft_available, is_xpu_available\n from .offload import load_offloaded_weight, offload_weight, save_offload_index\n from .tqdm import is_tqdm_available, tqdm\n \n", "difficulty": 0, "changed_files": ["src/accelerate/utils/__init__.py", "src/accelerate/utils/modeling.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/028ad1efee2c41691d78e5a4de90ebd6f8236cad"}
 
 
data/python/02d7747.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/034b195.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 157, "repo_owner": "modin-project", "repo_name": "modin", "head_branch": "merge_without_to_pandas", "workflow_name": "fuzzy", "workflow_filename": "fuzzydata-test.yml", "workflow_path": ".github/workflows/fuzzydata-test.yml", "contributor": "arunjose696", "sha_fail": "034b1956a0914b739c3e5fc2e19ffcfee0f08c02", "sha_success": "ac24182c5553424d2d00a7e573b5aa289c79d955", "workflow": "name: fuzzy\non:\n pull_request:\n paths:\n # NOTE: keep these paths in sync with the paths that trigger the CI Github\n # Actions in .github/workflows/ci.yml\n - .github/workflows/**\n - '!.github/workflows/push-to-master.yml'\n - asv_bench/**\n - modin/**\n - requirements/**\n - scripts/**\n - environment-dev.yml\n - requirements-dev.txt\n - setup.cfg\n - setup.py\n - versioneer.py\nconcurrency:\n # Cancel other jobs in the same branch. We don't care whether CI passes\n # on old commits.\n group: ${{ github.workflow }}-${{ github.ref }}\n cancel-in-progress: ${{ startsWith(github.ref, 'refs/pull/') }}\nenv:\n MODIN_GITHUB_CI: true\njobs:\n test-fuzzydata:\n runs-on: ubuntu-latest\n defaults:\n run:\n shell: bash -l {0}\n strategy:\n matrix:\n python-version: [\"3.9\"]\n engine: [\"ray\", \"dask\"]\n steps:\n - uses: actions/checkout@v3\n - uses: ./.github/actions/mamba-env\n with:\n environment-file: environment-dev.yml\n python-version: ${{matrix.python-version}}\n - name: test-fuzzydata (engine ${{matrix.engine}}, python ${{matrix.python-version}})\n run: python -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-${{matrix.engine}}/run.log --log-file-level=INFO\n env:\n MODIN_ENGINE: ${{matrix.engine}}\n - uses: actions/upload-artifact@v3\n if: success() || failure()\n with:\n name: fuzzydata-test-workflow-${{matrix.engine}}\n path: /tmp/fuzzydata-test-wf-${{matrix.engine}}/* # Must match output dir in test_fuzzydata.py\n if-no-files-found: error\n", "logs": [{"step_name": "test-fuzzydata (3.9, ray)/4_test-fuzzydata (engine ray, python 3.9).txt", "log": "##[group]Run python -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-ray/run.log --log-file-level=INFO\n\u001b[36;1mpython -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-ray/run.log --log-file-level=INFO\u001b[0m\nshell: /usr/bin/bash -l {0}\nenv:\n MODIN_GITHUB_CI: true\n INPUT_RUN_POST: true\n CONDA: /usr/share/miniconda3\n CONDA_PKGS_DIR: /home/runner/conda_pkgs_dir\n MODIN_ENGINE: ray\n##[endgroup]\n============================= test session starts ==============================\nplatform linux -- Python 3.9.18, pytest-7.4.4, pluggy-1.3.0\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/modin/modin\nconfigfile: setup.cfg\nplugins: benchmark-4.0.0, xdist-3.5.0, anyio-4.2.0, Faker-22.1.0, cov-4.1.0\ncollected 1 item\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py F [100%]\n\n=================================== FAILURES ===================================\n________________________ test_fuzzydata_sample_workflow ________________________\n\n def test_fuzzydata_sample_workflow():\n # Workflow Generation Options\n wf_name = str(uuid.uuid4())[:8] # Unique name for the generated workflow\n num_versions = 10 # Number of unique CSV files to generate\n cols = 33 # Columns in Base Artifact\n rows = 1000 # Rows in Base Artifact\n bfactor = 1.0 # Branching Factor - 0.1 is linear, 10.0 is star-like\n exclude_ops = [\"groupby\"] # In-Memory groupby operations cause issue #4287\n matfreq = 2 # How many operations to chain before materialization\n \n engine = Engine.get().lower()\n \n # Create Output Directory for Workflow Data\n base_out_directory = (\n f\"/tmp/fuzzydata-test-wf-{engine}/\" # Must match corresponding github-action\n )\n if os.path.exists(base_out_directory):\n shutil.rmtree(base_out_directory)\n output_directory = f\"{base_out_directory}/{wf_name}/\"\n os.makedirs(output_directory, exist_ok=True)\n \n # Start Workflow Generation\n> workflow = generate_workflow(\n workflow_class=ModinWorkflow,\n name=wf_name,\n num_versions=num_versions,\n base_shape=(cols, rows),\n out_directory=output_directory,\n bfactor=bfactor,\n exclude_ops=exclude_ops,\n matfreq=matfreq,\n wf_options={\"modin_engine\": engine},\n )\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py:47: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/generator.py:381: in generate_workflow\n raise e\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/generator.py:367: in generate_workflow\n wf.execute_current_operation(next_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/workflow.py:197: in execute_current_operation\n raise e\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/workflow.py:168: in execute_current_operation\n new_artifact = self.current_operation.execute(new_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/operation.py:172: in execute\n result = self.materialize(new_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/clients/pandas.py:114: in materialize\n new_df = eval(self.code)\n<string>:1: in <module>\n ???\nmodin/logging/logger_decorator.py:129: in run_and_log\n return obj(*args, **kwargs)\nmodin/pandas/dataframe.py:1301: in merge\n query_compiler=self._query_compiler.merge(\nmodin/logging/logger_decorator.py:129: in run_and_log\n return obj(*args, **kwargs)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <modin.core.storage_formats.pandas.query_compiler.PandasQueryCompiler object at 0x7f9fccfaedf0>\nright = <modin.core.storage_formats.pandas.query_compiler.PandasQueryCompiler object at 0x7f9fccfae940>\nkwargs = {'copy': True, 'how': 'inner', 'indicator': False, 'left_index': False, ...}\nhow = 'inner', left_index = False, right_index = False, sort = False\nmap_func = <function PandasQueryCompiler.merge.<locals>.map_func at 0x7f9fd4036dc0>\nnew_columns = Index(['urhjs__ipv4_public', '19caY__sha1', 'uqdyN__bothify',\n 'Fpmv9__credit_card_number', 'cTFks__ipv4_private... 'NN7sX__day_of_month', 'Z2GvM__safe_color_name', 'RZGm2__pyfloat',\n 'JW1cO__address'],\n dtype='object')\nnew_dtypes = ModinDtypes:\n\tvalue type: <class 'modin.core.dataframe.pandas.metadata.dtypes.DtypesDescriptor'>;\n\tvalue:\n\tDtypesDescr...graph', '7tHsJ__sentence'];\n\tschema is known: False;\n\thas parent df: False;\n\tcolumns order: None;\n\tknow all names: True\n\n def merge(self, right, **kwargs):\n how = kwargs.get(\"how\", \"inner\")\n on = kwargs.get(\"on\", None)\n left_on = kwargs.get(\"left_on\", None)\n right_on = kwargs.get(\"right_on\", None)\n left_index = kwargs.get(\"left_index\", False)\n right_index = kwargs.get(\"right_index\", False)\n sort = kwargs.get(\"sort\", False)\n \n if how in [\"left\", \"inner\"] and left_index is False and right_index is False:\n kwargs[\"sort\"] = False\n \n def should_keep_index(left, right):\n keep_index = False\n if left_on is not None and right_on is not None:\n keep_index = any(\n o in left.index.names\n and o in right_on\n and o in right.index.names\n for o in left_on\n )\n elif on is not None:\n keep_index = any(\n o in left.index.names and o in right.index.names for o in on\n )\n return keep_index\n \n def map_func(\n left, right, *axis_lengths, kwargs=kwargs, **service_kwargs\n ): # pragma: no cover\n df = pandas.merge(left, right, **kwargs)\n \n if kwargs[\"how\"] == \"left\":\n partition_idx = service_kwargs[\"partition_idx\"]\n if len(axis_lengths):\n if not should_keep_index(left, right):\n # Doesn't work for \"inner\" case, since the partition sizes of the\n # left dataframe may change\n start = sum(axis_lengths[:partition_idx])\n stop = sum(axis_lengths[: partition_idx + 1])\n \n df.index = pandas.RangeIndex(start, stop)\n \n return df\n \n # Want to ensure that these are python lists\n if left_on is not None and right_on is not None:\n left_on = list(left_on) if is_list_like(left_on) else [left_on]\n right_on = list(right_on) if is_list_like(right_on) else [right_on]\n elif on is not None:\n on = list(on) if is_list_like(on) else [on]\n \n new_columns = None\n new_dtypes = None\n if self._modin_frame.has_materialized_columns:\n if left_on is None and right_on is None:\n if on is None:\n on = [c for c in self.columns if c in right.columns]\n _left_on, _right_on = on, on\n else:\n if left_on is None or right_on is None:\n raise MergeError(\n \"Must either pass only 'on' or 'left_on' and 'right_on', not combination of them.\"\n )\n _left_on, _right_on = left_on, right_on\n \n try:\n new_columns, left_renamer, right_renamer = join_columns(\n self.columns,\n right.columns,\n _left_on,\n _right_on,\n kwargs.get(\"suffixes\", (\"_x\", \"_y\")),\n )\n except NotImplementedError:\n # This happens when one of the keys to join is an index level. Pandas behaviour\n # is really complicated in this case, so we're not computing resulted columns for now.\n pass\n else:\n # renamers may contain columns from 'index', so trying to merge index and column dtypes here\n right_index_dtypes = (\n right.index.dtypes\n if isinstance(right.index, pandas.MultiIndex)\n else pandas.Series(\n [right.index.dtype], index=[right.index.name]\n )\n )\n right_dtypes = pandas.concat([right.dtypes, right_index_dtypes])[\n right_renamer.keys()\n ].rename(right_renamer)\n \n left_index_dtypes = (\n self._modin_frame._index_cache.maybe_get_dtypes()\n )\n left_dtypes = (\n ModinDtypes.concat(\n [self._modin_frame._dtypes, left_index_dtypes]\n )\n .lazy_get(left_renamer.keys())\n .set_index(list(left_renamer.values()))\n )\n new_dtypes = ModinDtypes.concat([left_dtypes, right_dtypes])\n \n new_self = self.__constructor__(\n self._modin_frame.broadcast_apply_full_axis(\n axis=1,\n func=map_func,\n enumerate_partitions=how == \"left\",\n other=right._modin_frame,\n # We're going to explicitly change the shape across the 1-axis,\n # so we want for partitioning to adapt as well\n keep_partitioning=False,\n num_splits=merge_partitioning(\n self._modin_frame, right._modin_frame, axis=1\n ),\n new_columns=new_columns,\n sync_labels=False,\n dtypes=new_dtypes,\n pass_axis_lengths_to_partitions=how == \"left\",\n )\n )\n \n # Here we want to understand whether we're joining on a column or on an index level.\n # It's cool if indexes are already materialized so we can easily check that, if not\n # it's fine too, we can also decide that by columns, which tend to be already\n # materialized quite often compared to the indexes.\n keep_index = False\n if self._modin_frame.has_materialized_index:\n> keep_index = should_keep_index(self, right_pandas)\nE NameError: name 'right_pandas' is not defined\n\nmodin/core/storage_formats/pandas/query_compiler.py:643: NameError\n----------------------------- Captured stderr call -----------------------------\n2024-01-10 15:52:55,130\tINFO worker.py:1664 -- Started a local Ray instance. View the dashboard at \u001b[1m\u001b[32m127.0.0.1:8265 \u001b[39m\u001b[22m\n------------------------------ Captured log call -------------------------------\nINFO fuzzydata.core.workflow:workflow.py:53 Creating new Workflow cc54e466\nINFO fuzzydata.core.generator:generator.py:72 Generating base df with 1000 rows and 33 columns\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO fuzzydata.core.generator:generator.py:301 Selected Artifact: Artifact(label=artifact_0), initializing operation chain\nINFO fuzzydata.core.generator:generator.py:339 Chaining Operation: sample\nINFO fuzzydata.core.generator:generator.py:70 Generating right-merge df df with 998 rows and 5 columns\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO fuzzydata.core.generator:generator.py:339 Chaining Operation: merge\nINFO fuzzydata.core.generator:generator.py:365 Executing current operation list: {'sources': ['artifact_0', 'artifact_1'], 'new_label': None, 'op_list': [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]}\nERROR fuzzydata.core.workflow:workflow.py:192 Could not execute Operation: [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]\nERROR fuzzydata.core.generator:generator.py:376 Error during generation, stopping...\nERROR fuzzydata.core.generator:generator.py:377 Was trying to execute operation: {'sources': ['artifact_0', 'artifact_1'], 'new_label': None, 'op_list': [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]} on soruce artifact(s): [Artifact(label=artifact_0), Artifact(label=artifact_1)]\nERROR fuzzydata.core.generator:generator.py:378 Was trying to execute code: self.sources[0].table.sample(frac=0.73).merge(self.sources[1].table, on=\"uqdyN__bothify\")\nERROR fuzzydata.core.generator:generator.py:379 Writing out all files to /tmp/fuzzydata-test-wf-ray//cc54e466/\n=============================== warnings summary ===============================\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google.cloud')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n=========================== short test summary info ============================\nFAILED modin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow - NameError: name 'right_pandas' is not defined\n======================== 1 failed, 6 warnings in 11.54s ========================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/modin/core/storage_formats/pandas/query_compiler.py b/modin/core/storage_formats/pandas/query_compiler.py\nindex 32b379994..7ca5bb8f1 100644\n--- a/modin/core/storage_formats/pandas/query_compiler.py\n+++ b/modin/core/storage_formats/pandas/query_compiler.py\n@@ -640,19 +640,19 @@ class PandasQueryCompiler(BaseQueryCompiler):\n # materialized quite often compared to the indexes.\n keep_index = False\n if self._modin_frame.has_materialized_index:\n- keep_index = should_keep_index(self, right_pandas)\n+ keep_index = should_keep_index(self, right)\n else:\n # Have to trigger columns materialization. Hope they're already available at this point.\n if left_on is not None and right_on is not None:\n keep_index = any(\n- o not in right_pandas.columns\n+ o not in right.columns\n and o in left_on\n and o not in self.columns\n for o in right_on\n )\n elif on is not None:\n keep_index = any(\n- o not in right_pandas.columns and o not in self.columns\n+ o not in right.columns and o not in self.columns\n for o in on\n )\n \n", "difficulty": 2, "changed_files": ["modin/core/storage_formats/pandas/query_compiler.py"], "commit_link": "https://github.com/modin-project/modin/tree/034b1956a0914b739c3e5fc2e19ffcfee0f08c02"}
 
 
data/python/03669a5.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 153, "repo_owner": "mikel-brostrom", "repo_name": "yolo_tracking", "head_branch": "centroid-asso-support", "workflow_name": "CI CPU testing", "workflow_filename": "ci.yml", "workflow_path": ".github/workflows/ci.yml", "contributor": "mikel-brostrom", "sha_fail": "03669a5d72130c57575bedd657b82c601f08a982", "sha_success": "0948605abb9d6d962450f1bbf9a0b9c96c429b29", "workflow": "# name of the workflow, what it is doing (optional)\nname: CI CPU testing\n\n# events that trigger the workflow (required)\non:\n push:\n branches: [master, CIdebug]\n pull_request:\n # pull request where master is target\n branches: [master]\n\nenv:\n # Directory of PyPi package to be tested\n PACKAGE_DIR: boxmot\n # Minimum acceptable test coverage\n # Increase as you add more tests to increase coverage\n COVERAGE_FAIL_UNDER: 29\n\n# the workflow that gets triggerd\njobs:\n build:\n runs-on: ${{ matrix.os }}\n strategy:\n fail-fast: false\n matrix:\n os: [ubuntu-latest] # skip windows-latest for\n python-version: ['3.8', '3.9', '3.10']\n #model: ['yolov8n', 'yolo_nas_s', yolox_n] # yolo models to test\n #tracking-methods: ['deepocsort', 'ocsort', 'botsort', 'strongsort', 'bytetrack'] # tracking methods to test\n\n # Timeout: https://stackoverflow.com/a/59076067/4521646\n timeout-minutes: 50\n steps:\n\n - uses: actions/checkout@v4 # Check out the repository\n - uses: actions/setup-python@v5 # Prepare environment with python 3.9\n with:\n python-version: ${{ matrix.python-version }}\n cache: 'pip' # caching pip dependencies\n - name: Install requirements\n shell: bash # for Windows compatibility\n run: |\n python -m pip install --upgrade pip setuptools wheel\n pip install -e . pytest pytest-cov --extra-index-url https://download.pytorch.org/whl/cpu\n python --version\n pip --version\n pip list\n\n - name: Tests all tracking options\n shell: bash # for Windows compatibility\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n run: |\n # deepocsort fro all supported yolo models\n python examples/track.py --tracking-method deepocsort --source $IMG --imgsz 320\n python examples/track.py --yolo-model yolo_nas_s --tracking-method deepocsort --source $IMG --imgsz 320\n # python examples/track.py --yolo-model yolox_n --tracking-method deepocsort --source $IMG --imgsz 320\n\n # hybridsort\n python examples/track.py --tracking-method hybridsort --source $IMG --imgsz 320\n\n # botsort\n python examples/track.py --tracking-method botsort --source $IMG --imgsz 320\n\n # strongsort\n python examples/track.py --tracking-method strongsort --source $IMG --imgsz 320\n\n # ocsort\n python examples/track.py --tracking-method ocsort --source $IMG --imgsz 320\n\n # bytetrack\n python examples/track.py --tracking-method bytetrack --source $IMG --imgsz 320\n\n - name: Pytest tests # after tracking options as this does not download models\n shell: bash # for Windows compatibility\n run: |\n\n # needed in TFLite export\n wget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\n tar -zxvf flatc.tar.gz\n sudo chmod +x flatc\n sudo mv flatc /usr/bin/\n\n pytest --cov=$PACKAGE_DIR --cov-report=html -v tests\n coverage report --fail-under=$COVERAGE_FAIL_UNDER\n\n - name: Tests exported reid models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n\n # test exported reid model\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17.torchscript --source $IMG --imgsz 320\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17.onnx --source $IMG --imgsz 320\n #python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17_saved_model/osnet_x0_25_msmt17_float16.tflite --source $IMG --imgsz 320\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17_openvino_model --source $IMG --imgsz 320\n\n - name: Test tracking with seg models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n # tracking with SEG models\n python examples/track.py --tracking-method deepocsort --yolo-model yolov8n-seg.pt --source $IMG\n\n - name: Test tracking with pose models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n # tracking with POSE models\n python3 examples/track.py --yolo-model weights/yolov8n.pt --source $IMG --imgsz 320\n\n - name: Test validation on MOT17 subset\n shell: bash # for Windows compatibility\n run: |\n # validation on a few MOT17 imges\n python examples/val.py --tracking-method deepocsort --yolo-model yolov8n.pt --benchmark MOT17-mini --imgsz 320 --conf 0.2\n\n - name: Test evolution on MOT17 subset\n shell: bash # for Windows compatibility\n run: |\n # evolve a for a single set of parameters\n python examples/evolve.py --objectives HOTA,MOTA,IDF1 --benchmark MOT17-mini --n-trials 1 --imgsz 320 --conf 0.2\n", "logs": [{"step_name": "build (ubuntu-latest, 3.8)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n--2024-01-12 20:11:25-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 140.82.112.3\nConnecting to github.com (github.com)|140.82.112.3|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201125Z&X-Amz-Expires=300&X-Amz-Signature=3bfcceeafbe583d085cec55f86a8526c6e9f64b963ff80c4ac3797875e856d28&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:11:25-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201125Z&X-Amz-Expires=300&X-Amz-Signature=3bfcceeafbe583d085cec55f86a8526c6e9f64b963ff80c4ac3797875e856d28&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.110.133, 185.199.108.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.110.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 28.1M 0s\n 50K .......... .......... .......... .......... .......... 7% 36.5M 0s\n 100K .......... .......... .......... .......... .......... 11% 143M 0s\n 150K .......... .......... .......... .......... .......... 14% 42.2M 0s\n 200K .......... .......... .......... .......... .......... 18% 131M 0s\n 250K .......... .......... .......... .......... .......... 22% 167M 0s\n 300K .......... .......... .......... .......... .......... 25% 200M 0s\n 350K .......... .......... .......... .......... .......... 29% 74.0M 0s\n 400K .......... .......... .......... .......... .......... 33% 187M 0s\n 450K .......... .......... .......... .......... .......... 37% 172M 0s\n 500K .......... .......... .......... .......... .......... 40% 132M 0s\n 550K .......... .......... .......... .......... .......... 44% 233M 0s\n 600K .......... .......... .......... .......... .......... 48% 155M 0s\n 650K .......... .......... .......... .......... .......... 51% 208M 0s\n 700K .......... .......... .......... .......... .......... 55% 136M 0s\n 750K .......... .......... .......... .......... .......... 59% 223M 0s\n 800K .......... .......... .......... .......... .......... 62% 136M 0s\n 850K .......... .......... .......... .......... .......... 66% 140M 0s\n 900K .......... .......... .......... .......... .......... 70% 238M 0s\n 950K .......... .......... .......... .......... .......... 74% 241M 0s\n 1000K .......... .......... .......... .......... .......... 77% 242M 0s\n 1050K .......... .......... .......... .......... .......... 81% 215M 0s\n 1100K .......... .......... .......... .......... .......... 85% 238M 0s\n 1150K .......... .......... .......... .......... .......... 88% 240M 0s\n 1200K .......... .......... .......... .......... .......... 92% 236M 0s\n 1250K .......... .......... .......... .......... .......... 96% 198M 0s\n 1300K .......... .......... .......... .......... .......... 99% 251M 0s\n 1350K 100% 572G=0.01s\n\n2024-01-12 20:11:25 (120 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.8.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7f1dba152ca0>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 9%|\u2589 | 2.10M/22.4M [00:00<00:01, 19.5MB/s]\n 40%|\u2588\u2588\u2588\u2589 | 8.91M/22.4M [00:00<00:00, 43.6MB/s]\n 82%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 18.4M/22.4M [00:00<00:00, 65.1MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 64.2MB/s]\n\u001b[32m2024-01-12 20:11:55.585\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:11:55.738\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('mpl_toolkits')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.8.18-final-0 -----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 9 warnings in 31.85s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (ubuntu-latest, 3.9)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n--2024-01-12 20:11:58-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 192.30.255.113\nConnecting to github.com (github.com)|192.30.255.113|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201159Z&X-Amz-Expires=300&X-Amz-Signature=709497883dd2c2a9471410daf12c63e704acb6b779fb216aa16d512cc5577229&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:11:59-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201159Z&X-Amz-Expires=300&X-Amz-Signature=709497883dd2c2a9471410daf12c63e704acb6b779fb216aa16d512cc5577229&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.111.133, 185.199.108.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.111.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 10.7M 0s\n 50K .......... .......... .......... .......... .......... 7% 44.1M 0s\n 100K .......... .......... .......... .......... .......... 11% 89.2M 0s\n 150K .......... .......... .......... .......... .......... 14% 62.8M 0s\n 200K .......... .......... .......... .......... .......... 18% 108M 0s\n 250K .......... .......... .......... .......... .......... 22% 163M 0s\n 300K .......... .......... .......... .......... .......... 25% 186M 0s\n 350K .......... .......... .......... .......... .......... 29% 202M 0s\n 400K .......... .......... .......... .......... .......... 33% 108M 0s\n 450K .......... .......... .......... .......... .......... 37% 199M 0s\n 500K .......... .......... .......... .......... .......... 40% 190M 0s\n 550K .......... .......... .......... .......... .......... 44% 198M 0s\n 600K .......... .......... .......... .......... .......... 48% 171M 0s\n 650K .......... .......... .......... .......... .......... 51% 212M 0s\n 700K .......... .......... .......... .......... .......... 55% 170M 0s\n 750K .......... .......... .......... .......... .......... 59% 185M 0s\n 800K .......... .......... .......... .......... .......... 62% 199M 0s\n 850K .......... .......... .......... .......... .......... 66% 259M 0s\n 900K .......... .......... .......... .......... .......... 70% 191M 0s\n 950K .......... .......... .......... .......... .......... 74% 205M 0s\n 1000K .......... .......... .......... .......... .......... 77% 193M 0s\n 1050K .......... .......... .......... .......... .......... 81% 267M 0s\n 1100K .......... .......... .......... .......... .......... 85% 271M 0s\n 1150K .......... .......... .......... .......... .......... 88% 239M 0s\n 1200K .......... .......... .......... .......... .......... 92% 268M 0s\n 1250K .......... .......... .......... .......... .......... 96% 260M 0s\n 1300K .......... .......... .......... .......... .......... 99% 269M 0s\n 1350K 100% 572G=0.01s\n\n2024-01-12 20:11:59 (103 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.9.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.9.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7ff526351e50>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 5%|\u258d | 1.05M/22.4M [00:00<00:02, 9.56MB/s]\n 21%|\u2588\u2588 | 4.72M/22.4M [00:00<00:00, 21.4MB/s]\n 54%|\u2588\u2588\u2588\u2588\u2588\u258e | 12.1M/22.4M [00:00<00:00, 32.0MB/s]\n 77%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b | 17.3M/22.4M [00:00<00:00, 32.2MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 37.0MB/s]\n\u001b[32m2024-01-12 20:12:31.441\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:12:31.599\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.9.18-final-0 -----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 8 warnings in 33.86s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (ubuntu-latest, 3.10)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n--2024-01-12 20:12:02-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 140.82.113.3\nConnecting to github.com (github.com)|140.82.113.3|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201203Z&X-Amz-Expires=300&X-Amz-Signature=f339441879645b502b6bf0bbc8e04deb0dcdf54ed1dcd223e8a7e04adef107ad&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:12:03-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201203Z&X-Amz-Expires=300&X-Amz-Signature=f339441879645b502b6bf0bbc8e04deb0dcdf54ed1dcd223e8a7e04adef107ad&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.108.133, 185.199.110.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.108.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 5.77M 0s\n 50K .......... .......... .......... .......... .......... 7% 6.18M 0s\n 100K .......... .......... .......... .......... .......... 11% 30.6M 0s\n 150K .......... .......... .......... .......... .......... 14% 39.6M 0s\n 200K .......... .......... .......... .......... .......... 18% 9.54M 0s\n 250K .......... .......... .......... .......... .......... 22% 49.5M 0s\n 300K .......... .......... .......... .......... .......... 25% 65.1M 0s\n 350K .......... .......... .......... .......... .......... 29% 45.1M 0s\n 400K .......... .......... .......... .......... .......... 33% 47.7M 0s\n 450K .......... .......... .......... .......... .......... 37% 86.6M 0s\n 500K .......... .......... .......... .......... .......... 40% 13.9M 0s\n 550K .......... .......... .......... .......... .......... 44% 44.2M 0s\n 600K .......... .......... .......... .......... .......... 48% 74.1M 0s\n 650K .......... .......... .......... .......... .......... 51% 127M 0s\n 700K .......... .......... .......... .......... .......... 55% 88.8M 0s\n 750K .......... .......... .......... .......... .......... 59% 126M 0s\n 800K .......... .......... .......... .......... .......... 62% 178M 0s\n 850K .......... .......... .......... .......... .......... 66% 77.3M 0s\n 900K .......... .......... .......... .......... .......... 70% 209M 0s\n 950K .......... .......... .......... .......... .......... 74% 180M 0s\n 1000K .......... .......... .......... .......... .......... 77% 159M 0s\n 1050K .......... .......... .......... .......... .......... 81% 14.6M 0s\n 1100K .......... .......... .......... .......... .......... 85% 52.9M 0s\n 1150K .......... .......... .......... .......... .......... 88% 186M 0s\n 1200K .......... .......... .......... .......... .......... 92% 127M 0s\n 1250K .......... .......... .......... .......... .......... 96% 210M 0s\n 1300K .......... .......... .......... .......... .......... 99% 216M 0s\n 1350K 100% 8.87M=0.04s\n\n2024-01-12 20:12:03 (31.1 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.10.13, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.10.13/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7f20a69f0700>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 2%|\u258f | 524k/22.4M [00:00<00:05, 3.80MB/s]\n 35%|\u2588\u2588\u2588\u258c | 7.86M/22.4M [00:00<00:00, 38.7MB/s]\n 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258d | 16.8M/22.4M [00:00<00:00, 58.3MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 54.7MB/s]\n\u001b[32m2024-01-12 20:12:35.371\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:12:35.534\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.10.13-final-0 ----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 8 warnings in 33.78s ===================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/boxmot/utils/iou.py b/boxmot/utils/iou.py\nindex d8bfc7e..144f40f 100644\n--- a/boxmot/utils/iou.py\n+++ b/boxmot/utils/iou.py\n@@ -200,10 +200,10 @@ def run_asso_func(func, *args):\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n- if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n+ if func in (iou_batch, giou_batch, diou_batch, ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n- return func(*args)\n+ return func(*args[0:2])\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n", "difficulty": 2, "changed_files": ["boxmot/utils/iou.py"], "commit_link": "https://github.com/mikel-brostrom/yolo_tracking/tree/03669a5d72130c57575bedd657b82c601f08a982"}
 
 
data/python/06bc6b6.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 93, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "composer_lora", "workflow_name": "Code Quality Checks", "workflow_filename": "code-quality.yaml", "workflow_path": ".github/workflows/code-quality.yaml", "contributor": "dakinggg", "sha_fail": "06bc6b627d6afb938950382ba50b6b71432d7cf6", "sha_success": "23961743885e704b3f4bd77ba8410b2508d5127c", "workflow": "name: Code Quality Checks\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n code-quality:\n runs-on: ubuntu-20.04\n timeout-minutes: 15\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n pip_deps:\n - \"[dev]\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .${{ matrix.pip_deps }}\n - name: Run checks\n run: |\n pre-commit run --all-files\n", "logs": [{"step_name": "code-quality (3.9, [dev])/5_Run checks.txt", "log": "##[group]Run pre-commit run --all-files\n\u001b[36;1mpre-commit run --all-files\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n[INFO] Initializing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Initializing environment for https://github.com/google/yapf.\n[INFO] Initializing environment for https://github.com/google/yapf:toml.\n[INFO] Initializing environment for https://github.com/pycqa/isort.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle:toml.\n[INFO] Initializing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/kynan/nbstripout.\n[INFO] Initializing environment for local:[email protected].\n[INFO] Initializing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Installing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/google/yapf.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pycqa/isort.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/kynan/nbstripout.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for local.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\nruff.....................................................................Passed\nyapf.....................................................................Passed\nisort....................................................................Failed\n- hook id: isort\n- files were modified by this hook\n\nFixing /home/runner/work/composer/composer/tests/models/test_hf_model.py\nSkipped 1 files\n\npydocstyle...............................................................Passed\ncheck for added large files..............................................Passed\ncheck python ast.........................................................Passed\ncheck builtin type constructor use.......................................Passed\ncheck for case conflicts.................................................Passed\ncheck docstring is first.................................................Passed\ncheck that executables have shebangs.....................................Passed\ncheck json...............................................................Passed\ncheck that scripts with shebangs are executable..........................Passed\npretty format json.......................................................Passed\ncheck for merge conflicts................................................Passed\ncheck for broken symlinks................................................Passed\ncheck toml...............................................................Passed\ncheck vcs permalinks.....................................................Passed\ncheck xml................................................................Passed\ncheck yaml...............................................................Passed\ndebug statements (python)................................................Passed\ndetect destroyed symlinks................................................Passed\nfix double quoted strings................................................Passed\nfix end of files.........................................................Passed\nfix utf-8 byte order marker..............................................Passed\nmixed line ending........................................................Passed\ntrim trailing whitespace.................................................Passed\nInsert license in comments...............................................Passed\nnbstripout...............................................................Passed\npyright..................................................................Passed\nsecret scan..............................................................Passed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/composer/models/huggingface.py b/composer/models/huggingface.py\nindex f3855ce6..c223a917 100644\n--- a/composer/models/huggingface.py\n+++ b/composer/models/huggingface.py\n@@ -111,6 +111,13 @@ class HuggingFaceModel(ComposerModel):\n conda_package='peft',\n conda_channel='conda-forge')\n \n+ if peft_config is not None:\n+ # Hugging Face requires the peft type and task type to be upper case, so we do that here\n+ # https://github.com/huggingface/peft/blob/ebbff4023ad276cbcb2466fd7e99be7d3ae0ae11/src/peft/utils/peft_types.py#L22-L51\n+ if isinstance(peft_config.peft_type, str):\n+ peft_config.peft_type = peft_config.peft_type.upper()\n+ if isinstance(peft_config.task_type, str):\n+ peft_config.task_type = peft_config.task_type.upper()\n if peft_config is not None and peft_config.peft_type != 'LORA':\n raise ValueError(\n f'PEFT type {peft_config.peft_type} is not supported by HuggingFaceModel. Only LORA is supported.')\ndiff --git a/tests/models/test_hf_model.py b/tests/models/test_hf_model.py\nindex c3c59e1c..a69b56b5 100644\n--- a/tests/models/test_hf_model.py\n+++ b/tests/models/test_hf_model.py\n@@ -28,8 +28,6 @@ from tests.common.markers import device, world_size\n from tests.common.models import (configure_tiny_bert_model, configure_tiny_bert_tokenizer, configure_tiny_gpt2_model,\n configure_tiny_gpt2_tokenizer, configure_tiny_mistral_model,\n configure_tiny_mistral_tokenizer, configure_tiny_t5_model, configure_tiny_t5_tokenizer)\n-from tests.common.models import (configure_tiny_bert_model, configure_tiny_bert_tokenizer, configure_tiny_gpt2_model,\n- configure_tiny_gpt2_tokenizer, configure_tiny_t5_model, configure_tiny_t5_tokenizer)\n from tests.loggers.test_remote_uploader_downloader import DummyObjectStore\n \n if TYPE_CHECKING:\n", "difficulty": 0, "changed_files": ["composer/models/huggingface.py", "tests/models/test_hf_model.py"], "commit_link": "https://github.com/mosaicml/composer/tree/06bc6b627d6afb938950382ba50b6b71432d7cf6"}
 
 
data/python/077f6aa.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 116, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "main", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "c-core-labs", "sha_fail": "077f6aaac3ebb96626ac747fb126a0b4d752489c", "sha_success": "4c315803bfdb46fa7818cb08fd7224fb0bb599f1", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 26.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 90.4 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 116.1 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 81.8 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 92.1 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|22e1ac8b31209d20dbc1af062b7bb70cb716a8ce5ddbf255b52ad3a34004f0dc\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nReceived 0 of 48057368 (0.0%), 0.0 MBs/sec\nCache Size: ~46 MB (48057368 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/516a3dc9-7b6f-410a-9989-4b9d88d8a69b/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48057368 of 48057368 (100.0%), 22.9 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|22e1ac8b31209d20dbc1af062b7bb70cb716a8ce5ddbf255b52ad3a34004f0dc\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nruff.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: ruff\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\nFound 1 error (1 fixed, 0 remaining).\n\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted wandb/integration/ultralytics/callback.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m870 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/wandb/integration/ultralytics/bbox_utils.py b/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[1mindex 798aa57..9cee915 100644\u001b[m\n\u001b[1m--- a/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[1m+++ b/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[36m@@ -3,6 +3,7 @@\u001b[m \u001b[mfrom typing import Any, Dict, List, Optional, Tuple, Union\u001b[m\n import torch\u001b[m\n from ultralytics.engine.results import Results\u001b[m\n from ultralytics.models.yolo.detect import DetectionPredictor\u001b[m\n\u001b[32m+\u001b[m\n try:\u001b[m\n from ultralytics.yolo.utils import ops\u001b[m\n except ModuleNotFoundError:\u001b[m\n\u001b[1mdiff --git a/wandb/integration/ultralytics/callback.py b/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[1mindex ed8a224..618fc5a 100644\u001b[m\n\u001b[1m--- a/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[1m+++ b/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[36m@@ -42,6 +42,7 @@\u001b[m \u001b[mtry:\u001b[m\n SegmentationValidator,\u001b[m\n )\u001b[m\n from ultralytics.utils.torch_utils import de_parallel\u001b[m\n\u001b[32m+\u001b[m\n try:\u001b[m\n from ultralytics.yolo.utils import RANK, __version__\u001b[m\n except ModuleNotFoundError:\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/wandb/cli/cli.py b/wandb/cli/cli.py\nindex 94a3ece90..1d0031488 100644\n--- a/wandb/cli/cli.py\n+++ b/wandb/cli/cli.py\n@@ -1496,6 +1496,8 @@ def launch(\n if cli_template_vars:\n if queue is None:\n raise LaunchError(\"'--set-var' flag requires queue to be set\")\n+ if entity is None:\n+ entity = launch_utils.get_default_entity(api, config)\n public_api = PublicApi()\n runqueue = RunQueue(client=public_api.client, name=queue, entity=entity)\n template_variables = launch_utils.fetch_and_validate_template_variables(\ndiff --git a/wandb/integration/ultralytics/bbox_utils.py b/wandb/integration/ultralytics/bbox_utils.py\nindex 798aa573a..8b924697d 100644\n--- a/wandb/integration/ultralytics/bbox_utils.py\n+++ b/wandb/integration/ultralytics/bbox_utils.py\n@@ -3,10 +3,7 @@ from typing import Any, Dict, List, Optional, Tuple, Union\n import torch\n from ultralytics.engine.results import Results\n from ultralytics.models.yolo.detect import DetectionPredictor\n-try:\n- from ultralytics.yolo.utils import ops\n-except ModuleNotFoundError:\n- from ultralytics.utils import ops\n+from ultralytics.yolo.utils import ops\n \n import wandb\n \ndiff --git a/wandb/integration/ultralytics/callback.py b/wandb/integration/ultralytics/callback.py\nindex ed8a22465..2e133a163 100644\n--- a/wandb/integration/ultralytics/callback.py\n+++ b/wandb/integration/ultralytics/callback.py\n@@ -42,10 +42,7 @@ try:\n SegmentationValidator,\n )\n from ultralytics.utils.torch_utils import de_parallel\n- try:\n- from ultralytics.yolo.utils import RANK, __version__\n- except ModuleNotFoundError:\n- from ultralytics.utils import RANK, __version__\n+ from ultralytics.yolo.utils import RANK, __version__\n \n from wandb.integration.ultralytics.bbox_utils import (\n plot_predictions,\ndiff --git a/wandb/integration/yolov8/yolov8.py b/wandb/integration/yolov8/yolov8.py\nindex 1c5dbd0fe..310a39114 100644\n--- a/wandb/integration/yolov8/yolov8.py\n+++ b/wandb/integration/yolov8/yolov8.py\n@@ -2,13 +2,8 @@ from typing import Any, Callable, Dict, List, Optional\n \n from ultralytics.yolo.engine.model import YOLO\n from ultralytics.yolo.engine.trainer import BaseTrainer\n-\n-try:\n- from ultralytics.yolo.utils import RANK\n- from ultralytics.yolo.utils.torch_utils import get_flops, get_num_params\n-except ModuleNotFoundError:\n- from ultralytics.utils import RANK\n- from ultralytics.utils.torch_utils import get_flops, get_num_params\n+from ultralytics.yolo.utils import RANK\n+from ultralytics.yolo.utils.torch_utils import get_flops, get_num_params\n from ultralytics.yolo.v8.classify.train import ClassificationTrainer\n \n import wandb\ndiff --git a/wandb/sdk/launch/utils.py b/wandb/sdk/launch/utils.py\nindex 05efe128e..69f4e57bb 100644\n--- a/wandb/sdk/launch/utils.py\n+++ b/wandb/sdk/launch/utils.py\n@@ -204,10 +204,7 @@ def set_project_entity_defaults(\n config_project = launch_config.get(\"project\")\n project = config_project or source_uri or \"\"\n if entity is None:\n- config_entity = None\n- if launch_config:\n- config_entity = launch_config.get(\"entity\")\n- entity = config_entity or api.default_entity\n+ entity = get_default_entity(api, launch_config)\n prefix = \"\"\n if platform.system() != \"Windows\" and sys.stdout.encoding == \"UTF-8\":\n prefix = \"\ud83d\ude80 \"\n@@ -217,6 +214,13 @@ def set_project_entity_defaults(\n return project, entity\n \n \n+def get_default_entity(api: Api, launch_config: Optional[Dict[str, Any]]):\n+ config_entity = None\n+ if launch_config:\n+ config_entity = launch_config.get(\"entity\")\n+ return config_entity or api.default_entity\n+\n+\n def construct_launch_spec(\n uri: Optional[str],\n job: Optional[str],\n", "difficulty": 0, "changed_files": ["wandb/cli/cli.py", "wandb/integration/ultralytics/bbox_utils.py", "wandb/integration/ultralytics/callback.py", "wandb/integration/yolov8/yolov8.py", "wandb/sdk/launch/utils.py"], "commit_link": "https://github.com/wandb/wandb/tree/077f6aaac3ebb96626ac747fb126a0b4d752489c"}
 
 
data/python/07bc10c.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 189, "repo_owner": "openvinotoolkit", "repo_name": "open_model_zoo", "head_branch": "ac_sub_evaluation", "workflow_name": "github-actions", "workflow_filename": "accuracy_checker.yml", "workflow_path": ".github/workflows/accuracy_checker.yml", "contributor": "pwolnows", "sha_fail": "07bc10c0e7858b22e9345812af8e6bb6c4ef18be", "sha_success": "76a4c47015ae5c1280dcf23d3d8592df40dbced8", "workflow": "name: github-actions\non: pull_request\njobs:\n accuracy_checker:\n strategy:\n matrix:\n runs-on: [ubuntu-20.04, ubuntu-22.04]\n python-version: [3.8, 3.9, '3.10']\n runs-on: ${{ matrix.runs-on }}\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n - name: Pip install dependencies\n run: |\n python -m pip install --upgrade pip\n python -m pip install tools/accuracy_checker openvino -r tools/accuracy_checker/requirements-test.in\n - name: python -m pytest tools/accuracy_checker\n run: |\n python -m pytest tools/accuracy_checker\n - name: Pip install and run pylint\n working-directory: ./tools/accuracy_checker\n run: |\n python -m pip install pylint==2.10.2\n PYTHONPATH=. python -m pylint --rcfile=.pylintrc `find -wholename '?*/**/*.py' -not -path \"./tests/*\" -not -path \"./build/*\"`\n", "logs": [{"step_name": "accuracy_checker (ubuntu-20.04, 3.8)/6_Pip install and run pylint.txt", "log": "##[group]Run python -m pip install pylint==2.10.2\n\u001b[36;1mpython -m pip install pylint==2.10.2\u001b[0m\n\u001b[36;1mPYTHONPATH=. python -m pylint --rcfile=.pylintrc `find -wholename '?*/**/*.py' -not -path \"./tests/*\" -not -path \"./build/*\"`\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nCollecting pylint==2.10.2\n Downloading pylint-2.10.2-py3-none-any.whl (392 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 392.6/392.6 kB 15.3 MB/s eta 0:00:00\nCollecting platformdirs>=2.2.0 (from pylint==2.10.2)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting astroid<2.8,>=2.7.2 (from pylint==2.10.2)\n Downloading astroid-2.7.3-py3-none-any.whl (240 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 240.6/240.6 kB 46.7 MB/s eta 0:00:00\nCollecting isort<6,>=4.2.5 (from pylint==2.10.2)\n Downloading isort-5.13.2-py3-none-any.whl.metadata (12 kB)\nCollecting mccabe<0.7,>=0.6 (from pylint==2.10.2)\n Downloading mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\nCollecting toml>=0.7.1 (from pylint==2.10.2)\n Downloading toml-0.10.2-py2.py3-none-any.whl (16 kB)\nCollecting lazy-object-proxy>=1.4.0 (from astroid<2.8,>=2.7.2->pylint==2.10.2)\n Downloading lazy_object_proxy-1.10.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.8 kB)\nCollecting wrapt<1.13,>=1.11 (from astroid<2.8,>=2.7.2->pylint==2.10.2)\n Downloading wrapt-1.12.1.tar.gz (27 kB)\n Installing build dependencies: started\n Installing build dependencies: finished with status 'done'\n Getting requirements to build wheel: started\n Getting requirements to build wheel: finished with status 'done'\n Preparing metadata (pyproject.toml): started\n Preparing metadata (pyproject.toml): finished with status 'done'\nRequirement already satisfied: setuptools>=20.0 in /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages (from astroid<2.8,>=2.7.2->pylint==2.10.2) (56.0.0)\nDownloading isort-5.13.2-py3-none-any.whl (92 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 92.3/92.3 kB 21.2 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading lazy_object_proxy-1.10.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (66 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 66.4/66.4 kB 16.7 MB/s eta 0:00:00\nBuilding wheels for collected packages: wrapt\n Building wheel for wrapt (pyproject.toml): started\n Building wheel for wrapt (pyproject.toml): finished with status 'done'\n Created wheel for wrapt: filename=wrapt-1.12.1-cp38-cp38-linux_x86_64.whl size=81820 sha256=be1d80d6ab9cc886aa44b352ed5ead61a5f3ec49ae66639ac53a214d9cb2af26\n Stored in directory: /home/runner/.cache/pip/wheels/5f/fd/9e/b6cf5890494cb8ef0b5eaff72e5d55a70fb56316007d6dfe73\nSuccessfully built wrapt\nInstalling collected packages: wrapt, mccabe, toml, platformdirs, lazy-object-proxy, isort, astroid, pylint\nSuccessfully installed astroid-2.7.3 isort-5.13.2 lazy-object-proxy-1.10.0 mccabe-0.6.1 platformdirs-4.1.0 pylint-2.10.2 toml-0.10.2 wrapt-1.12.1\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n************* Module openvino.tools.accuracy_checker.evaluators.model_evaluator\nopenvino/tools/accuracy_checker/evaluators/model_evaluator.py:805:0: C0303: Trailing whitespace (trailing-whitespace)\n\n------------------------------------\nYour code has been rated at 10.00/10\n\n##[error]Process completed with exit code 16.\n"}], "diff": "diff --git a/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py b/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\nindex adb41abf5..5a310a979 100644\n--- a/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\n+++ b/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\n@@ -802,5 +802,4 @@ def get_config_metrics(config):\n # first subset_metrics or matching subsample_size\n metrics = item.get('metrics')\n break\n- \n return config.get('metrics',[]) if (metrics is None) else metrics\n", "difficulty": 0, "changed_files": ["tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py"], "commit_link": "https://github.com/openvinotoolkit/open_model_zoo/tree/07bc10c0e7858b22e9345812af8e6bb6c4ef18be"}
 
 
data/python/0828c8d.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/08c47c4.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 118, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "WB-11095", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "wandb", "sha_fail": "08c47c420629666a284cb013a39c907ac7451af5", "sha_success": "143c1ae682a944d2bfc8ebe9fb6ef7368cc58797", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.3-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 6.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 27.8 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 33.6 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 97.3 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 74.3 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.3-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 96.0 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.3 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.3\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCache Size: ~47 MB (48931996 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/b3bd0166-47b9-4cba-b7da-654927a1ab90/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48931996 of 48931996 (100.0%), 46.6 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nclippy...................................................................\u001b[42mPassed\u001b[m\nruff.....................................................................\u001b[42mPassed\u001b[m\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted tests/pytest_tests/unit_tests/test_public_api.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m887 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/tests/pytest_tests/unit_tests/test_public_api.py b/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[1mindex 6fb0799..bb0190b 100644\u001b[m\n\u001b[1m--- a/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[1m+++ b/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[36m@@ -97,6 +97,7 @@\u001b[m \u001b[mdef test_parse_path_proj():\u001b[m\n assert project == \"proj\"\u001b[m\n assert run == \"proj\"\u001b[m\n \u001b[m\n\u001b[32m+\u001b[m\n @pytest.mark.usefixtures(\"patch_apikey\", \"patch_prompt\")\u001b[m\n def test_parse_path_id():\u001b[m\n with mock.patch.dict(\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/tests/pytest_tests/unit_tests/test_public_api.py b/tests/pytest_tests/unit_tests/test_public_api.py\nindex 6fb079974..bb0190be6 100644\n--- a/tests/pytest_tests/unit_tests/test_public_api.py\n+++ b/tests/pytest_tests/unit_tests/test_public_api.py\n@@ -97,6 +97,7 @@ def test_parse_path_proj():\n assert project == \"proj\"\n assert run == \"proj\"\n \n+\n @pytest.mark.usefixtures(\"patch_apikey\", \"patch_prompt\")\n def test_parse_path_id():\n with mock.patch.dict(\n", "difficulty": 0, "changed_files": ["tests/pytest_tests/unit_tests/test_public_api.py"], "commit_link": "https://github.com/wandb/wandb/tree/08c47c420629666a284cb013a39c907ac7451af5"}
 
 
data/python/0b08b8e.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/0b5c0d2.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/0b93d2d.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/0d26cc1.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 38, "repo_owner": "beetbox", "repo_name": "beets", "head_branch": "master", "workflow_name": "Check code formatting", "workflow_filename": "formatting_check.yml", "workflow_path": ".github/workflows/formatting_check.yml", "contributor": "niacdoial", "sha_fail": "0d26cc1482ff5080ec579b17b29f22657a20c562", "sha_success": "828c6e7dbd1a1fe77b4b7be61aa2e3d220ff86d1", "workflow": "name: formatting_check\nrun-name: Check code formatting\non: [push, pull_request]\njobs:\n formatting_check:\n runs-on: ubuntu-latest\n steps:\n - name: Install dependencies\n uses: actions/checkout@v3\n\n - name: Run formatting check\n uses: paolorechia/[email protected]\n with:\n tox_env: \"format_check\"\n", "logs": [{"step_name": "formatting_check/3_Run formatting check.txt", "log": "##[group]Run paolorechia/[email protected]\nwith:\n tox_env: format_check\n##[endgroup]\nstdout: Defaulting to user installation because normal site-packages is not writeable\n\nstdout: Collecting tox\n\nstdout: Downloading tox-4.11.4-py3-none-any.whl (153 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 KB 4.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Requirement already satisfied: platformdirs>=3.10 in /usr/local/lib/python3.10/dist-packages (from tox) (4.1.0)\n\nstdout: Collecting pluggy>=1.3\n\nstdout: Downloading pluggy-1.3.0-py3-none-any.whl (18 kB)\n\nstdout: Collecting virtualenv>=20.24.3\n\nstdout: Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 12.0 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting cachetools>=5.3.1\n\nstdout: Downloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\n\nstdout: Requirement already satisfied: packaging>=23.1 in /usr/local/lib/python3.10/dist-packages (from tox) (23.2)\n\nstdout: Requirement already satisfied: tomli>=2.0.1 in /usr/local/lib/python3.10/dist-packages (from tox) (2.0.1)\n\nstdout: Collecting filelock>=3.12.3\n\nstdout: Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\n\nstdout: Collecting colorama>=0.4.6\n\nstdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\n\nstdout: Collecting pyproject-api>=1.6.1\n\nstdout: Downloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\n\nstdout: Collecting chardet>=5.2\n\nstdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 KB 21.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting distlib<1,>=0.3.7\n\nstdout: Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 15.4 MB/s eta 0:00:00\nstdout: \n\nstdout: Installing collected packages: distlib, pyproject-api, pluggy, filelock, colorama, chardet, cachetools, virtualenv, tox\n\nstdout: Successfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\nReceived 6565789 of 6565789 (100.0%), 18.9 MBs/sec\nCache Size: ~6 MB (6565789 B)\n[command]/usr/bin/tar -z -xf /home/runner/work/_temp/d6197f94-ef69-4efe-9ad2-cb7265f9e34d/cache.tgz -P -C /home/runner/work/beets/beets\nCache restored successfully\nHit cache on tox\nstdout: format_check: recreate env because python changed virtualenv version='20.24.5'->'20.25.0'\n\nstdout: format_check: remove tox env folder /home/runner/work/beets/beets/.tox/format_check\n\nstdout: format_check: install_deps> python -I -m pip install black isort\n\nstdout: format_check: commands[0]> isort beets beetsplug test --check\n\nstderr: ERROR: /home/runner/work/beets/beets/beetsplug/edit.py Imports are incorrectly sorted and/or formatted.\n\nstdout: format_check: exit 1 (0.54 seconds) /home/runner/work/beets/beets> isort beets beetsplug test --check pid=1759\n\nstdout: format_check: FAIL code 1 (3.49=setup[2.95]+cmd[0.54] seconds)\n evaluation failed :( (3.64 seconds)\n\nchild process exited with code 1\n##[warning]The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/\n##[error]Tox failed\n"}], "diff": "diff --git a/beetsplug/advancedrewrite.py b/beetsplug/advancedrewrite.py\nindex 639aa5247..20f2b7e03 100644\n--- a/beetsplug/advancedrewrite.py\n+++ b/beetsplug/advancedrewrite.py\n@@ -161,7 +161,7 @@ class AdvancedRewritePlugin(BeetsPlugin):\n )\n elif isinstance(replacement, str):\n if Item._fields[fieldname] is MULTI_VALUE_DSV:\n- replacement = list(replacement)\n+ replacement = [replacement]\n else:\n raise UserError(\n f\"Invalid type of replacement {replacement} \"\ndiff --git a/beetsplug/edit.py b/beetsplug/edit.py\nindex 8a07fccaa..323dd9e41 100644\n--- a/beetsplug/edit.py\n+++ b/beetsplug/edit.py\n@@ -23,7 +23,7 @@ from tempfile import NamedTemporaryFile\n \n import yaml\n \n-from beets import plugins, ui, util, config\n+from beets import plugins, ui, util\n from beets.dbcore import types\n from beets.importer import action\n from beets.ui.commands import PromptChoice, _do_query\n@@ -155,8 +155,6 @@ class EditPlugin(plugins.BeetsPlugin):\n }\n )\n \n- self.has_shown_ui = False\n-\n self.register_listener(\n \"before_choose_candidate\", self.before_choose_candidate_listener\n )\n@@ -238,23 +236,11 @@ class EditPlugin(plugins.BeetsPlugin):\n # Get the content to edit as raw data structures.\n old_data = [flatten(o, fields) for o in objs]\n \n- # take set fields into account\n- set_fields = config[\"import\"][\"set_fields\"]\n- if set_fields and not self.has_shown_ui:\n- old_str = \"\\n\\n# note: the following fields will be reset to their current values:\\n\"\n- for key in set_fields:\n- old_str += f\"# - {key}\\n\"\n- for obj in old_data:\n- # those values will be enforced later anyway\n- obj.update({k:v.get() for k,v in set_fields.items()})\n- else:\n- old_str = \"\"\n-\n # Set up a temporary file with the initial data for editing.\n new = NamedTemporaryFile(\n mode=\"w\", suffix=\".yaml\", delete=False, encoding=\"utf-8\"\n )\n- old_str = dump(old_data) + old_str\n+ old_str = dump(old_data)\n new.write(old_str)\n new.close()\n \n@@ -263,7 +249,6 @@ class EditPlugin(plugins.BeetsPlugin):\n while True:\n # Ask the user to edit the data.\n edit(new.name, self._log)\n- self.has_shown_ui = True\n \n # Read the data back after editing and check whether anything\n # changed.\ndiff --git a/beetsplug/web/__init__.py b/beetsplug/web/__init__.py\nindex cebb0be0a..cd7e8a3fc 100644\n--- a/beetsplug/web/__init__.py\n+++ b/beetsplug/web/__init__.py\n@@ -336,7 +336,6 @@ def item_file(item_id):\n response = flask.send_file(\n item_path, as_attachment=True, download_name=safe_filename\n )\n- response.headers[\"Content-Length\"] = os.path.getsize(item_path)\n return response\n \n \ndiff --git a/docs/changelog.rst b/docs/changelog.rst\nindex 73cc10f3d..b3b63f1c2 100644\n--- a/docs/changelog.rst\n+++ b/docs/changelog.rst\n@@ -279,6 +279,8 @@ Bug fixes:\n :bug:`4973`\n * Fix bug regarding displaying tracks that have been changed not being\n displayed unless the detail configuration is enabled.\n+* :doc:`/plugins/web`: Fix range request support, allowing to play large audio/\n+ opus files using e.g. a browser/firefox or gstreamer/mopidy directly.\n \n For plugin developers:\n \n", "difficulty": 0, "changed_files": ["beetsplug/advancedrewrite.py", "beetsplug/edit.py", "beetsplug/web/__init__.py", "docs/changelog.rst"], "commit_link": "https://github.com/beetbox/beets/tree/0d26cc1482ff5080ec579b17b29f22657a20c562"}
 
 
data/python/0d367ce.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/0f71221.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 20, "repo_owner": "scrapy", "repo_name": "scrapy", "head_branch": "component-getters", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "Gallaecio", "sha_fail": "0f71221cf9875ed8ef3400e1008408e79b6691e6", "sha_success": "c1ba9ccdf916b89d875628ba143dc5c9f6977430", "workflow": "name: Checks\non: [push, pull_request]\n\nconcurrency:\n group: ${{github.workflow}}-${{ github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n include:\n - python-version: \"3.12\"\n env:\n TOXENV: pylint\n - python-version: 3.8\n env:\n TOXENV: typing\n - python-version: \"3.11\" # Keep in sync with .readthedocs.yml\n env:\n TOXENV: docs\n - python-version: \"3.12\"\n env:\n TOXENV: twinecheck\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Run check\n env: ${{ matrix.env }}\n run: |\n pip install -U tox\n tox\n\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "checks (3.12, pylint)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n TOXENV: pylint\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 16.6 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 48.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 16.1 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 104.2 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 81.3 MB/s eta 0:00:00\nInstalling collected packages: distlib, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.3.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\npylint: install_deps> python -I -m pip install -ctests/upper-constraints.txt boto3 google-cloud-storage 'markupsafe<2.1.0' 'mitmproxy<8,>=4.0.4; python_version < \"3.9\" and implementation_name != \"pypy\"' Pillow pylint==3.0.1 robotexclusionrulesparser 'Twisted[http2]' -r tests/requirements.txt\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\npylint: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\npylint: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\npylint: freeze> python -m pip freeze --all\npylint: astroid==3.0.2,asttokens==2.4.1,attrs==23.1.0,Automat==22.10.0,blessed==1.20.0,boto3==1.34.3,botocore==1.34.3,bpython==0.24,Brotli==1.1.0,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,coverage==7.3.3,cryptography==41.0.7,cssselect==1.2.0,curtsies==0.4.2,cwcwidth==0.1.9,decorator==5.1.1,dill==0.3.7,execnet==2.0.2,executing==2.0.1,filelock==3.13.1,google-api-core==2.15.0,google-auth==2.25.2,google-cloud-core==2.4.1,google-cloud-storage==2.14.0,google-crc32c==1.5.0,google-resumable-media==2.7.0,googleapis-common-protos==1.62.0,greenlet==3.0.2,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,iniconfig==2.0.0,ipython==8.18.1,isort==5.13.2,itemadapter==0.8.0,itemloaders==1.1.0,jedi==0.19.1,jmespath==1.0.1,lxml==4.9.3,MarkupSafe==2.0.1,matplotlib-inline==0.1.6,mccabe==0.7.0,packaging==23.2,parsel==1.8.1,parso==0.8.3,pexpect==4.9.0,Pillow==10.1.0,pip==23.3.2,platformdirs==4.1.0,pluggy==1.3.0,priority==1.3.0,prompt-toolkit==3.0.43,Protego==0.3.0,protobuf==4.25.1,ptyprocess==0.7.0,pure-eval==0.2.2,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,PyDispatcher==2.0.7,pyftpdlib==1.5.9,Pygments==2.17.2,pylint==3.0.1,pyOpenSSL==23.3.0,pytest==7.4.3,pytest-cov==4.0.0,pytest-xdist==3.5.0,python-dateutil==2.8.2,pyxdg==0.28,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,robotexclusionrulesparser==1.7.1,rsa==4.9,s3transfer==0.9.0,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=6f43aa5d37ec10736bd947dbba89df54ebc97175083ab360a721b3360f0d895b,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,stack-data==0.6.3,sybil==6.0.2,testfixtures==7.2.2,tldextract==5.1.1,tomlkit==0.12.3,traitlets==5.14.0,Twisted==23.10.0,typing_extensions==4.9.0,urllib3==2.0.7,uvloop==0.19.0,w3lib==2.1.2,wcwidth==0.2.12,zope.interface==6.1,zstandard==0.22.0\npylint: commands[0]> pylint conftest.py docs extras scrapy setup.py tests\n<unknown>:230: SyntaxWarning: invalid escape sequence '\\d'\n************* Module scrapy.crawler\nscrapy/crawler.py:182:4: W0211: Static method with 'cls' as first argument (bad-staticmethod-argument)\nscrapy/crawler.py:184:15: C0123: Use isinstance() rather than type() for a typecheck. (unidiomatic-typecheck)\n\n------------------------------------\nYour code has been rated at 10.00/10\n\npylint: exit 20 (56.60 seconds) /home/runner/work/scrapy/scrapy> pylint conftest.py docs extras scrapy setup.py tests pid=2086\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n pylint: FAIL code 20 (100.72=setup[44.12]+cmd[56.60] seconds)\n evaluation failed :( (100.86 seconds)\n##[error]Process completed with exit code 20.\n"}], "diff": "diff --git a/scrapy/crawler.py b/scrapy/crawler.py\nindex 76f4048e4..333b39f33 100644\n--- a/scrapy/crawler.py\n+++ b/scrapy/crawler.py\n@@ -179,9 +179,12 @@ class Crawler:\n yield maybeDeferred(self.engine.stop)\n \n @staticmethod\n- def _get_component(cls, components):\n+ def _get_component(component_class, components):\n for component in components:\n- if type(component) is cls:\n+ if (\n+ type(component) # pylint: disable=unidiomatic-typecheck\n+ is component_class\n+ ):\n return component\n return None\n \n", "difficulty": 0, "changed_files": ["scrapy/crawler.py"], "commit_link": "https://github.com/scrapy/scrapy/tree/0f71221cf9875ed8ef3400e1008408e79b6691e6"}
 
 
data/python/102f918.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 21, "repo_owner": "mindsdb", "repo_name": "mindsdb", "head_branch": "staging", "workflow_name": "MindsDB Code Checks", "workflow_filename": "code_checks.yml", "workflow_path": ".github/workflows/code_checks.yml", "contributor": "mindsdb", "sha_fail": "102f918deb2532bb7b825f00258f2c1414cf94da", "sha_success": "1cf75d7a1071fba82d4190725e984e8d55db48fb", "workflow": "name: MindsDB Code Checks\n\non:\n pull_request:\n branches: [stable, staging]\n \njobs:\n check_requirements:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/[email protected]\n with:\n python-version: 3.9\n - name: Check main requirements\n shell: bash\n run: |\n pip install -r requirements/requirements-dev.txt\n\n python tests/scripts/check_requirements.py\n\n - name: Check requirements files are installable\n shell: bash\n run: |\n pip install --dry-run --ignore-installed . # Install only the default handlers. We can expand this to all handlers later with: .[all_handlers_extras]\n", "logs": [{"step_name": "check_requirements/4_Check main requirements.txt", "log": "##[group]Run pip install -r requirements/requirements-dev.txt\n\u001b[36;1mpip install -r requirements/requirements-dev.txt\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpython tests/scripts/check_requirements.py\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nCollecting black==23.3.0 (from -r requirements/requirements-dev.txt (line 1))\n Downloading black-23.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.7 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.7/1.7 MB 18.4 MB/s eta 0:00:00\nCollecting flake8==5.0.4 (from -r requirements/requirements-dev.txt (line 2))\n Downloading flake8-5.0.4-py2.py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.9/61.9 kB 19.0 MB/s eta 0:00:00\nCollecting isort==5.10.1 (from -r requirements/requirements-dev.txt (line 3))\n Downloading isort-5.10.1-py3-none-any.whl (103 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 103.4/103.4 kB 26.5 MB/s eta 0:00:00\nCollecting pre-commit>=2.16.0 (from -r requirements/requirements-dev.txt (line 4))\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting watchfiles==0.19.0 (from -r requirements/requirements-dev.txt (line 5))\n Downloading watchfiles-0.19.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.3/1.3 MB 31.1 MB/s eta 0:00:00\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from -r requirements/requirements-dev.txt (line 6)) (58.1.0)\nCollecting wheel (from -r requirements/requirements-dev.txt (line 7))\n Downloading wheel-0.42.0-py3-none-any.whl.metadata (2.2 kB)\nCollecting deptry==0.12.0 (from -r requirements/requirements-dev.txt (line 8))\n Downloading deptry-0.12.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting twine (from -r requirements/requirements-dev.txt (line 9))\n Downloading twine-4.0.2-py3-none-any.whl (36 kB)\nCollecting click>=8.0.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading click-8.1.7-py3-none-any.whl.metadata (3.0 kB)\nCollecting mypy-extensions>=0.4.3 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading mypy_extensions-1.0.0-py3-none-any.whl (4.7 kB)\nCollecting packaging>=22.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting pathspec>=0.9.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading pathspec-0.12.1-py3-none-any.whl.metadata (21 kB)\nCollecting platformdirs>=2 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting tomli>=1.1.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting typing-extensions>=3.10.0.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading typing_extensions-4.9.0-py3-none-any.whl.metadata (3.0 kB)\nCollecting mccabe<0.8.0,>=0.7.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading mccabe-0.7.0-py2.py3-none-any.whl (7.3 kB)\nCollecting pycodestyle<2.10.0,>=2.9.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading pycodestyle-2.9.1-py2.py3-none-any.whl (41 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.5/41.5 kB 11.9 MB/s eta 0:00:00\nCollecting pyflakes<2.6.0,>=2.5.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading pyflakes-2.5.0-py2.py3-none-any.whl (66 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 66.1/66.1 kB 20.4 MB/s eta 0:00:00\nCollecting anyio>=3.0.0 (from watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading anyio-4.2.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting chardet>=4.0.0 (from deptry==0.12.0->-r requirements/requirements-dev.txt (line 8))\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting cfgv>=2.0.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting pkginfo>=1.8.1 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pkginfo-1.9.6-py3-none-any.whl (30 kB)\nCollecting readme-renderer>=35.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading readme_renderer-42.0-py3-none-any.whl.metadata (2.8 kB)\nCollecting requests>=2.20 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting requests-toolbelt!=0.9.0,>=0.8.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading requests_toolbelt-1.0.0-py2.py3-none-any.whl (54 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 17.0 MB/s eta 0:00:00\nCollecting urllib3>=1.26.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading urllib3-2.1.0-py3-none-any.whl.metadata (6.4 kB)\nCollecting importlib-metadata>=3.6 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading importlib_metadata-7.0.1-py3-none-any.whl.metadata (4.9 kB)\nCollecting keyring>=15.1 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading keyring-24.3.0-py3-none-any.whl.metadata (20 kB)\nCollecting rfc3986>=1.4.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading rfc3986-2.0.0-py2.py3-none-any.whl (31 kB)\nCollecting rich>=12.0.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading rich-13.7.0-py3-none-any.whl.metadata (18 kB)\nCollecting idna>=2.8 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading idna-3.6-py3-none-any.whl.metadata (9.9 kB)\nCollecting sniffio>=1.1 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading sniffio-1.3.0-py3-none-any.whl (10 kB)\nCollecting exceptiongroup>=1.0.2 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading exceptiongroup-1.2.0-py3-none-any.whl.metadata (6.6 kB)\nCollecting zipp>=0.5 (from importlib-metadata>=3.6->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading zipp-3.17.0-py3-none-any.whl.metadata (3.7 kB)\nCollecting jaraco.classes (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading jaraco.classes-3.3.0-py3-none-any.whl.metadata (2.9 kB)\nCollecting SecretStorage>=3.2 (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading SecretStorage-3.3.3-py3-none-any.whl (15 kB)\nCollecting jeepney>=0.4.2 (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading jeepney-0.8.0-py3-none-any.whl (48 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 48.4/48.4 kB 17.1 MB/s eta 0:00:00\nCollecting nh3>=0.2.14 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading nh3-0.2.15-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.7 kB)\nCollecting docutils>=0.13.1 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading docutils-0.20.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting Pygments>=2.5.1 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pygments-2.17.2-py3-none-any.whl.metadata (2.6 kB)\nCollecting charset-normalizer<4,>=2 (from requests>=2.20->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (33 kB)\nCollecting certifi>=2017.4.17 (from requests>=2.20->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading certifi-2023.11.17-py3-none-any.whl.metadata (2.2 kB)\nCollecting markdown-it-py>=2.2.0 (from rich>=12.0.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading markdown_it_py-3.0.0-py3-none-any.whl.metadata (6.9 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=12.0.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB)\nCollecting cryptography>=2.0 (from SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading cryptography-41.0.7-cp37-abi3-manylinux_2_28_x86_64.whl.metadata (5.2 kB)\nCollecting more-itertools (from jaraco.classes->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading more_itertools-10.1.0-py3-none-any.whl.metadata (33 kB)\nCollecting cffi>=1.12 (from cryptography>=2.0->SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading cffi-1.16.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.5 kB)\nCollecting pycparser (from cffi>=1.12->cryptography>=2.0->SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pycparser-2.21-py2.py3-none-any.whl (118 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 118.7/118.7 kB 35.3 MB/s eta 0:00:00\nDownloading deptry-0.12.0-py3-none-any.whl (41 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.7/41.7 kB 13.1 MB/s eta 0:00:00\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 42.9 MB/s eta 0:00:00\nDownloading wheel-0.42.0-py3-none-any.whl (65 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 65.4/65.4 kB 22.9 MB/s eta 0:00:00\nDownloading anyio-4.2.0-py3-none-any.whl (85 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.5/85.5 kB 29.1 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 43.8 MB/s eta 0:00:00\nDownloading click-8.1.7-py3-none-any.whl (97 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 97.9/97.9 kB 33.7 MB/s eta 0:00:00\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.2 MB/s eta 0:00:00\nDownloading importlib_metadata-7.0.1-py3-none-any.whl (23 kB)\nDownloading keyring-24.3.0-py3-none-any.whl (38 kB)\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 18.2 MB/s eta 0:00:00\nDownloading pathspec-0.12.1-py3-none-any.whl (31 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (738 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 738.9/738.9 kB 37.6 MB/s eta 0:00:00\nDownloading readme_renderer-42.0-py3-none-any.whl (13 kB)\nDownloading requests-2.31.0-py3-none-any.whl (62 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 62.6/62.6 kB 21.9 MB/s eta 0:00:00\nDownloading rich-13.7.0-py3-none-any.whl (240 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 240.6/240.6 kB 51.0 MB/s eta 0:00:00\nDownloading typing_extensions-4.9.0-py3-none-any.whl (32 kB)\nDownloading urllib3-2.1.0-py3-none-any.whl (104 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.6/104.6 kB 33.0 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 53.1 MB/s eta 0:00:00\nDownloading certifi-2023.11.17-py3-none-any.whl (162 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 162.5/162.5 kB 43.3 MB/s eta 0:00:00\nDownloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (142 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 142.3/142.3 kB 40.8 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 68.6 MB/s eta 0:00:00\nDownloading docutils-0.20.1-py3-none-any.whl (572 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 572.7/572.7 kB 67.0 MB/s eta 0:00:00\nDownloading exceptiongroup-1.2.0-py3-none-any.whl (16 kB)\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading idna-3.6-py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.6/61.6 kB 20.6 MB/s eta 0:00:00\nDownloading markdown_it_py-3.0.0-py3-none-any.whl (87 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 87.5/87.5 kB 28.4 MB/s eta 0:00:00\nDownloading nh3-0.2.15-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.7 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.7/1.7 MB 72.8 MB/s eta 0:00:00\nDownloading pygments-2.17.2-py3-none-any.whl (1.2 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.2/1.2 MB 81.6 MB/s eta 0:00:00\nDownloading zipp-3.17.0-py3-none-any.whl (7.4 kB)\nDownloading jaraco.classes-3.3.0-py3-none-any.whl (5.9 kB)\nDownloading cryptography-41.0.7-cp37-abi3-manylinux_2_28_x86_64.whl (4.4 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.4/4.4 MB 94.7 MB/s eta 0:00:00\nDownloading more_itertools-10.1.0-py3-none-any.whl (55 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 55.8/55.8 kB 17.6 MB/s eta 0:00:00\nDownloading cffi-1.16.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (443 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 443.4/443.4 kB 72.9 MB/s eta 0:00:00\nInstalling collected packages: nh3, distlib, zipp, wheel, urllib3, typing-extensions, tomli, sniffio, rfc3986, pyyaml, Pygments, pyflakes, pycparser, pycodestyle, platformdirs, pkginfo, pathspec, packaging, nodeenv, mypy-extensions, more-itertools, mdurl, mccabe, jeepney, isort, idna, identify, filelock, exceptiongroup, docutils, click, charset-normalizer, chardet, cfgv, certifi, virtualenv, requests, readme-renderer, markdown-it-py, jaraco.classes, importlib-metadata, flake8, deptry, cffi, black, anyio, watchfiles, rich, requests-toolbelt, pre-commit, cryptography, SecretStorage, keyring, twine\nSuccessfully installed Pygments-2.17.2 SecretStorage-3.3.3 anyio-4.2.0 black-23.3.0 certifi-2023.11.17 cffi-1.16.0 cfgv-3.4.0 chardet-5.2.0 charset-normalizer-3.3.2 click-8.1.7 cryptography-41.0.7 deptry-0.12.0 distlib-0.3.8 docutils-0.20.1 exceptiongroup-1.2.0 filelock-3.13.1 flake8-5.0.4 identify-2.5.33 idna-3.6 importlib-metadata-7.0.1 isort-5.10.1 jaraco.classes-3.3.0 jeepney-0.8.0 keyring-24.3.0 markdown-it-py-3.0.0 mccabe-0.7.0 mdurl-0.1.2 more-itertools-10.1.0 mypy-extensions-1.0.0 nh3-0.2.15 nodeenv-1.8.0 packaging-23.2 pathspec-0.12.1 pkginfo-1.9.6 platformdirs-4.1.0 pre-commit-3.6.0 pycodestyle-2.9.1 pycparser-2.21 pyflakes-2.5.0 pyyaml-6.0.1 readme-renderer-42.0 requests-2.31.0 requests-toolbelt-1.0.0 rfc3986-2.0.0 rich-13.7.0 sniffio-1.3.0 tomli-2.0.1 twine-4.0.2 typing-extensions-4.9.0 urllib3-2.1.0 virtualenv-20.25.0 watchfiles-0.19.0 wheel-0.42.0 zipp-3.17.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\n--- Checking requirements files for duplicates ---\n\n--- Checking that requirements match imports ---\n- requirements/requirements.txt\n None:None: DEP002 'type_infer' defined as a dependency but not used in the codebase\n\n\n--- Checking handlers that require other handlers ---\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/mindsdb/__main__.py b/mindsdb/__main__.py\nindex cac3482883..6cc92a7899 100644\n--- a/mindsdb/__main__.py\n+++ b/mindsdb/__main__.py\n@@ -35,6 +35,7 @@ from mindsdb.utilities.fs import create_dirs_recursive, clean_process_marks, cle\n from mindsdb.utilities.telemetry import telemetry_file_exists, disable_telemetry\n from mindsdb.utilities.context import context as ctx\n from mindsdb.utilities.auth import register_oauth_client, get_aws_meta_data\n+import type_infer # noqa\n \n try:\n import torch.multiprocessing as mp\n", "difficulty": 1, "changed_files": ["mindsdb/__main__.py"], "commit_link": "https://github.com/mindsdb/mindsdb/tree/102f918deb2532bb7b825f00258f2c1414cf94da"}
 
 
data/python/1168b92.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 131, "repo_owner": "scrapy", "repo_name": "scrapy", "head_branch": "simplify-attempt2", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "monicaq21", "sha_fail": "1168b9244d680437e2a2683294a55c0b52c118f6", "sha_success": "b5b7c99e24c722fa09c426ed0245df1ae792d486", "workflow": "name: Checks\non: [push, pull_request]\n\nconcurrency:\n group: ${{github.workflow}}-${{ github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n include:\n - python-version: \"3.12\"\n env:\n TOXENV: pylint\n - python-version: 3.8\n env:\n TOXENV: typing\n - python-version: \"3.11\" # Keep in sync with .readthedocs.yml\n env:\n TOXENV: docs\n - python-version: \"3.12\"\n env:\n TOXENV: twinecheck\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Run check\n env: ${{ matrix.env }}\n run: |\n pip install -U tox\n tox\n\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "checks (3.12, pylint)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n TOXENV: pylint\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 19.9 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 51.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 16.5 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 117.4 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 82.7 MB/s eta 0:00:00\nInstalling collected packages: distlib, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.7 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\npylint: install_deps> python -I -m pip install -ctests/upper-constraints.txt boto3 google-cloud-storage 'markupsafe<2.1.0' 'mitmproxy<8,>=4.0.4; python_version < \"3.9\" and implementation_name != \"pypy\"' Pillow pylint==3.0.1 robotexclusionrulesparser 'Twisted[http2]' -r tests/requirements.txt\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\npylint: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\npylint: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\npylint: freeze> python -m pip freeze --all\npylint: annotated-types==0.6.0,astroid==3.0.1,asttokens==2.4.1,attrs==23.1.0,Automat==22.10.0,blessed==1.20.0,boto3==1.33.11,botocore==1.33.11,bpython==0.24,Brotli==1.1.0,cachetools==5.3.2,Cerberus==1.3.5,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,coverage==7.3.2,cryptography==41.0.7,cssselect==1.2.0,curtsies==0.4.2,cwcwidth==0.1.9,decorator==5.1.1,dill==0.3.7,distlib==0.3.7,docopt==0.6.2,execnet==2.0.2,executing==2.0.1,filelock==3.13.1,google-api-core==2.15.0,google-auth==2.25.2,google-cloud-core==2.4.1,google-cloud-storage==2.13.0,google-crc32c==1.5.0,google-resumable-media==2.6.0,googleapis-common-protos==1.62.0,greenlet==3.0.2,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,iniconfig==2.0.0,ipython==8.18.1,isort==5.13.0,itemadapter==0.8.0,itemloaders==1.1.0,jedi==0.19.1,jmespath==1.0.1,lxml==4.9.3,MarkupSafe==2.0.1,matplotlib-inline==0.1.6,mccabe==0.7.0,packaging==23.2,parsel==1.8.1,parso==0.8.3,pep517==0.13.1,pexpect==4.9.0,Pillow==10.1.0,pip==23.3.1,pip-api==0.0.30,pipreqs==0.4.13,platformdirs==4.1.0,plette==0.4.4,pluggy==1.3.0,priority==1.3.0,prompt-toolkit==3.0.41,Protego==0.3.0,protobuf==4.25.1,ptyprocess==0.7.0,pure-eval==0.2.2,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,pydantic==2.5.2,pydantic_core==2.14.5,PyDispatcher==2.0.7,pyftpdlib==1.5.9,Pygments==2.17.2,pylint==3.0.1,pyOpenSSL==23.3.0,pytest==7.4.3,pytest-cov==4.0.0,pytest-xdist==3.5.0,python-dateutil==2.8.2,pyxdg==0.28,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,requirementslib==3.0.0,robotexclusionrulesparser==1.7.1,rsa==4.9,s3transfer==0.8.2,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=badae26f961983adc76ef32b8b07c6ba5b979a31f891e5f2e57590381cc5b815,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,stack-data==0.6.3,sybil==6.0.2,testfixtures==7.2.2,tldextract==5.1.1,tomlkit==0.12.3,traitlets==5.14.0,Twisted==23.10.0,typing_extensions==4.9.0,urllib3==2.0.7,uvloop==0.19.0,w3lib==2.1.2,wcwidth==0.2.12,yarg==0.1.9,zope.interface==6.1,zstandard==0.22.0\npylint: commands[0]> pylint conftest.py docs extras scrapy setup.py tests\n<unknown>:230: SyntaxWarning: invalid escape sequence '\\d'\n************* Module scrapy.crawler\nscrapy/crawler.py:112:37: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.engine\nscrapy/core/engine.py:361:20: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.scheduler\nscrapy/core/scheduler.py:325:15: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/scheduler.py:336:12: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.contextfactory\nscrapy/core/downloader/contextfactory.py:168:26: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:168:26: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:174:26: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:174:26: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers\nscrapy/core/downloader/handlers/__init__.py:58:17: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/__init__.py:58:17: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers.http10\nscrapy/core/downloader/handlers/http10.py:33:37: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/http10.py:33:37: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers.s3\nscrapy/core/downloader/handlers/s3.py:53:24: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/s3.py:53:24: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module tests.test_downloader_handlers\ntests/test_downloader_handlers.py:831:22: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:831:22: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:861:17: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:861:17: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:888:12: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:888:12: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module tests.test_webclient\ntests/test_webclient.py:473:33: W1117: 'settings' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_webclient.py:485:33: W1117: 'settings' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n\n-----------------------------------\nYour code has been rated at 9.99/10\n\npylint: exit 4 (53.83 seconds) /home/runner/work/scrapy/scrapy> pylint conftest.py docs extras scrapy setup.py tests pid=2037\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n pylint: FAIL code 4 (100.80=setup[46.97]+cmd[53.83] seconds)\n evaluation failed :( (100.88 seconds)\n##[error]Process completed with exit code 4.\n"}, {"step_name": "checks (3.8, typing)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n TOXENV: typing\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting tomli>=2.0.1 (from tox)\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 8.4 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 32.2 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 17.1 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 82.9 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 79.6 MB/s eta 0:00:00\nInstalling collected packages: distlib, tomli, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.7 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tomli-2.0.1 tox-4.11.4 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.1\n[notice] To update, run: pip install --upgrade pip\ntyping: install_deps> python -I -m pip install -ctests/upper-constraints.txt mypy==1.6.1 types-attrs==19.1.0 types-lxml==2023.10.21 types-Pillow==10.1.0.0 types-Pygments==2.16.0.0 types-pyOpenSSL==23.3.0.0 types-setuptools==68.2.0.0 typing-extensions==4.8.0 'w3lib>=2.1.2'\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\ntyping: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\ntyping: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\ntyping: freeze> python -m pip freeze --all\ntyping: attrs==23.1.0,Automat==22.10.0,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,cryptography==41.0.7,cssselect==1.2.0,filelock==3.13.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,itemadapter==0.8.0,itemloaders==1.1.0,jmespath==1.0.1,lxml==4.9.3,mypy==1.6.1,mypy-extensions==1.0.0,packaging==23.2,parsel==1.8.1,pip==23.3.1,Protego==0.3.0,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,PyDispatcher==2.0.7,pyOpenSSL==23.3.0,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=ebe501ee4a51f734f99dff82a0f62429c4051d3976e2377b4c07e9f5fb55d533,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,tldextract==5.1.1,tomli==2.0.1,Twisted==23.10.0,types-attrs==19.1.0,types-beautifulsoup4==4.12.0.7,types-docutils==0.20.0.3,types-html5lib==1.1.11.15,types-lxml==2023.10.21,types-Pillow==10.1.0.0,types-Pygments==2.16.0.0,types-pyOpenSSL==23.3.0.0,types-setuptools==68.2.0.0,typing_extensions==4.8.0,urllib3==2.1.0,w3lib==2.1.2,wheel==0.42.0,zope.interface==6.1\ntyping: commands[0]> mypy scrapy tests\nscrapy/core/downloader/handlers/__init__.py:58: error: Too few arguments for \"build_from_crawler\" [call-arg]\nscrapy/core/engine.py:361: error: Too few arguments for \"build_from_crawler\" [call-arg]\nscrapy/crawler.py:112: error: Too few arguments for \"build_from_crawler\" [call-arg]\nFound 3 errors in 3 files (checked 333 source files)\ntyping: exit 1 (8.99 seconds) /home/runner/work/scrapy/scrapy> mypy scrapy tests pid=1984\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n typing: FAIL code 1 (25.59=setup[16.60]+cmd[8.99] seconds)\n evaluation failed :( (25.71 seconds)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/scrapy/core/downloader/contextfactory.py b/scrapy/core/downloader/contextfactory.py\nindex 9047a5b83..05d6e7a09 100644\n--- a/scrapy/core/downloader/contextfactory.py\n+++ b/scrapy/core/downloader/contextfactory.py\n@@ -166,14 +166,14 @@ def load_context_factory_from_settings(settings, crawler):\n # try method-aware context factory\n try:\n context_factory = build_from_crawler(\n- objcls=context_factory_cls,\n- crawler=crawler,\n+ context_factory_cls,\n+ crawler,\n method=ssl_method,\n )\n except TypeError:\n context_factory = build_from_crawler(\n- objcls=context_factory_cls,\n- crawler=crawler,\n+ context_factory_cls,\n+ crawler,\n )\n msg = (\n f\"{settings['DOWNLOADER_CLIENTCONTEXTFACTORY']} does not accept \"\ndiff --git a/scrapy/core/downloader/handlers/__init__.py b/scrapy/core/downloader/handlers/__init__.py\nindex 466393b80..416669b7f 100644\n--- a/scrapy/core/downloader/handlers/__init__.py\n+++ b/scrapy/core/downloader/handlers/__init__.py\n@@ -56,8 +56,8 @@ class DownloadHandlers:\n if skip_lazy and getattr(dhcls, \"lazy\", True):\n return None\n dh = build_from_crawler(\n- objcls=dhcls,\n- crawler=self._crawler,\n+ dhcls,\n+ self._crawler,\n )\n except NotConfigured as ex:\n self._notconfigured[scheme] = str(ex)\ndiff --git a/scrapy/core/downloader/handlers/http10.py b/scrapy/core/downloader/handlers/http10.py\nindex 817a420c0..b6ac7a251 100644\n--- a/scrapy/core/downloader/handlers/http10.py\n+++ b/scrapy/core/downloader/handlers/http10.py\n@@ -31,8 +31,8 @@ class HTTP10DownloadHandler:\n host, port = to_unicode(factory.host), factory.port\n if factory.scheme == b\"https\":\n client_context_factory = build_from_crawler(\n- objcls=self.ClientContextFactory,\n- crawler=self._crawler,\n+ self.ClientContextFactory,\n+ self._crawler,\n )\n return reactor.connectSSL(host, port, factory, client_context_factory)\n return reactor.connectTCP(host, port, factory)\ndiff --git a/scrapy/core/downloader/handlers/s3.py b/scrapy/core/downloader/handlers/s3.py\nindex a37f29aac..1f7533759 100644\n--- a/scrapy/core/downloader/handlers/s3.py\n+++ b/scrapy/core/downloader/handlers/s3.py\n@@ -51,8 +51,8 @@ class S3DownloadHandler:\n )\n \n _http_handler = build_from_crawler(\n- objcls=httpdownloadhandler,\n- crawler=crawler,\n+ httpdownloadhandler,\n+ crawler,\n )\n self._download_http = _http_handler.download_request\n \ndiff --git a/scrapy/core/engine.py b/scrapy/core/engine.py\nindex 281dc8a54..545cd401f 100644\n--- a/scrapy/core/engine.py\n+++ b/scrapy/core/engine.py\n@@ -358,7 +358,7 @@ class ExecutionEngine:\n raise RuntimeError(f\"No free spider slot when opening {spider.name!r}\")\n logger.info(\"Spider opened\", extra={\"spider\": spider})\n nextcall = CallLaterOnce(self._next_request)\n- scheduler = build_from_crawler(self.scheduler_cls, crawler=self.crawler)\n+ scheduler = build_from_crawler(self.scheduler_cls, self.crawler)\n start_requests = yield self.scraper.spidermw.process_start_requests(\n start_requests, spider\n )\ndiff --git a/scrapy/core/scheduler.py b/scrapy/core/scheduler.py\nindex bf027e1f9..f41b83a67 100644\n--- a/scrapy/core/scheduler.py\n+++ b/scrapy/core/scheduler.py\n@@ -324,7 +324,7 @@ class Scheduler(BaseScheduler):\n \"\"\"Create a new priority queue instance, with in-memory storage\"\"\"\n return build_from_crawler(\n self.pqclass,\n- crawler=self.crawler,\n+ self.crawler,\n downstream_queue_cls=self.mqclass,\n key=\"\",\n )\n@@ -335,7 +335,7 @@ class Scheduler(BaseScheduler):\n state = self._read_dqs_state(self.dqdir)\n q = build_from_crawler(\n self.pqclass,\n- crawler=self.crawler,\n+ self.crawler,\n downstream_queue_cls=self.dqclass,\n key=self.dqdir,\n startprios=state,\ndiff --git a/scrapy/crawler.py b/scrapy/crawler.py\nindex b7a6df2f3..844d5f759 100644\n--- a/scrapy/crawler.py\n+++ b/scrapy/crawler.py\n@@ -111,7 +111,7 @@ class Crawler:\n \n self.request_fingerprinter = build_from_crawler(\n load_object(self.settings[\"REQUEST_FINGERPRINTER_CLASS\"]),\n- crawler=self,\n+ self,\n )\n \n reactor_class: str = self.settings[\"TWISTED_REACTOR\"]\ndiff --git a/tests/test_downloader_handlers.py b/tests/test_downloader_handlers.py\nindex 37a53643b..dd07d33f1 100644\n--- a/tests/test_downloader_handlers.py\n+++ b/tests/test_downloader_handlers.py\n@@ -829,8 +829,8 @@ class S3AnonTestCase(unittest.TestCase):\n skip_if_no_boto()\n crawler = get_crawler()\n self.s3reqh = build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n httpdownloadhandler=HttpDownloadHandlerMock,\n # anon=True, # implicit\n )\n@@ -859,8 +859,8 @@ class S3TestCase(unittest.TestCase):\n skip_if_no_boto()\n crawler = get_crawler()\n s3reqh = build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n aws_access_key_id=self.AWS_ACCESS_KEY_ID,\n aws_secret_access_key=self.AWS_SECRET_ACCESS_KEY,\n httpdownloadhandler=HttpDownloadHandlerMock,\n@@ -886,8 +886,8 @@ class S3TestCase(unittest.TestCase):\n try:\n crawler = get_crawler()\n build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n extra_kw=True,\n )\n except Exception as e:\ndiff --git a/tests/test_webclient.py b/tests/test_webclient.py\nindex a69d9c1b0..d4b6ba15b 100644\n--- a/tests/test_webclient.py\n+++ b/tests/test_webclient.py\n@@ -471,7 +471,7 @@ class WebClientCustomCiphersSSLTestCase(WebClientSSLTestCase):\n s = \"0123456789\" * 10\n settings = Settings({\"DOWNLOADER_CLIENT_TLS_CIPHERS\": self.custom_ciphers})\n client_context_factory = build_from_settings(\n- ScrapyClientContextFactory, settings=settings\n+ ScrapyClientContextFactory, settings\n )\n return getPage(\n self.getURL(\"payload\"), body=s, contextFactory=client_context_factory\n@@ -483,7 +483,7 @@ class WebClientCustomCiphersSSLTestCase(WebClientSSLTestCase):\n {\"DOWNLOADER_CLIENT_TLS_CIPHERS\": \"ECDHE-RSA-AES256-GCM-SHA384\"}\n )\n client_context_factory = build_from_settings(\n- ScrapyClientContextFactory, settings=settings\n+ ScrapyClientContextFactory, settings\n )\n d = getPage(\n self.getURL(\"payload\"), body=s, contextFactory=client_context_factory\n", "difficulty": 2, "changed_files": ["scrapy/core/downloader/contextfactory.py", "scrapy/core/downloader/handlers/__init__.py", "scrapy/core/downloader/handlers/http10.py", "scrapy/core/downloader/handlers/s3.py", "scrapy/core/engine.py", "scrapy/core/scheduler.py", "scrapy/crawler.py", "tests/test_downloader_handlers.py", "tests/test_webclient.py"], "commit_link": "https://github.com/scrapy/scrapy/tree/1168b9244d680437e2a2683294a55c0b52c118f6"}
 
 
data/python/124065b.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 121, "repo_owner": "autogluon", "repo_name": "autogluon", "head_branch": "modify-sem-seg", "workflow_name": "Codespell", "workflow_filename": "codespell.yml", "workflow_path": ".github/workflows/codespell.yml", "contributor": "Harry-zzh", "sha_fail": "124065b09d5b45bfbd2a38262aa8e9fc11ded7a6", "sha_success": "0ed9a432f338d4a05f3b729c99c17f2546c96573", "workflow": "---\nname: Codespell\n\non:\n push:\n branches: [master]\n pull_request:\n branches: [master]\n\npermissions:\n contents: read\n\njobs:\n codespell:\n name: Check for spelling errors\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout\n uses: actions/checkout@v3\n - name: Codespell\n uses: codespell-project/actions-codespell@v2\n", "logs": [{"step_name": "Check for spelling errors/4_Codespell.txt", "log": "##[group]Run codespell-project/actions-codespell@v2\nwith:\n skip: ./.git\n##[endgroup]\n##[command]/usr/bin/docker run --name da69a10d75df8948c097276461d46605f8_7bca0f --label 8689da --workdir /github/workspace --rm -e \"INPUT_CHECK_FILENAMES\" -e \"INPUT_CHECK_HIDDEN\" -e \"INPUT_EXCLUDE_FILE\" -e \"INPUT_SKIP\" -e \"INPUT_BUILTIN\" -e \"INPUT_IGNORE_WORDS_FILE\" -e \"INPUT_IGNORE_WORDS_LIST\" -e \"INPUT_URI_IGNORE_WORDS_LIST\" -e \"INPUT_PATH\" -e \"INPUT_ONLY_WARN\" -e \"HOME\" -e \"GITHUB_JOB\" -e \"GITHUB_REF\" -e \"GITHUB_SHA\" -e \"GITHUB_REPOSITORY\" -e \"GITHUB_REPOSITORY_OWNER\" -e \"GITHUB_REPOSITORY_OWNER_ID\" -e \"GITHUB_RUN_ID\" -e \"GITHUB_RUN_NUMBER\" -e \"GITHUB_RETENTION_DAYS\" -e \"GITHUB_RUN_ATTEMPT\" -e \"GITHUB_REPOSITORY_ID\" -e \"GITHUB_ACTOR_ID\" -e \"GITHUB_ACTOR\" -e \"GITHUB_TRIGGERING_ACTOR\" -e \"GITHUB_WORKFLOW\" -e \"GITHUB_HEAD_REF\" -e \"GITHUB_BASE_REF\" -e \"GITHUB_EVENT_NAME\" -e \"GITHUB_SERVER_URL\" -e \"GITHUB_API_URL\" -e \"GITHUB_GRAPHQL_URL\" -e \"GITHUB_REF_NAME\" -e \"GITHUB_REF_PROTECTED\" -e \"GITHUB_REF_TYPE\" -e \"GITHUB_WORKFLOW_REF\" -e \"GITHUB_WORKFLOW_SHA\" -e \"GITHUB_WORKSPACE\" -e \"GITHUB_ACTION\" -e \"GITHUB_EVENT_PATH\" -e \"GITHUB_ACTION_REPOSITORY\" -e \"GITHUB_ACTION_REF\" -e \"GITHUB_PATH\" -e \"GITHUB_ENV\" -e \"GITHUB_STEP_SUMMARY\" -e \"GITHUB_STATE\" -e \"GITHUB_OUTPUT\" -e \"RUNNER_OS\" -e \"RUNNER_ARCH\" -e \"RUNNER_NAME\" -e \"RUNNER_ENVIRONMENT\" -e \"RUNNER_TOOL_CACHE\" -e \"RUNNER_TEMP\" -e \"RUNNER_WORKSPACE\" -e \"ACTIONS_RUNTIME_URL\" -e \"ACTIONS_RUNTIME_TOKEN\" -e \"ACTIONS_CACHE_URL\" -e \"ACTIONS_RESULTS_URL\" -e GITHUB_ACTIONS=true -e CI=true -v \"/var/run/docker.sock\":\"/var/run/docker.sock\" -v \"/home/runner/work/_temp/_github_home\":\"/github/home\" -v \"/home/runner/work/_temp/_github_workflow\":\"/github/workflow\" -v \"/home/runner/work/_temp/_runner_file_commands\":\"/github/file_commands\" -v \"/home/runner/work/autogluon/autogluon\":\"/github/workspace\" 8689da:69a10d75df8948c097276461d46605f8\nRunning codespell on '' with the following options...\nCheck filenames? ''\nCheck hidden? ''\nExclude file ''\nSkipping './.git'\nBuiltin dictionaries ''\nIgnore words file ''\nIgnore words list ''\nIgnore URI words list ''\nResulting CLI options --skip ./.git\n1\n##[error]./multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py:752: evalution ==> evaluation, evolution\nCodespell found one or more problems\n"}], "diff": "diff --git a/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py b/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\nindex 27032888..e6c62802 100644\n--- a/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\n+++ b/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\n@@ -749,7 +749,7 @@ class MAE(COD):\n COD_METRICS_NAMES = {\"sm\": SM(), \"fm\": FM(), \"em\": EM(), \"mae\": MAE()}\n \n \n-# TODO: Modify multi-gpu evalution error. Maybe there will be a more elegant way.\n+# TODO: Modify multi-gpu evaluation error. Maybe there will be a more elegant way.\n class Multiclass_IoU_Pred:\n \"\"\"\n Compute the IoU for multi-class semantic segmentation based on https://github.com/xieenze/Trans2Seg/blob/master/segmentron/utils/score.py.\n", "difficulty": 0, "changed_files": ["multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py"], "commit_link": "https://github.com/autogluon/autogluon/tree/124065b09d5b45bfbd2a38262aa8e9fc11ded7a6"}
 
 
data/python/16a0c04.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 140, "repo_owner": "dask", "repo_name": "dask", "head_branch": "milesgranger/10801-more-deprecations", "workflow_name": "Linting", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "milesgranger", "sha_fail": "16a0c04d06205527ec5e379df2596b399ee5dadc", "sha_success": "3716f88d477a969d0910c75b4268949af11c0e6e", "workflow": "name: Linting\n\non:\n push:\n branches: main\n pull_request:\n branches: main\n\njobs:\n checks:\n name: pre-commit hooks\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - uses: actions/setup-python@v5\n with:\n python-version: '3.9'\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "pre-commit hooks/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from nodeenv>=0.11.1->pre-commit) (58.1.0)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 19.7 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 27.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (738 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 738.9/738.9 kB 80.2 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 114.1 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 78.9 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nInstalling collected packages: distlib, pyyaml, platformdirs, nodeenv, identify, filelock, cfgv, virtualenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.9.18/x64|2f6023be5c7e4f75d9ed9a00c3fcb4c3fa7d8493a70885078264f9accc27f632\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nReceived 0 of 69094681 (0.0%), 0.0 MBs/sec\nCache Size: ~66 MB (69094681 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/2e80fd5c-4883-4bdb-be22-ed0a545b7db7/cache.tzst -P -C /home/runner/work/dask/dask --use-compress-program unzstd\nReceived 69094681 of 69094681 (100.0%), 32.9 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.9.18/x64|2f6023be5c7e4f75d9ed9a00c3fcb4c3fa7d8493a70885078264f9accc27f632\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nfix end of files.........................................................\u001b[42mPassed\u001b[m\ndebug statements (python)................................................\u001b[42mPassed\u001b[m\nabsolufy-imports.........................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\npyupgrade................................................................\u001b[42mPassed\u001b[m\nblack....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted dask/utils.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m257 files \u001b[0mleft unchanged.\n\nflake8...................................................................\u001b[42mPassed\u001b[m\ncodespell................................................................\u001b[42mPassed\u001b[m\nmypy.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: mypy\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\ndask/utils.py:231: error: Unsupported operand types for + (\"str\" and \"None\") [operator]\ndask/utils.py:231: note: Right operand is of type \"Optional[str]\"\ndask/utils.py:257: error: Unsupported operand types for + (\"str\" and \"None\") [operator]\ndask/utils.py:257: note: Right operand is of type \"Optional[str]\"\nFound 2 errors in 1 file (checked 258 source files)\n\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/dask/utils.py b/dask/utils.py\u001b[m\n\u001b[1mindex 80ea6e8..26f1176 100644\u001b[m\n\u001b[1m--- a/dask/utils.py\u001b[m\n\u001b[1m+++ b/dask/utils.py\u001b[m\n\u001b[36m@@ -148,7 +148,7 @@\u001b[m \u001b[mdef _deprecated_kwarg(\u001b[m\n new_arg_name: str | None = None,\u001b[m\n mapping: Mapping[Any, Any] | Callable[[Any], Any] | None = None,\u001b[m\n stacklevel: int = 2,\u001b[m\n\u001b[31m- comment: str | None = None\u001b[m\n\u001b[32m+\u001b[m\u001b[32m comment: str | None = None,\u001b[m\n ) -> Callable[[F], F]:\u001b[m\n \"\"\"\u001b[m\n Decorator to deprecate a keyword argument of a function.\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/dask/utils.py b/dask/utils.py\nindex 80ea6e8d6..0da2953c8 100644\n--- a/dask/utils.py\n+++ b/dask/utils.py\n@@ -148,7 +148,7 @@ def _deprecated_kwarg(\n new_arg_name: str | None = None,\n mapping: Mapping[Any, Any] | Callable[[Any], Any] | None = None,\n stacklevel: int = 2,\n- comment: str | None = None\n+ comment: str | None = None,\n ) -> Callable[[F], F]:\n \"\"\"\n Decorator to deprecate a keyword argument of a function.\n@@ -219,7 +219,7 @@ def _deprecated_kwarg(\n \"mapping from old to new argument values must be dict or callable!\"\n )\n \n- comment = f\"\\n{comment}\" or \"\"\n+ comment_ = f\"\\n{comment}\" or \"\"\n \n def _deprecated_kwarg(func: F) -> F:\n @wraps(func)\n@@ -232,7 +232,7 @@ def _deprecated_kwarg(\n f\"the {repr(old_arg_name)} keyword is deprecated and \"\n \"will be removed in a future version. Please take \"\n f\"steps to stop the use of {repr(old_arg_name)}\"\n- ) + comment\n+ ) + comment_\n warnings.warn(msg, FutureWarning, stacklevel=stacklevel)\n kwargs[old_arg_name] = old_arg_value\n return func(*args, **kwargs)\n@@ -254,7 +254,7 @@ def _deprecated_kwarg(\n f\"use {repr(new_arg_name)} instead.\"\n )\n \n- warnings.warn(msg + comment, FutureWarning, stacklevel=stacklevel)\n+ warnings.warn(msg + comment_, FutureWarning, stacklevel=stacklevel)\n if kwargs.get(new_arg_name) is not None:\n msg = (\n f\"Can only specify {repr(old_arg_name)} \"\n", "difficulty": 0, "changed_files": ["dask/utils.py"], "commit_link": "https://github.com/dask/dask/tree/16a0c04d06205527ec5e379df2596b399ee5dadc"}
 
 
data/python/1afe2c9.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 107, "repo_owner": "encode", "repo_name": "httpx", "head_branch": "add-ssl-context-argument", "workflow_name": "Test Suite", "workflow_filename": "test-suite.yml", "workflow_path": ".github/workflows/test-suite.yml", "contributor": "karpetrosyan", "sha_fail": "1afe2c9cb192d3760d59190cc7892e7ac37d5e27", "sha_success": "6be802e03bdb78d20f9c8df63a43f9167dcbfd49", "workflow": "---\nname: Test Suite\n\non:\n push:\n branches: [\"master\"]\n pull_request:\n branches: [\"master\"]\n\njobs:\n tests:\n name: \"Python ${{ matrix.python-version }}\"\n runs-on: \"ubuntu-latest\"\n\n strategy:\n matrix:\n python-version: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n\n steps:\n - uses: \"actions/checkout@v4\"\n - uses: \"actions/setup-python@v4\"\n with:\n python-version: \"${{ matrix.python-version }}\"\n allow-prereleases: true\n - name: \"Install dependencies\"\n run: \"scripts/install\"\n - name: \"Run linting checks\"\n run: \"scripts/check\"\n - name: \"Build package & docs\"\n run: \"scripts/build\"\n - name: \"Run tests\"\n run: \"scripts/test\"\n - name: \"Enforce coverage\"\n run: \"scripts/coverage\"\n", "logs": [{"step_name": "Python 3.10/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Python 3.11/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Python 3.12/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/httpx/_client.py b/httpx/_client.py\nindex e5cff5d..4dd2d07 100644\n--- a/httpx/_client.py\n+++ b/httpx/_client.py\n@@ -1455,8 +1455,6 @@ class AsyncClient(BaseClient):\n ) -> AsyncBaseTransport:\n return AsyncHTTPTransport(\n ssl_context=ssl_context,\n- verify=verify,\n- cert=cert,\n http1=http1,\n http2=http2,\n limits=limits,\n", "difficulty": 2, "changed_files": ["httpx/_client.py"], "commit_link": "https://github.com/encode/httpx/tree/1afe2c9cb192d3760d59190cc7892e7ac37d5e27"}
 
 
data/python/1b22eb7.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 88, "repo_owner": "hugsy", "repo_name": "gef", "head_branch": "main", "workflow_name": "CI Test for GEF", "workflow_filename": "run-tests.yml", "workflow_path": ".github/workflows/run-tests.yml", "contributor": "Angelo942", "sha_fail": "1b22eb71da460e7d3a301842f65c256b920f6195", "sha_success": "a5bb1bce7421428c7c5bd36344708c7a418baa87", "workflow": "name: CI Test for GEF\n\non:\n push:\n branches:\n - main\n\n pull_request:\n branches:\n - main\n\njobs:\n build:\n strategy:\n fail-fast: false\n matrix:\n os:\n - ubuntu-22.04\n - ubuntu-20.04\n name: \"Run Unit tests on ${{ matrix.os }}\"\n runs-on: ${{ matrix.os }}\n defaults:\n run:\n shell: bash\n\n steps:\n - uses: actions/checkout@v3\n\n - name: Install python and toolchain\n run: |\n sudo apt-get update\n sudo apt-get install -y gdb-multiarch python3-dev python3-pip python3-wheel python3-setuptools git cmake gcc g++ pkg-config libglib2.0-dev gdbserver qemu-user\n sudo python3 -m pip install --upgrade pip\n\n - name: Set runtime environment variables\n run: |\n echo PY_VER=`gdb -q -nx -ex \"pi print('.'.join(map(str, sys.version_info[:2])))\" -ex quit` >> $GITHUB_ENV\n echo GEF_CI_NB_CPU=`grep -c ^processor /proc/cpuinfo` >> $GITHUB_ENV\n echo GEF_CI_ARCH=`uname --processor` >> $GITHUB_ENV\n echo GEF_CI_CACHE_DIR=`python3 -m pip cache dir` >> $GITHUB_ENV\n\n - name: Cache dependencies\n uses: actions/cache@v3\n id: cache-deps\n env:\n cache-name: cache-deps\n with:\n key: ${{ runner.os }}-pip-${{ hashFiles('**/requirements.txt') }}\n path: ${{ env.GEF_CI_CACHE_DIR }}\n restore-keys:\n ${{ runner.os }}-pip-${{ env.cache-name }}-\n ${{ runner.os }}-pip-\n ${{ runner.os }}-${{ env.cache-name }}-\n ${{ runner.os }}-\n\n - name: Install Python Requirements\n run: |\n mkdir -p ${{ env.GEF_CI_CACHE_DIR }}\n python${{ env.PY_VER }} -m pip install --user --upgrade -r tests/requirements.txt\n\n - name: Setup GEF\n run: |\n echo \"source $(pwd)/gef.py\" > ~/.gdbinit\n gdb -q -ex 'gef missing' -ex 'gef help' -ex 'gef config' -ex start -ex continue -ex quit /bin/pwd\n\n - name: Run Tests\n run: |\n make -C tests/binaries -j ${{ env.GEF_CI_NB_CPU }}\n python${{ env.PY_VER }} -m pytest --forked -n ${{ env.GEF_CI_NB_CPU }} -v -k \"not benchmark\" tests/\n", "logs": [{"step_name": "Run Unit tests on ubuntu-22.04/8_Run Tests.txt", "log": "##[group]Run make -C tests/binaries -j 4\n\u001b[36;1mmake -C tests/binaries -j 4\u001b[0m\n\u001b[36;1mpython3.10 -m pytest --forked -n 4 -v -k \"not benchmark\" tests/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PY_VER: 3.10\n GEF_CI_NB_CPU: 4\n GEF_CI_ARCH: x86_64\n GEF_CI_CACHE_DIR: /home/runner/.cache/pip\n##[endgroup]\nmake: Entering directory '/home/runner/work/gef/gef/tests/binaries'\n[+] Building '/tmp/bss.out'\n[+] Building '/tmp/canary.out'\n[+] Building '/tmp/checksec-no-canary.out'\n[+] Building '/tmp/checksec-no-nx.out'\n[+] Building '/tmp/checksec-no-pie.out'\n[+] Building '/tmp/default.out'\n[+] Building '/tmp/format-string-helper.out'\n[+] Building '/tmp/heap-analysis.out'\n[+] Building '/tmp/heap-bins.out'\n[+] Building '/tmp/heap-fastbins.out'\n[+] Building '/tmp/heap-multiple-heaps.out'\n[+] Building '/tmp/heap-non-main.out'\n[+] Building '/tmp/heap-tcache.out'\n[+] Building '/tmp/heap.out'\n[+] Building '/tmp/memwatch.out'\n[+] Building '/tmp/mmap-known-address.out'\n[+] Building '/tmp/nested.out'\n[+] Building '/tmp/nested2.out'\n[+] Building '/tmp/pattern.out'\n[+] Building '/tmp/pcustom.out'\n[+] Building '/tmp/class.out'\nmake: Leaving directory '/home/runner/work/gef/gef/tests/binaries'\n/home/runner/.local/lib/python3.10/site-packages/pytest_benchmark/logger.py:46: PytestBenchmarkWarning: Benchmarks are automatically disabled because xdist plugin is active.Benchmarks cannot be performed reliably in a parallelized environment.\n warner(PytestBenchmarkWarning(text))\n============================= test session starts ==============================\nplatform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.3.0 -- /usr/bin/python3.10\ncachedir: .pytest_cache\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/gef/gef/tests\nconfigfile: pytest.ini\nplugins: benchmark-4.0.0, forked-1.6.0, xdist-3.5.0, cov-4.1.0\ncreated: 4/4 workers\n4 workers [155 items]\n\nscheduling tests via LoadScheduling\n\ntests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \n[gw1] [ 0%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/gef_session.py::GefSessionApi::test_func_get_filepath \n[gw3] [ 1%] PASSED tests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \n[gw1] [ 1%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_filepath \ntests/api/gef_session.py::GefSessionApi::test_func_get_pid \n[gw2] [ 2%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \n[gw1] [ 3%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_pid \ntests/api/gef_session.py::GefSessionApi::test_root_dir_local \n[gw3] [ 3%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \n[gw1] [ 4%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_local \ntests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \n[gw2] [ 5%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \ntests/api/misc.py::MiscFunctionTest::test_func_update_gef \n[gw2] [ 5%] SKIPPED tests/api/misc.py::MiscFunctionTest::test_func_update_gef \ntests/api/misc.py::MiscFunctionTest::test_func_which \n[gw1] [ 6%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \ntests/api/gef_session.py::GefSessionApi::test_root_dir_remote \n[gw3] [ 7%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \n[gw2] [ 7%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_which \ntests/commands/aliases.py::AliasesCommand::test_cmd_aliases \n[gw3] [ 8%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \n[gw1] [ 9%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_remote \ntests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \n[gw1] [ 9%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \ntests/api/misc.py::MiscFunctionTest::test_func_parse_address \n[gw2] [ 10%] PASSED tests/commands/aliases.py::AliasesCommand::test_cmd_aliases \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \n[gw3] [ 10%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \n[gw0] [ 11%] PASSED tests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \n[gw2] [ 12%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \n[gw1] [ 12%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_address \ntests/api/misc.py::MiscFunctionTest::test_func_parse_maps \n[gw0] [ 13%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \ntests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \n[gw3] [ 14%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \n[gw0] [ 14%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \n[gw2] [ 15%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \ntests/commands/canary.py::CanaryCommand::test_cmd_canary \n[gw0] [ 16%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \n[gw3] [ 16%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \ntests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \n[gw0] [ 17%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \ntests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \n[gw3] [ 18%] PASSED tests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \ntests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \n[gw2] [ 18%] PASSED tests/commands/canary.py::CanaryCommand::test_cmd_canary \ntests/commands/canary.py::CanaryCommand::test_overwrite_canary \n[gw3] [ 19%] PASSED tests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \n[gw2] [ 20%] PASSED tests/commands/canary.py::CanaryCommand::test_overwrite_canary \ntests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \n[gw0] [ 20%] PASSED tests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \n[gw2] [ 21%] PASSED tests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \ntests/commands/functions.py::FunctionsCommand::test_cmd_functions \n[gw2] [ 21%] PASSED tests/commands/functions.py::FunctionsCommand::test_cmd_functions \ntests/commands/gef.py::GefCommand::test_cmd_gef \n[gw1] [ 22%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_maps \ntests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \n[gw2] [ 23%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef \ntests/commands/gef.py::GefCommand::test_cmd_gef_config \n[gw3] [ 23%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \n[gw2] [ 24%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_get \n[gw1] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \ntests/commands/gef.py::GefCommand::test_cmd_gef_save \n[gw2] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_get \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_set \n[gw3] [ 26%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_set \ntests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw1] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_save \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \n[gw2] [ 28%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_help \ntests/commands/gef.py::GefCommand::test_cmd_gef_install \n[gw3] [ 29%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \n[gw2] [ 29%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_install \ntests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \n[gw0] [ 30%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \n[gw3] [ 30%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \n[gw1] [ 31%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \n[gw2] [ 32%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \ntests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \n[gw1] [ 32%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \n[gw2] [ 33%] PASSED tests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \ntests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \n[gw3] [ 34%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \n[gw1] [ 34%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \ntests/commands/got.py::GotCommand::test_cmd_got \n[gw0] [ 35%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \n[gw1] [ 36%] PASSED tests/commands/got.py::GotCommand::test_cmd_got \ntests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \n[gw3] [ 36%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \n[gw2] [ 37%] PASSED tests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \ntests/commands/highlight.py::HighlightCommand::test_cmd_highlight \n[gw1] [ 38%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \n[gw2] [ 38%] PASSED tests/commands/highlight.py::HighlightCommand::test_cmd_highlight \ntests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \n[gw0] [ 39%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \ntests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \n[gw2] [ 40%] PASSED tests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \ntests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \n[gw1] [ 40%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \n[gw3] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \n[gw2] [ 41%] PASSED tests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_list \n[gw1] [ 42%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \n[gw0] [ 43%] PASSED tests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \ntests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \n[gw3] [ 43%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \n[gw1] [ 44%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \n[gw2] [ 45%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_list \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \n[gw1] [ 45%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \n[gw0] [ 46%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes \n[gw3] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \n[gw2] [ 47%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \n[gw1] [ 48%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \ntests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \n[gw0] [ 49%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw1] [ 49%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \ntests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \n[gw0] [ 50%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw2] [ 50%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \n[gw3] [ 51%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \n[gw1] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \n[gw0] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \n[gw1] [ 53%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \n[gw0] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \n[gw1] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \ntests/commands/nop.py::NopCommand::test_cmd_nop_inactive \n[gw0] [ 55%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \n[gw3] [ 56%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \n[gw1] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_inactive \ntests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \n[gw0] [ 57%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \n[gw3] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch \n[gw1] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \n[gw3] [ 59%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte \n[gw1] [ 60%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \n[gw3] [ 60%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \n[gw0] [ 61%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \ntests/commands/patch.py::PatchCommand::test_cmd_patch_string \n[gw3] [ 61%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \ntests/commands/patch.py::PatchCommand::test_cmd_patch_dword \n[gw0] [ 62%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_string \ntests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw1] [ 63%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \n[gw3] [ 63%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_dword \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword \n[gw0] [ 64%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_word \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \n[gw2] [ 65%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \n[gw3] [ 65%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \n[gw0] [ 66%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \n[gw2] [ 67%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop \n[gw3] [ 67%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \n[gw1] [ 68%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \ntests/commands/pie.py::PieCommand::test_cmd_pie \n[gw2] [ 69%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \n[gw3] [ 69%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \n[gw2] [ 70%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \n[gw1] [ 70%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \n[gw1] [ 71%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \ntests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \n[gw2] [ 72%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \n[gw1] [ 72%] PASSED tests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \ntests/commands/scan.py::ScanCommand::test_cmd_scan \n[gw2] [ 73%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \n[gw0] [ 74%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \ntests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \n[gw2] [ 74%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \n[gw3] [ 75%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \n[gw2] [ 76%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \n[gw1] [ 76%] PASSED tests/commands/scan.py::ScanCommand::test_cmd_scan \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \n[gw0] [ 77%] PASSED tests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \ntests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \n[gw3] [ 78%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 78%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \n[gw1] [ 79%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \ntests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \n[gw0] [ 80%] PASSED tests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \ntests/commands/registers.py::RegistersCommand::test_cmd_registers \n[gw1] [ 80%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \n[gw3] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \n[gw2] [ 82%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \n[gw1] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \ntests/commands/stub.py::StubCommand::test_cmd_stub \n[gw0] [ 83%] PASSED tests/commands/registers.py::RegistersCommand::test_cmd_registers \ntests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \n[gw3] [ 84%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \ntests/commands/theme.py::ThemeCommand::test_cmd_theme \n[gw2] [ 85%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \ntests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \n[gw1] [ 85%] PASSED tests/commands/stub.py::StubCommand::test_cmd_stub \ntests/commands/version.py::VersionCommand::test_cmd_version \n[gw2] [ 86%] PASSED tests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \n[gw1] [ 87%] PASSED tests/commands/version.py::VersionCommand::test_cmd_version \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \n[gw0] [ 87%] PASSED tests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \ntests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \n[gw1] [ 88%] FAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n[gw2] [ 89%] PASSED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \n[gw1] [ 89%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \ntests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \n[gw0] [ 90%] PASSED tests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \ntests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \n[gw2] [ 90%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \ntests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \n[gw0] [ 91%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \n[gw2] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \n[gw1] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \ntests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \n[gw2] [ 93%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \n[gw0] [ 94%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \n[gw1] [ 94%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw0] [ 95%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw2] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \ntests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw1] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw0] [ 97%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw1] [ 98%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw2] [ 98%] PASSED tests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw3] [ 99%] PASSED tests/commands/theme.py::ThemeCommand::test_cmd_theme \ntests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n[gw3] [100%] PASSED tests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n\n=================================== FAILURES ===================================\n_____________________ XinfoCommand.test_cmd_xinfo_on_class _____________________\n[gw1] linux -- Python 3.10.12 /usr/bin/python3.10\nself = <tests.commands.xinfo.XinfoCommand testMethod=test_cmd_xinfo_on_class>\n\n def test_cmd_xinfo_on_class(self):\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n self.assertNoException(res)\n> self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\nE AssertionError: 'Symbol: B<TraitA, TraitB>::Run()+4' not found in 'Reading symbols from /tmp/class.out...\\nError while writing index for `/tmp/class.out\\': mkstemp: No such file or directory.\\nGEF for linux ready, type `gef\\' to start, `gef config\\' to configure\\n89 commands loaded and 5 functions added for GDB 12.1 in 0.00ms using Python engine 3.10\\nBreakpoint 1 at 0x1328: file class.cpp, line 21.\\nStarting program: /tmp/class.out \\n[Thread debugging using libthread_db enabled]\\nUsing host libthread_db library \"/lib/x86_64-linux-gnu/libthread_db.so.1\".\\n\\nBreakpoint 1, B<TraitA, TraitB>::Run (this=0x55555556b2b0) at class.cpp:21\\n21\\t virtual void Run() { printf(\"I am B\\\\n\"); }\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500 xinfo: 0x55555555532c \u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\nPage: 0x0000555555555000 \u2192 0x0000555555556000 (size=0x1000)\\nPermissions: r-x\\nPathname: /tmp/class.out\\nOffset (from page): 0x32c\\nInode: 37098\\nSegment: .text (0x00005555555550a0-0x000055555555533a)\\nOffset (from segment): 0x28c\\nSymbol: B<TraitA, TraitB>::Run()+20'\n\ntests/commands/xinfo.py:27: AssertionError\n=========================== short test summary info ============================\nFAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class\n================== 1 failed, 153 passed, 1 skipped in 52.95s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Run Unit tests on ubuntu-20.04/8_Run Tests.txt", "log": "##[group]Run make -C tests/binaries -j 4\n\u001b[36;1mmake -C tests/binaries -j 4\u001b[0m\n\u001b[36;1mpython3.8 -m pytest --forked -n 4 -v -k \"not benchmark\" tests/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PY_VER: 3.8\n GEF_CI_NB_CPU: 4\n GEF_CI_ARCH: x86_64\n GEF_CI_CACHE_DIR: /home/runner/.cache/pip\n##[endgroup]\nmake: Entering directory '/home/runner/work/gef/gef/tests/binaries'\n[+] Building '/tmp/heap.out'\n[+] Building '/tmp/bss.out'\n[+] Building '/tmp/canary.out'\n[+] Building '/tmp/heap-multiple-heaps.out'\n[+] Building '/tmp/checksec-no-canary.out'\n[+] Building '/tmp/format-string-helper.out'\n[+] Building '/tmp/checksec-no-nx.out'\n[+] Building '/tmp/heap-non-main.out'\n[+] Building '/tmp/nested2.out'\n[+] Building '/tmp/mmap-known-address.out'\n[+] Building '/tmp/heap-analysis.out'\n[+] Building '/tmp/default.out'\n[+] Building '/tmp/heap-tcache.out'\n[+] Building '/tmp/nested.out'\n[+] Building '/tmp/checksec-no-pie.out'\n[+] Building '/tmp/heap-bins.out'\n[+] Building '/tmp/pattern.out'\n[+] Building '/tmp/memwatch.out'\n[+] Building '/tmp/pcustom.out'\n[+] Building '/tmp/heap-fastbins.out'\n[+] Building '/tmp/class.out'\nmake: Leaving directory '/home/runner/work/gef/gef/tests/binaries'\n============================= test session starts ==============================\nplatform linux -- Python 3.8.10, pytest-7.4.4, pluggy-1.3.0 -- /usr/bin/python3.8\ncachedir: .pytest_cache\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/gef/gef/tests\nconfigfile: pytest.ini\nplugins: forked-1.6.0, xdist-3.5.0, benchmark-4.0.0, cov-4.1.0\ncreated: 4/4 workers\n4 workers [155 items]\n\nscheduling tests via LoadScheduling\n\ntests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \n[gw1] [ 0%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/gef_session.py::GefSessionApi::test_func_get_filepath \n[gw3] [ 1%] PASSED tests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \n[gw1] [ 1%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_filepath \ntests/api/gef_session.py::GefSessionApi::test_func_get_pid \n[gw2] [ 2%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \n[gw1] [ 3%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_pid \ntests/api/gef_session.py::GefSessionApi::test_root_dir_local \n[gw3] [ 3%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \n[gw1] [ 4%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_local \ntests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \n[gw2] [ 5%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \ntests/api/misc.py::MiscFunctionTest::test_func_update_gef \n[gw2] [ 5%] SKIPPED tests/api/misc.py::MiscFunctionTest::test_func_update_gef \ntests/api/misc.py::MiscFunctionTest::test_func_which \n[gw1] [ 6%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \ntests/api/gef_session.py::GefSessionApi::test_root_dir_remote \n[gw3] [ 7%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \n[gw2] [ 7%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_which \ntests/commands/aliases.py::AliasesCommand::test_cmd_aliases \n[gw3] [ 8%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \n[gw1] [ 9%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_remote \ntests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \n[gw1] [ 9%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \ntests/api/misc.py::MiscFunctionTest::test_func_parse_address \n[gw3] [ 10%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \n[gw0] [ 10%] PASSED tests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \n[gw2] [ 11%] PASSED tests/commands/aliases.py::AliasesCommand::test_cmd_aliases \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \n[gw0] [ 12%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \ntests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \n[gw1] [ 12%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_address \ntests/api/misc.py::MiscFunctionTest::test_func_parse_maps \n[gw2] [ 13%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \n[gw3] [ 14%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \n[gw0] [ 14%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \n[gw0] [ 15%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \n[gw2] [ 16%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \ntests/commands/canary.py::CanaryCommand::test_cmd_canary \n[gw3] [ 16%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \ntests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \n[gw0] [ 17%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \ntests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \n[gw3] [ 18%] PASSED tests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \ntests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \n[gw2] [ 18%] PASSED tests/commands/canary.py::CanaryCommand::test_cmd_canary \ntests/commands/canary.py::CanaryCommand::test_overwrite_canary \n[gw3] [ 19%] PASSED tests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \n[gw0] [ 20%] PASSED tests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \n[gw2] [ 20%] PASSED tests/commands/canary.py::CanaryCommand::test_overwrite_canary \ntests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \n[gw2] [ 21%] PASSED tests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \ntests/commands/functions.py::FunctionsCommand::test_cmd_functions \n[gw2] [ 21%] PASSED tests/commands/functions.py::FunctionsCommand::test_cmd_functions \ntests/commands/gef.py::GefCommand::test_cmd_gef \n[gw1] [ 22%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_maps \ntests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \n[gw2] [ 23%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef \ntests/commands/gef.py::GefCommand::test_cmd_gef_config \n[gw3] [ 23%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \n[gw2] [ 24%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_get \n[gw1] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \ntests/commands/gef.py::GefCommand::test_cmd_gef_save \n[gw2] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_get \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_set \n[gw3] [ 26%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_set \ntests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw1] [ 28%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_save \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef.py::GefCommand::test_cmd_gef_install \n[gw3] [ 29%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \n[gw0] [ 29%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \n[gw3] [ 30%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \n[gw2] [ 30%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_install \ntests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \n[gw1] [ 31%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \n[gw2] [ 32%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \ntests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \n[gw1] [ 32%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \n[gw0] [ 33%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \n[gw2] [ 34%] PASSED tests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \ntests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \n[gw3] [ 34%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \n[gw1] [ 35%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \ntests/commands/got.py::GotCommand::test_cmd_got \n[gw3] [ 36%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \n[gw1] [ 36%] PASSED tests/commands/got.py::GotCommand::test_cmd_got \ntests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \n[gw0] [ 37%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \ntests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \n[gw2] [ 38%] PASSED tests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \ntests/commands/highlight.py::HighlightCommand::test_cmd_highlight \n[gw1] [ 38%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \n[gw2] [ 39%] PASSED tests/commands/highlight.py::HighlightCommand::test_cmd_highlight \ntests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \n[gw0] [ 40%] PASSED tests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \ntests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \n[gw2] [ 40%] PASSED tests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \ntests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \n[gw3] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \n[gw1] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \n[gw2] [ 42%] PASSED tests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_list \n[gw1] [ 43%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \n[gw0] [ 43%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes \n[gw3] [ 44%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \n[gw2] [ 45%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_list \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \n[gw1] [ 45%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \n[gw0] [ 46%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw3] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \n[gw1] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \n[gw0] [ 48%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \n[gw2] [ 49%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \n[gw1] [ 49%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \ntests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \n[gw3] [ 50%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \n[gw0] [ 50%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \n[gw2] [ 51%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \n[gw1] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \ntests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \n[gw0] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \n[gw1] [ 53%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \n[gw0] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \n[gw1] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \n[gw3] [ 55%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \n[gw0] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \n[gw1] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \ntests/commands/nop.py::NopCommand::test_cmd_nop_inactive \n[gw3] [ 57%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop \n[gw1] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_inactive \ntests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \n[gw0] [ 58%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \ntests/commands/patch.py::PatchCommand::test_cmd_patch_string \n[gw3] [ 59%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \n[gw1] [ 60%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \n[gw0] [ 60%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_string \ntests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw3] [ 61%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \n[gw1] [ 61%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \n[gw0] [ 62%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw3] [ 63%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \n[gw1] [ 63%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \n[gw3] [ 64%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \n[gw0] [ 65%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \n[gw3] [ 65%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \n[gw2] [ 66%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \n[gw3] [ 67%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \n[gw2] [ 67%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch \n[gw3] [ 68%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \n[gw2] [ 69%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte \n[gw1] [ 69%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \ntests/commands/pie.py::PieCommand::test_cmd_pie \n[gw2] [ 70%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \n[gw1] [ 70%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \n[gw2] [ 71%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \ntests/commands/patch.py::PatchCommand::test_cmd_patch_dword \n[gw0] [ 72%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \ntests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \n[gw1] [ 72%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \ntests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \n[gw2] [ 73%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_dword \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword \n[gw1] [ 74%] PASSED tests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \ntests/commands/scan.py::ScanCommand::test_cmd_scan \n[gw2] [ 74%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \n[gw3] [ 75%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \n[gw0] [ 76%] PASSED tests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \ntests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \n[gw2] [ 76%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \n[gw3] [ 77%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw0] [ 78%] PASSED tests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \ntests/commands/registers.py::RegistersCommand::test_cmd_registers \n[gw2] [ 78%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \n[gw1] [ 79%] PASSED tests/commands/scan.py::ScanCommand::test_cmd_scan \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \n[gw0] [ 80%] PASSED tests/commands/registers.py::RegistersCommand::test_cmd_registers \ntests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \n[gw3] [ 80%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \n[gw1] [ 81%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \ntests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \n[gw0] [ 82%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \n[gw2] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \n[gw0] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \ntests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \n[gw3] [ 84%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \ntests/commands/stub.py::StubCommand::test_cmd_stub \n[gw2] [ 85%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \ntests/commands/version.py::VersionCommand::test_cmd_version \n[gw2] [ 85%] PASSED tests/commands/version.py::VersionCommand::test_cmd_version \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \n[gw0] [ 86%] PASSED tests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \ntests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \n[gw1] [ 87%] PASSED tests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \ntests/commands/theme.py::ThemeCommand::test_cmd_theme \n[gw3] [ 87%] PASSED tests/commands/stub.py::StubCommand::test_cmd_stub \ntests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n[gw2] [ 88%] PASSED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \n[gw3] [ 89%] PASSED tests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \ntests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \n[gw0] [ 89%] PASSED tests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \n[gw2] [ 90%] FAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \ntests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \n[gw3] [ 90%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \ntests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \n[gw0] [ 91%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \ntests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \n[gw3] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \n[gw2] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \n[gw3] [ 93%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \n[gw0] [ 94%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \n[gw0] [ 94%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw2] [ 95%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw3] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \ntests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw0] [ 96%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw2] [ 97%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw0] [ 98%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw3] [ 98%] PASSED tests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw1] [ 99%] PASSED tests/commands/theme.py::ThemeCommand::test_cmd_theme \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n[gw1] [100%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n\n=================================== FAILURES ===================================\n_____________________ XinfoCommand.test_cmd_xinfo_on_class _____________________\n[gw2] linux -- Python 3.8.10 /usr/bin/python3.8\nself = <tests.commands.xinfo.XinfoCommand testMethod=test_cmd_xinfo_on_class>\n\n def test_cmd_xinfo_on_class(self):\n/home/runner/.local/lib/python3.8/site-packages/pytest_benchmark/logger.py:46: PytestBenchmarkWarning: Benchmarks are automatically disabled because xdist plugin is active.Benchmarks cannot be performed reliably in a parallelized environment.\n warner(PytestBenchmarkWarning(text))\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n self.assertNoException(res)\n> self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\nE AssertionError: 'Symbol: B<TraitA, TraitB>::Run()+4' not found in 'Reading symbols from /tmp/class.out...\\nError while writing index for `/tmp/class.out\\': mkstemp: No such file or directory.\\nGEF for linux ready, type `gef\\' to start, `gef config\\' to configure\\n89 commands loaded and 5 functions added for GDB 10.2 in 0.00ms using Python engine 3.8\\nBreakpoint 1 at 0x1328: file class.cpp, line 21.\\nStarting program: /tmp/class.out \\n\\nBreakpoint 1, B<TraitA, TraitB>::Run (this=0x55555556aeb0) at class.cpp:21\\n21\\t virtual void Run() { printf(\"I am B\\\\n\"); }\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500 xinfo: 0x55555555532c \u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\nPage: 0x0000555555555000 \u2192 0x0000555555556000 (size=0x1000)\\nPermissions: r-x\\nPathname: /tmp/class.out\\nOffset (from page): 0x32c\\nInode: 36000\\nSegment: .text (0x00005555555550a0-0x00005555555553b5)\\nOffset (from segment): 0x28c\\nSymbol: B<TraitA, TraitB>::Run()+20'\n\ntests/commands/xinfo.py:27: AssertionError\n=========================== short test summary info ============================\nFAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class\n================== 1 failed, 153 passed, 1 skipped in 49.25s ===================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/tests/commands/xinfo.py b/tests/commands/xinfo.py\nindex 625615b..ea50e23 100644\n--- a/tests/commands/xinfo.py\n+++ b/tests/commands/xinfo.py\n@@ -22,6 +22,6 @@ class XinfoCommand(GefUnitTestGeneric):\n def test_cmd_xinfo_on_class(self):\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n- res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n+ res = gdb_run_silent_cmd(cmd, target=target, before=[\"b *'B<TraitA, TraitB>::Run'\"])\n self.assertNoException(res)\n self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\n", "difficulty": 2, "changed_files": ["tests/commands/xinfo.py"], "commit_link": "https://github.com/hugsy/gef/tree/1b22eb71da460e7d3a301842f65c256b920f6195"}
 
 
data/python/1e48195.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 120, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "update/diffusers", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "wandb", "sha_fail": "1e48195d38e0d8de001e0f55bfbc830abf2e3f41", "sha_success": "5c24819003ab2cf1fdaa07e78a120fbf53875392", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.3-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 31.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 87.6 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 116.8 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 49.6 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.3-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 92.1 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.3 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.3\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nReceived 0 of 48931996 (0.0%), 0.0 MBs/sec\nCache Size: ~47 MB (48931996 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/2e6b382b-db1b-4411-add2-ddbcf402d5ce/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48931996 of 48931996 (100.0%), 23.3 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nclippy...................................................................\u001b[42mPassed\u001b[m\nruff.....................................................................\u001b[42mPassed\u001b[m\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted wandb/integration/diffusers/autologger.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m887 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/wandb/integration/diffusers/autologger.py b/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[1mindex ab88d9c..e10ee15 100644\u001b[m\n\u001b[1m--- a/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[1m+++ b/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[36m@@ -65,7 +65,7 @@\u001b[m \u001b[mautolog = AutologAPI(\u001b[m\n \"TextToVideoSDPipeline.__call__\",\u001b[m\n \"TextToVideoZeroPipeline.__call__\",\u001b[m\n \"StableVideoDiffusionPipeline.__call__\",\u001b[m\n\u001b[31m- \"AmusedPipeline.__call__\"\u001b[m\n\u001b[32m+\u001b[m\u001b[32m \"AmusedPipeline.__call__\",\u001b[m\n ),\u001b[m\n resolver=DiffusersPipelineResolver(),\u001b[m\n telemetry_feature=\"diffusers_autolog\",\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/wandb/integration/diffusers/autologger.py b/wandb/integration/diffusers/autologger.py\nindex ab88d9c7d..e10ee1549 100644\n--- a/wandb/integration/diffusers/autologger.py\n+++ b/wandb/integration/diffusers/autologger.py\n@@ -65,7 +65,7 @@ autolog = AutologAPI(\n \"TextToVideoSDPipeline.__call__\",\n \"TextToVideoZeroPipeline.__call__\",\n \"StableVideoDiffusionPipeline.__call__\",\n- \"AmusedPipeline.__call__\"\n+ \"AmusedPipeline.__call__\",\n ),\n resolver=DiffusersPipelineResolver(),\n telemetry_feature=\"diffusers_autolog\",\ndiff --git a/wandb/integration/diffusers/pipeline_resolver.py b/wandb/integration/diffusers/pipeline_resolver.py\nindex 37cf75fd7..9d3736d2d 100644\n--- a/wandb/integration/diffusers/pipeline_resolver.py\n+++ b/wandb/integration/diffusers/pipeline_resolver.py\n@@ -18,6 +18,7 @@ class DiffusersPipelineResolver:\n \n def __init__(self) -> None:\n self.wandb_table = None\n+ self.pipeline_call_count = 1\n \n def __call__(\n self,\n@@ -43,7 +44,10 @@ class DiffusersPipelineResolver:\n pipeline_name = args[0].__class__.__name__\n resolver = None\n if pipeline_name in SUPPORTED_MULTIMODAL_PIPELINES:\n- resolver = DiffusersMultiModalPipelineResolver(pipeline_name)\n+ resolver = DiffusersMultiModalPipelineResolver(\n+ pipeline_name, self.pipeline_call_count\n+ )\n+ self.pipeline_call_count += 1\n elif pipeline_name in SUPPORTED_SDXL_PIPELINES:\n resolver = SDXLResolver(pipeline_name)\n loggable_dict = resolver(args, kwargs, response, start_time, time_elapsed)\ndiff --git a/wandb/integration/diffusers/resolvers/multimodal.py b/wandb/integration/diffusers/resolvers/multimodal.py\nindex 2948778f9..021efcf2c 100644\n--- a/wandb/integration/diffusers/resolvers/multimodal.py\n+++ b/wandb/integration/diffusers/resolvers/multimodal.py\n@@ -502,8 +502,9 @@ class DiffusersMultiModalPipelineResolver:\n pipeline_name: (str) The name of the Diffusion Pipeline.\n \"\"\"\n \n- def __init__(self, pipeline_name: str) -> None:\n+ def __init__(self, pipeline_name: str, pipeline_call_count: int) -> None:\n self.pipeline_name = pipeline_name\n+ self.pipeline_call_count = pipeline_call_count\n columns = []\n if pipeline_name in SUPPORTED_MULTIMODAL_PIPELINES:\n columns += SUPPORTED_MULTIMODAL_PIPELINES[pipeline_name][\"table-schema\"]\n@@ -543,9 +544,30 @@ class DiffusersMultiModalPipelineResolver:\n pipeline_configs = dict(pipeline.config)\n pipeline_configs[\"pipeline-name\"] = self.pipeline_name\n \n- wandb.config.update(\n- {\"workflow\": {\"pipeline\": pipeline_configs, \"params\": kwargs}}\n- )\n+ if \"workflow\" not in wandb.config:\n+ wandb.config.update(\n+ {\n+ \"workflow\": [\n+ {\n+ \"pipeline\": pipeline_configs,\n+ \"params\": kwargs,\n+ \"stage\": f\"Pipeline-Call-{self.pipeline_call_count}\",\n+ }\n+ ]\n+ }\n+ )\n+ else:\n+ existing_workflow = wandb.config.workflow\n+ updated_workflow = existing_workflow + [\n+ {\n+ \"pipeline\": pipeline_configs,\n+ \"params\": kwargs,\n+ \"stage\": f\"Pipeline-Call-{self.pipeline_call_count}\",\n+ }\n+ ]\n+ wandb.config.update(\n+ {\"workflow\": updated_workflow}, allow_val_change=True\n+ )\n \n # Return the WandB loggable dict\n loggable_dict = self.prepare_loggable_dict(response, kwargs)\n@@ -596,7 +618,13 @@ class DiffusersMultiModalPipelineResolver:\n caption = loggable_kwarg_chunks[prompt_index][idx]\n except ValueError:\n caption = None\n- wandb.log({\"Generated-Image\": wandb.Image(image, caption=caption)})\n+ wandb.log(\n+ {\n+ f\"Generated-Image/Pipeline-Call-{self.pipeline_call_count}\": wandb.Image(\n+ image, caption=caption\n+ )\n+ }\n+ )\n else:\n if (\n SUPPORTED_MULTIMODAL_PIPELINES[self.pipeline_name][\"output-type\"]\n@@ -611,7 +639,7 @@ class DiffusersMultiModalPipelineResolver:\n caption = None\n wandb.log(\n {\n- \"Generated-Video\": wandb.Video(\n+ f\"Generated-Video/Pipeline-Call-{self.pipeline_call_count}\": wandb.Video(\n postprocess_pils_to_np(image), fps=4, caption=caption\n )\n }\n@@ -629,7 +657,7 @@ class DiffusersMultiModalPipelineResolver:\n caption = None\n wandb.log(\n {\n- \"Generated-Audio\": wandb.Audio(\n+ f\"Generated-Audio/Pipeline-Call-{self.pipeline_call_count}\": wandb.Audio(\n image, sample_rate=16000, caption=caption\n )\n }\n@@ -696,7 +724,11 @@ class DiffusersMultiModalPipelineResolver:\n images, normalize=self.pipeline_name == \"TextToVideoZeroPipeline\"\n )\n wandb.log(\n- {\"Generated-Video\": wandb.Video(video, fps=4, caption=kwargs[\"prompt\"])}\n+ {\n+ f\"Generated-Video/Pipeline-Call-{self.pipeline_call_count}\": wandb.Video(\n+ video, fps=4, caption=kwargs[\"prompt\"]\n+ )\n+ }\n )\n loggable_kwarg_ids = SUPPORTED_MULTIMODAL_PIPELINES[self.pipeline_name][\n \"kwarg-logging\"\n@@ -727,5 +759,6 @@ class DiffusersMultiModalPipelineResolver:\n self.log_media(image, loggable_kwarg_chunks, idx)\n # Populate the row of the wandb_table\n self.add_data_to_table(image, loggable_kwarg_chunks, idx)\n-\n- return {\"Result-Table\": self.wandb_table}\n+ return {\n+ f\"Result-Table/Pipeline-Call-{self.pipeline_call_count}\": self.wandb_table\n+ }\n", "difficulty": 0, "changed_files": ["wandb/integration/diffusers/autologger.py", "wandb/integration/diffusers/pipeline_resolver.py", "wandb/integration/diffusers/resolvers/multimodal.py"], "commit_link": "https://github.com/wandb/wandb/tree/1e48195d38e0d8de001e0f55bfbc830abf2e3f41"}
 
 
data/python/212ff27.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 52, "repo_owner": "yt-dlp", "repo_name": "yt-dlp", "head_branch": "mx3", "workflow_name": "Quick Test", "workflow_filename": "quick-test.yml", "workflow_path": ".github/workflows/quick-test.yml", "contributor": "martinxyz", "sha_fail": "212ff27cb59b69bad24a2e675d97e52ead5b12e8", "sha_success": "1cb1df5d401172bb7b6770605a72fdb9bd03b0a8", "workflow": "name: Quick Test\non: [push, pull_request]\npermissions:\n contents: read\n\njobs:\n tests:\n name: Core Test\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - name: Set up Python 3.11\n uses: actions/setup-python@v4\n with:\n python-version: '3.11'\n - name: Install test requirements\n run: pip install pytest -r requirements.txt\n - name: Run tests\n run: |\n python3 -m yt_dlp -v || true\n ./devscripts/run_tests.sh core\n flake8:\n name: Linter\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v4\n - name: Install flake8\n run: pip install flake8\n - name: Make lazy extractors\n run: python devscripts/make_lazy_extractors.py\n - name: Run flake8\n run: flake8 .\n", "logs": [{"step_name": "Core Test/5_Run tests.txt", "log": "##[group]Run python3 -m yt_dlp -v || true\n\u001b[36;1mpython3 -m yt_dlp -v || true\u001b[0m\n\u001b[36;1mpython3 ./devscripts/run_tests.py core\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n[debug] Command-line config: ['-v']\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/runpy.py\", line 194, in _run_module_as_main\n return _run_code(code, main_globals, None,\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__main__.py\", line 17, in <module>\n yt_dlp.main()\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__init__.py\", line 1009, in main\n _exit(*variadic(_real_main(argv)))\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__init__.py\", line 963, in _real_main\n with YoutubeDL(ydl_opts) as ydl:\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/YoutubeDL.py\", line 689, in __init__\n self.print_debug_header()\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/YoutubeDL.py\", line 3919, in print_debug_header\n from .extractor.extractors import _LAZY_LOADER\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/extractors.py\", line 17, in <module>\n from ._extractors import * # noqa: F403\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/_extractors.py\", line 1140, in <module>\n from .mx3 import (\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 80, in <module>\n class Mx3IE(Mx3BaseIE):\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 82, in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\nNameError: name '_Mx3BaseIE' is not defined\nRunning ['pytest', '-Werror', '--tb=short', '--color=yes', '-m', 'not download']\n\u001b[1m============================= test session starts ==============================\u001b[0m\nplatform linux -- Python 3.8.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yt-dlp/yt-dlp\nconfigfile: setup.cfg\n\u001b[1mcollecting ... \u001b[0mcollected 653 items / 9 errors / 6 deselected / 647 selected\n\n==================================== ERRORS ====================================\n\u001b[31m\u001b[1m_________________ ERROR collecting test/test_InfoExtractor.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_InfoExtractor.py\u001b[0m:16: in <module>\n from yt_dlp.extractor import YoutubeIE, get_info_extractor\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m___________________ ERROR collecting test/test_YoutubeDL.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_YoutubeDL.py\u001b[0m:17: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m____________________ ERROR collecting test/test_all_urls.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_all_urls.py\u001b[0m:14: in <module>\n from yt_dlp.extractor import FacebookIE, YoutubeIE, gen_extractors\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m____________________ ERROR collecting test/test_download.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_download.py\u001b[0m:286: in <module>\n inject_tests(normal_test_cases)\n\u001b[1m\u001b[31mtest/test_download.py\u001b[0m:275: in inject_tests\n for test_case in test_cases:\n\u001b[1m\u001b[31mtest/helper.py\u001b[0m:91: in gettestcases\n for ie in yt_dlp.extractor.gen_extractors():\n\u001b[1m\u001b[31myt_dlp/extractor/__init__.py\u001b[0m:20: in gen_extractors\n return [klass() for klass in gen_extractor_classes()]\n\u001b[1m\u001b[31myt_dlp/extractor/__init__.py\u001b[0m:11: in gen_extractor_classes\n from .extractors import _ALL_CLASSES\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_____________ ERROR collecting test/test_iqiyi_sdk_interpreter.py ______________\u001b[0m\n\u001b[1m\u001b[31mtest/test_iqiyi_sdk_interpreter.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import IqiyiIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m___________________ ERROR collecting test/test_subtitles.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_subtitles.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import (\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_________________ ERROR collecting test/test_youtube_lists.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_lists.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import YoutubeIE, YoutubeTabIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m__________________ ERROR collecting test/test_youtube_misc.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_misc.py\u001b[0m:11: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_______________ ERROR collecting test/test_youtube_signature.py ________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_signature.py\u001b[0m:17: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[36m\u001b[1m=========================== short test summary info ============================\u001b[0m\n\u001b[31mERROR\u001b[0m test/test_InfoExtractor.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_YoutubeDL.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_all_urls.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_download.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_iqiyi_sdk_interpreter.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_subtitles.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_lists.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_misc.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_signature.py - NameError: name '_Mx3BaseIE' is not defined\n!!!!!!!!!!!!!!!!!!! Interrupted: 9 errors during collection !!!!!!!!!!!!!!!!!!!!\n\u001b[31m======================= \u001b[33m6 deselected\u001b[0m, \u001b[31m\u001b[1m9 errors\u001b[0m\u001b[31m in 1.32s\u001b[0m\u001b[31m ========================\u001b[0m\n##[error]Process completed with exit code 2.\n"}, {"step_name": "Linter/5_Make lazy extractors.txt", "log": "##[group]Run python devscripts/make_lazy_extractors.py\n\u001b[36;1mpython devscripts/make_lazy_extractors.py\u001b[0m\nshell: /usr/bin/bash -e {0}\n##[endgroup]\nTraceback (most recent call last):\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 132, in <module>\n main()\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 41, in main\n _ALL_CLASSES = get_all_ies() # Must be before import\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 68, in get_all_ies\n from yt_dlp.extractor.extractors import _ALL_CLASSES\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/extractors.py\", line 17, in <module>\n from ._extractors import * # noqa: F403\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/_extractors.py\", line 1140, in <module>\n from .mx3 import (\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 80, in <module>\n class Mx3IE(Mx3BaseIE):\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 82, in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\nNameError: name '_Mx3BaseIE' is not defined. Did you mean: 'Mx3BaseIE'?\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/yt_dlp/extractor/mx3.py b/yt_dlp/extractor/mx3.py\nindex b4fd5ab72..e39334a6e 100644\n--- a/yt_dlp/extractor/mx3.py\n+++ b/yt_dlp/extractor/mx3.py\n@@ -79,7 +79,7 @@ def get_info_field(name):\n \n class Mx3IE(Mx3BaseIE):\n _DOMAIN = 'mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://mx3.ch/t/1Cru',\n 'md5': '7ba09e9826b4447d4e1ce9d69e0e295f',\n@@ -131,7 +131,7 @@ class Mx3IE(Mx3BaseIE):\n \n class Mx3NeoIE(Mx3BaseIE):\n _DOMAIN = 'neo.mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://neo.mx3.ch/t/1hpd',\n 'md5': '6d9986bbae5cac3296ec8813bf965eb2',\n@@ -152,7 +152,7 @@ class Mx3NeoIE(Mx3BaseIE):\n \n class Mx3VolksmusikIE(Mx3BaseIE):\n _DOMAIN = 'volksmusik.mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://volksmusik.mx3.ch/t/Zx',\n 'md5': 'dd967a7b0c1ef898f3e072cf9c2eae3c',\n", "difficulty": 1, "changed_files": ["yt_dlp/extractor/mx3.py"], "commit_link": "https://github.com/yt-dlp/yt-dlp/tree/212ff27cb59b69bad24a2e675d97e52ead5b12e8"}
 
 
data/python/2201be2.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 159, "repo_owner": "mwaskom", "repo_name": "seaborn", "head_branch": "plot/layout_rect", "workflow_name": "CI", "workflow_filename": "ci.yaml", "workflow_path": ".github/workflows/ci.yaml", "contributor": "mwaskom", "sha_fail": "2201be21886bb82201f3c3487f5f1468f6e6ac81", "sha_success": "596ad4f85f31f39bcedd2fbca935647b1a2fdf84", "workflow": "name: CI\n\non:\n push:\n branches: [master, v0.*]\n pull_request:\n branches: master\n schedule:\n - cron: '0 6 * * 1,4' # Each Monday and Thursday at 06:00 UTC\n workflow_dispatch:\n\npermissions:\n contents: read\n\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: ${{ github.workspace }}/seaborn-data\n\njobs:\n build-docs:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python 3.11\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n with:\n python-version: \"3.11\"\n\n - name: Install seaborn\n run: |\n pip install --upgrade pip\n pip install .[stats,docs]\n\n - name: Install pandoc\n run: |\n sudo apt-get install pandoc\n\n - name: Cache datasets\n run: |\n git clone https://github.com/mwaskom/seaborn-data.git\n ls $SEABORN_DATA\n\n - name: Build docs\n env:\n SPHINXOPTS: -j `nproc`\n run: |\n cd doc\n make -j `nproc` notebooks\n make html\n\n\n run-tests:\n runs-on: ubuntu-latest\n\n strategy:\n matrix:\n python: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n install: [full]\n deps: [latest]\n\n include:\n - python: \"3.8\"\n install: full\n deps: pinned\n - python: \"3.11\"\n install: light\n deps: latest\n\n steps:\n - uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python ${{ matrix.python }}\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n with:\n python-version: ${{ matrix.python }}\n allow-prereleases: true\n\n - name: Install seaborn\n run: |\n pip install --upgrade pip wheel\n if [[ ${{matrix.install}} == 'full' ]]; then EXTRAS=',stats'; fi\n if [[ ${{matrix.deps }} == 'pinned' ]]; then DEPS='-r ci/deps_pinned.txt'; fi\n pip install .[dev$EXTRAS] $DEPS\n\n - name: Run tests\n run: make test\n\n - name: Upload coverage\n uses: codecov/codecov-action@eaaf4bedf32dbdc6b720b63067d99c4d77d6047d # v3.1.4\n if: ${{ success() }}\n\n lint:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n steps:\n\n - name: Checkout\n uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n\n - name: Install tools\n run: pip install mypy flake8\n\n - name: Flake8\n run: make lint\n\n - name: Type checking\n run: make typecheck\n", "logs": [{"step_name": "build-docs/7_Build docs.txt", "log": "##[group]Run cd doc\n\u001b[36;1mcd doc\u001b[0m\n\u001b[36;1mmake -j `nproc` notebooks\u001b[0m\n\u001b[36;1mmake html\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n SPHINXOPTS: -j `nproc`\n##[endgroup]\nmake[1]: Entering directory '/home/runner/work/seaborn/seaborn/doc/_tutorial'\n../tools/nb_to_doc.py aesthetics.ipynb ../tutorial\n../tools/nb_to_doc.py axis_grids.ipynb ../tutorial\nmake[1]: Entering directory '/home/runner/work/seaborn/seaborn/doc/_docstrings'\n../tools/nb_to_doc.py FacetGrid.ipynb ../docstrings\n../tools/nb_to_doc.py categorical.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py JointGrid.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py PairGrid.ipynb ../docstrings\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py color_palettes.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py axes_style.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py data_structure.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py barplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py blend_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py distributions.ipynb ../tutorial\n../tools/nb_to_doc.py boxenplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py error_bars.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py boxplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py function_overview.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py catplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py clustermap.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py color_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py introduction.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py countplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py cubehelix_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py dark_palette.ipynb ../docstrings\n../tools/nb_to_doc.py objects_interface.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py displot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py diverging_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py ecdfplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py heatmap.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py histplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py hls_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py properties.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py husl_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py jointplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py regression.ipynb ../tutorial\n../tools/nb_to_doc.py kdeplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py relational.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py light_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py lineplot.ipynb ../docstrings\n../tools/nb_to_doc.py lmplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py move_legend.ipynb ../docstrings\n../tools/nb_to_doc.py mpl_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Agg.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Area.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Band.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Bar.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\nmake[1]: Leaving directory '/home/runner/work/seaborn/seaborn/doc/_tutorial'\n../tools/nb_to_doc.py objects.Bars.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Count.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Dash.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Dodge.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Dot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Dots.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Est.ipynb ../docstrings\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Hist.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Jitter.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.KDE.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Line.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Lines.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Norm.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Path.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Paths.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Perc.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.add.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.config.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.facet.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.label.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.layout.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.limit.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Plot.on.ipynb ../docstrings\nTraceback (most recent call last):\n File \"/home/runner/work/seaborn/seaborn/doc/_docstrings/../tools/nb_to_doc.py\", line 126, in <module>\n ep.preprocess(nb, {\"metadata\": {\"path\": basedir}})\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/preprocessors/execute.py\", line 102, in preprocess\n self.preprocess_cell(cell, resources, index)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/preprocessors/execute.py\", line 123, in preprocess_cell\n cell = self.execute_cell(cell, index, store_history=True)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/jupyter_core/utils/__init__.py\", line 173, in wrapped\n return loop.run_until_complete(inner)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/asyncio/base_events.py\", line 653, in run_until_complete\n return future.result()\n ^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbclient/client.py\", line 1062, in async_execute_cell\n await self._check_raise_for_error(cell, cell_index, exec_reply)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbclient/client.py\", line 918, in _check_raise_for_error\n raise CellExecutionError.from_cell_and_msg(cell, exec_reply_content)\nnbclient.exceptions.CellExecutionError: An error occurred while executing the following cell:\n------------------\np.layout(extent=[0, 0, .8, 1]).show()\n------------------\n\n\n\u001b[0;31m---------------------------------------------------------------------------\u001b[0m\n\u001b[0;31mNameError\u001b[0m Traceback (most recent call last)\nCell \u001b[0;32mIn[5], line 1\u001b[0m\n\u001b[0;32m----> 1\u001b[0m \u001b[43mp\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mlayout\u001b[49m\u001b[43m(\u001b[49m\u001b[43mextent\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m[\u001b[49m\u001b[38;5;241;43m0\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m0\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m.8\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m1\u001b[39;49m\u001b[43m]\u001b[49m\u001b[43m)\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mshow\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:930\u001b[0m, in \u001b[0;36mPlot.show\u001b[0;34m(self, **kwargs)\u001b[0m\n\u001b[1;32m 913\u001b[0m \u001b[38;5;250m\u001b[39m\u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 914\u001b[0m \u001b[38;5;124;03mCompile the plot and display it by hooking into pyplot.\u001b[39;00m\n\u001b[1;32m 915\u001b[0m \n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 922\u001b[0m \n\u001b[1;32m 923\u001b[0m \u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 924\u001b[0m \u001b[38;5;66;03m# TODO make pyplot configurable at the class level, and when not using,\u001b[39;00m\n\u001b[1;32m 925\u001b[0m \u001b[38;5;66;03m# import IPython.display and call on self to populate cell output?\u001b[39;00m\n\u001b[1;32m 926\u001b[0m \n\u001b[1;32m 927\u001b[0m \u001b[38;5;66;03m# Keep an eye on whether matplotlib implements \"attaching\" an existing\u001b[39;00m\n\u001b[1;32m 928\u001b[0m \u001b[38;5;66;03m# figure to pyplot: https://github.com/matplotlib/matplotlib/pull/14024\u001b[39;00m\n\u001b[0;32m--> 930\u001b[0m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mplot\u001b[49m\u001b[43m(\u001b[49m\u001b[43mpyplot\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m)\u001b[49m\u001b[38;5;241m.\u001b[39mshow(\u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:937\u001b[0m, in \u001b[0;36mPlot.plot\u001b[0;34m(self, pyplot)\u001b[0m\n\u001b[1;32m 933\u001b[0m \u001b[38;5;250m\u001b[39m\u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 934\u001b[0m \u001b[38;5;124;03mCompile the plot spec and return the Plotter object.\u001b[39;00m\n\u001b[1;32m 935\u001b[0m \u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 936\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m theme_context(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_theme_with_defaults()):\n\u001b[0;32m--> 937\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_plot\u001b[49m\u001b[43m(\u001b[49m\u001b[43mpyplot\u001b[49m\u001b[43m)\u001b[49m\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:971\u001b[0m, in \u001b[0;36mPlot._plot\u001b[0;34m(self, pyplot)\u001b[0m\n\u001b[1;32m 969\u001b[0m \u001b[38;5;66;03m# Add various figure decorations\u001b[39;00m\n\u001b[1;32m 970\u001b[0m plotter\u001b[38;5;241m.\u001b[39m_make_legend(\u001b[38;5;28mself\u001b[39m)\n\u001b[0;32m--> 971\u001b[0m \u001b[43mplotter\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_finalize_figure\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m)\u001b[49m\n\u001b[1;32m 973\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m plotter\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:1815\u001b[0m, in \u001b[0;36mPlotter._finalize_figure\u001b[0;34m(self, p)\u001b[0m\n\u001b[1;32m 1812\u001b[0m set_layout_engine(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure, \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtight\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 1814\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m (extent \u001b[38;5;241m:=\u001b[39m p\u001b[38;5;241m.\u001b[39m_layout_spec\u001b[38;5;241m.\u001b[39mget(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mextent\u001b[39m\u001b[38;5;124m\"\u001b[39m)) \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m-> 1815\u001b[0m engine \u001b[38;5;241m=\u001b[39m \u001b[43mget_layout_engine\u001b[49m(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure)\n\u001b[1;32m 1816\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m engine \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[1;32m 1817\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure\u001b[38;5;241m.\u001b[39msubplots_adjust(\u001b[38;5;241m*\u001b[39mextent)\n\n\u001b[0;31mNameError\u001b[0m: name 'get_layout_engine' is not defined\n\nmake[1]: *** [Makefile:7: ../docstrings/objects.Plot.layout.rst] Error 1\nmake[1]: *** Waiting for unfinished jobs....\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\nmake[1]: Leaving directory '/home/runner/work/seaborn/seaborn/doc/_docstrings'\nmake: *** [Makefile:60: docstrings] Error 2\n##[error]Process completed with exit code 2.\n"}, {"step_name": "lint/5_Flake8.txt", "log": "##[group]Run make lint\n\u001b[36;1mmake lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n##[endgroup]\nflake8 seaborn/ tests/\nseaborn/_core/plot.py:1815:22: F821 undefined name 'get_layout_engine'\nmake: *** [Makefile:7: lint] Error 1\n##[error]Process completed with exit code 2.\n"}, {"step_name": "run-tests (3.11, light, latest)/5_Run tests.txt", "log": "##[group]Run make test\n\u001b[36;1mmake test\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\npytest -n auto --cov=seaborn --cov=tests --cov-config=setup.cfg tests\n============================= test session starts ==============================\nplatform linux -- Python 3.11.7, pytest-7.4.3, pluggy-1.3.0\nrootdir: /home/runner/work/seaborn/seaborn\nconfigfile: pyproject.toml\nplugins: cov-4.1.0, xdist-3.5.0\ncreated: 4/4 workers\n4 workers [2379 items]\n\n........................................................................ [ 3%]\n........................................................................ [ 6%]\n........................................................................ [ 9%]\n........................................................................ [ 12%]\n........................................................................ [ 15%]\n........................................................................ [ 18%]\n........................................................................ [ 21%]\n........................................................................ [ 24%]\n........................................................................ [ 27%]\n........................................................................ [ 30%]\n........................................................................ [ 33%]\n.s...................................................................... [ 36%]\n........................................................................ [ 39%]\n........................................................................ [ 42%]\n.................................................s.s................ss.. [ 45%]\n..ss.............ssss................................................sss [ 48%]\nsssssssssssssssss.sssssssssssssssssss.ssssssssssssss.................... [ 51%]\n...............................................s........................ [ 54%]\n........................................................................ [ 57%]\n........................................................................ [ 60%]\n..............................................................s......... [ 63%]\n..............................................................x......... [ 66%]\n......x.........x.........................................x..x.......... [ 69%]\n.............................................s.....s.................... [ 72%]\n..................s...................................................... [ 75%]\n........................................................................ [ 78%]\n........................................................................ [ 81%]\n........................................................................ [ 84%]\n..................F.F..x................................................ [ 87%]\n...........................F............................................ [ 90%]\n........................................................................ [ 93%]\n........................................................................ [ 96%]\n........................s..s............................................ [ 99%]\n.. [100%]\n=================================== FAILURES ===================================\n_______________________ TestPlotting.test_layout_extent ________________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f06656abb10>\n\n @pytest.mark.skipif(\n _version_predates(mpl, \"3.6\"),\n reason=\"mpl<3.6 does not have get_layout_engine\",\n )\n def test_layout_extent(self):\n \n> p = Plot().layout(extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1100: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0654edebd0>\np = <seaborn._core.plot.Plot object at 0x7f06567fe150>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n_________________ TestPlotting.test_constrained_layout_extent __________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f0665702410>\n\n @pytest.mark.skipif(\n _version_predates(mpl, \"3.6\"),\n reason=\"mpl<3.6 does not have get_layout_engine\",\n )\n def test_constrained_layout_extent(self):\n \n> p = Plot().layout(engine=\"constrained\", extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1109: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0656f58050>\np = <seaborn._core.plot.Plot object at 0x7f0656f597d0>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n_____________________ TestPlotting.test_base_layout_extent _____________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f0664b34450>\n\n def test_base_layout_extent(self):\n \n> p = Plot().layout(engine=None, extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1114: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0656d300d0>\np = <seaborn._core.plot.Plot object at 0x7f0656d301d0>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n\n---------- coverage: platform linux, python 3.11.7-final-0 -----------\nName Stmts Miss Cover\n------------------------------------------------------\nseaborn/__init__.py 16 0 100%\nseaborn/_base.py 766 23 97%\nseaborn/_compat.py 77 47 39%\nseaborn/_core/__init__.py 0 0 100%\nseaborn/_core/data.py 116 5 96%\nseaborn/_core/exceptions.py 10 1 90%\nseaborn/_core/groupby.py 54 0 100%\nseaborn/_core/moves.py 109 0 100%\nseaborn/_core/plot.py 828 13 98%\nseaborn/_core/properties.py 425 4 99%\nseaborn/_core/rules.py 63 1 98%\nseaborn/_core/scales.py 502 49 90%\nseaborn/_core/subplots.py 140 0 100%\nseaborn/_core/typing.py 25 1 96%\nseaborn/_docstrings.py 40 5 88%\nseaborn/_marks/__init__.py 0 0 100%\nseaborn/_marks/area.py 86 3 97%\nseaborn/_marks/bar.py 123 0 100%\nseaborn/_marks/base.py 132 2 98%\nseaborn/_marks/dot.py 92 0 100%\nseaborn/_marks/line.py 116 0 100%\nseaborn/_marks/text.py 33 0 100%\nseaborn/_statistics.py 323 10 97%\nseaborn/_stats/__init__.py 0 0 100%\nseaborn/_stats/aggregation.py 42 1 98%\nseaborn/_stats/base.py 26 1 96%\nseaborn/_stats/counting.py 96 0 100%\nseaborn/_stats/density.py 99 3 97%\nseaborn/_stats/order.py 30 3 90%\nseaborn/_stats/regression.py 23 1 96%\nseaborn/_testing.py 52 4 92%\nseaborn/algorithms.py 60 0 100%\nseaborn/axisgrid.py 1039 32 97%\nseaborn/categorical.py 1204 13 99%\nseaborn/distributions.py 951 36 96%\nseaborn/matrix.py 551 317 42%\nseaborn/miscplot.py 24 0 100%\nseaborn/objects.py 15 0 100%\nseaborn/palettes.py 249 1 99%\nseaborn/rcmod.py 104 0 100%\nseaborn/regression.py 336 46 86%\nseaborn/relational.py 285 1 99%\nseaborn/utils.py 395 23 94%\ntests/__init__.py 0 0 100%\ntests/_core/__init__.py 0 0 100%\ntests/_core/test_data.py 293 2 99%\ntests/_core/test_groupby.py 83 0 100%\ntests/_core/test_moves.py 231 0 100%\ntests/_core/test_plot.py 1550 24 98%\ntests/_core/test_properties.py 365 4 99%\ntests/_core/test_rules.py 72 0 100%\ntests/_core/test_scales.py 547 1 99%\ntests/_core/test_subplots.py 368 0 100%\ntests/_marks/__init__.py 0 0 100%\ntests/_marks/test_area.py 84 0 100%\ntests/_marks/test_bar.py 152 0 100%\ntests/_marks/test_base.py 102 0 100%\ntests/_marks/test_dot.py 136 0 100%\ntests/_marks/test_line.py 298 0 100%\ntests/_marks/test_text.py 98 1 99%\ntests/_stats/__init__.py 0 0 100%\ntests/_stats/test_aggregation.py 84 0 100%\ntests/_stats/test_counting.py 180 0 100%\ntests/_stats/test_density.py 141 8 94%\ntests/_stats/test_order.py 64 1 98%\ntests/_stats/test_regression.py 36 0 100%\ntests/conftest.py 107 0 100%\ntests/test_algorithms.py 110 0 100%\ntests/test_axisgrid.py 1314 6 99%\ntests/test_base.py 1008 3 99%\ntests/test_categorical.py 2161 17 99%\ntests/test_distributions.py 1523 14 99%\ntests/test_docstrings.py 19 0 100%\ntests/test_matrix.py 864 454 47%\ntests/test_miscplot.py 24 0 100%\ntests/test_objects.py 11 0 100%\ntests/test_palettes.py 304 0 100%\ntests/test_rcmod.py 189 21 89%\ntests/test_regression.py 441 46 90%\ntests/test_relational.py 1213 8 99%\ntests/test_statistics.py 489 19 96%\ntests/test_utils.py 382 4 99%\n------------------------------------------------------\nTOTAL 24700 1279 95%\n\n=========================== short test summary info ============================\nFAILED tests/_core/test_plot.py::TestPlotting::test_layout_extent - NameError: name 'get_layout_engine' is not defined\nFAILED tests/_core/test_plot.py::TestPlotting::test_constrained_layout_extent - NameError: name 'get_layout_engine' is not defined\nFAILED tests/_core/test_plot.py::TestPlotting::test_base_layout_extent - NameError: name 'get_layout_engine' is not defined\n====== 3 failed, 2299 passed, 71 skipped, 6 xfailed in 244.97s (0:04:04) =======\nmake: *** [Makefile:4: test] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/seaborn/_core/plot.py b/seaborn/_core/plot.py\nindex af4669c6..39ccd2e0 100644\n--- a/seaborn/_core/plot.py\n+++ b/seaborn/_core/plot.py\n@@ -40,7 +40,7 @@ from seaborn._core.typing import (\n )\n from seaborn._core.exceptions import PlotSpecError\n from seaborn._core.rules import categorical_order\n-from seaborn._compat import set_layout_engine\n+from seaborn._compat import get_layout_engine, set_layout_engine\n from seaborn.rcmod import axes_style, plotting_context\n from seaborn.palettes import color_palette\n \n", "difficulty": 2, "changed_files": ["seaborn/_core/plot.py"], "commit_link": "https://github.com/mwaskom/seaborn/tree/2201be21886bb82201f3c3487f5f1468f6e6ac81"}
 
 
data/python/2327ca6.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/2a104bf.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 0, "repo_owner": "pyg-team", "repo_name": "pytorch_geometric", "head_branch": "master", "workflow_name": "Testing PyTorch nightly", "workflow_filename": "latest_testing.yml", "workflow_path": ".github/workflows/latest_testing.yml", "contributor": "pmpalang", "sha_fail": "2a104bfbc6cad12e95d941244955382e96f21c51", "sha_success": "060fc6a2edb9775331645512c1e6eb5ed4f0afbd", "workflow": "name: Testing PyTorch nightly\n\non: # yamllint disable-line rule:truthy\n push:\n branches:\n - master\n pull_request:\n\njobs:\n\n latest_pytest:\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 40\n\n # Skip workflow if only certain files have been changed.\n - name: Get changed files\n id: changed-files-specific\n uses: tj-actions/changed-files@v34\n with:\n files: |\n benchmark/**\n conda/**\n docker/**\n docs/**\n examples/**\n graphgym/**\n README.md\n CHANGELOG.md\n\n - name: Setup packages\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n uses: ./.github/actions/setup\n with:\n torch-version: nightly\n\n - name: Install main package\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n run: |\n pip install -e .[test]\n\n - name: Run tests\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n timeout-minutes: 10\n run: |\n pytest\n", "logs": [{"step_name": "latest_pytest/6_Run tests.txt", "log": "##[group]Run pytest\n\u001b[36;1mpytest\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n\u001b[1m============================= test session starts ==============================\u001b[0m\nplatform linux -- Python 3.8.18, pytest-7.4.3, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/pytorch_geometric/pytorch_geometric\nconfigfile: pyproject.toml\nplugins: cov-4.1.0\n\u001b[1mcollecting ... \u001b[0mcollected 5718 items / 1 error\n\n==================================== ERRORS ====================================\n\u001b[31m\u001b[1m_____________ ERROR collecting test/loader/test_neighbor_loader.py _____________\u001b[0m\n\u001b[31mImportError while importing test module '/home/runner/work/pytorch_geometric/pytorch_geometric/test/loader/test_neighbor_loader.py'.\nHint: make sure your test modules/packages have valid Python names.\nTraceback:\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/python.py:617: in _importtestmodule\n mod = import_path(self.path, mode=importmode, root=self.config.rootpath)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/pathlib.py:567: in import_path\n importlib.import_module(module_name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n<frozen importlib._bootstrap>:1014: in _gcd_import\n ???\n<frozen importlib._bootstrap>:991: in _find_and_load\n ???\n<frozen importlib._bootstrap>:975: in _find_and_load_unlocked\n ???\n<frozen importlib._bootstrap>:671: in _load_unlocked\n ???\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/assertion/rewrite.py:186: in exec_module\n exec(co, module.__dict__)\ntest/loader/test_neighbor_loader.py:24: in <module>\n from torch_geometric.typing import (\nE ImportError: cannot import name 'WITH_EDGE_TIME_NEIGHBOR_SAMPLE' from 'torch_geometric.typing' (/home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/typing.py)\u001b[0m\n\u001b[33m=============================== warnings summary ===============================\u001b[0m\ntorch_geometric/graphgym/config.py:19\n /home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/graphgym/config.py:19: UserWarning: Could not define global config object. Please install 'yacs' via 'pip install yacs' in order to use GraphGym\n warnings.warn(\"Could not define global config object. Please install \"\n\ntorch_geometric/graphgym/imports.py:14\n /home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/graphgym/imports.py:14: UserWarning: Please install 'pytorch_lightning' via 'pip install pytorch_lightning' in order to use GraphGym\n warnings.warn(\"Please install 'pytorch_lightning' via \"\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\u001b[36m\u001b[1m=========================== short test summary info ============================\u001b[0m\n\u001b[31mERROR\u001b[0m test/loader/test_neighbor_loader.py\n!!!!!!!!!!!!!!!!!!!! Interrupted: 1 error during collection !!!!!!!!!!!!!!!!!!!!\n\u001b[31m========================= \u001b[33m2 warnings\u001b[0m, \u001b[31m\u001b[1m1 error\u001b[0m\u001b[31m in 7.62s\u001b[0m\u001b[31m =========================\u001b[0m\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/torch_geometric/typing.py b/torch_geometric/typing.py\nindex d9e687601..085ecc613 100644\n--- a/torch_geometric/typing.py\n+++ b/torch_geometric/typing.py\n@@ -57,6 +57,7 @@ except Exception as e:\n WITH_SAMPLED_OP = False\n WITH_INDEX_SORT = False\n WITH_METIS = False\n+ WITH_EDGE_TIME_NEIGHBOR_SAMPLE = False\n WITH_WEIGHTED_NEIGHBOR_SAMPLE = False\n \n try:\n", "difficulty": 1, "changed_files": ["torch_geometric/typing.py"], "commit_link": "https://github.com/pyg-team/pytorch_geometric/tree/2a104bfbc6cad12e95d941244955382e96f21c51"}
 
 
data/python/2a59b55.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 84, "repo_owner": "django-import-export", "repo_name": "django-import-export", "head_branch": "remove_count_queries", "workflow_name": ".github/workflows/pre-commit.yml", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "PetrDlouhy", "sha_fail": "2a59b55e6124b33dca7f48c12845c78130b20fd5", "sha_success": "36786764b3ef1c4c3c8d0db451cea9e1b325616c", "workflow": "on:\n pull_request:\n push:\n branches:\n - main\n\njobs:\n main:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: 3.x\n - uses: pre-commit/[email protected]\n - uses: pre-commit-ci/[email protected]\n if: always()\n", "logs": [{"step_name": "main/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 6.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 26.4 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 31.5 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 92.4 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 78.2 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 93.4 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nCache Size: ~15 MB (15847689 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/83ce4811-68cf-4484-bf4e-4cfbe36bb31a/cache.tzst -P -C /home/runner/work/django-import-export/django-import-export --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: flake8\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\n\u001b[1mimport_export/admin.py\u001b[m\u001b[36m:\u001b[m749\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (96 > 88 characters)\n\u001b[1mimport_export/admin.py\u001b[m\u001b[36m:\u001b[m756\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (98 > 88 characters)\n\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/import_export/admin.py b/import_export/admin.py\nindex a1b288e9..8bb2db67 100644\n--- a/import_export/admin.py\n+++ b/import_export/admin.py\n@@ -746,14 +746,16 @@ class ExportMixin(BaseExportMixin, ImportExportMixinBase):\n def get_results(self, request):\n \"\"\"\n We override this method because we only call ChangeList.get_queryset()\n- so we don't need anything from this method. The get_results() gets called during\n- ChangeList.__init__() and we do want to avoid unnecessary COUNT queries.\n+ so we don't need anything from this method.\n+ The get_results() gets called during ChangeList.__init__()\n+ and we do want to avoid unnecessary COUNT queries.\n \"\"\"\n pass\n \n cl = ExportChangeList(**changelist_kwargs)\n \n- # get_queryset() is already called during initialization, it is enough to get it's results\n+ # get_queryset() is already called during initialization,\n+ # it is enough to get it's results\n if hasattr(cl, \"queryset\"):\n return cl.queryset\n \n", "difficulty": 0, "changed_files": ["import_export/admin.py"], "commit_link": "https://github.com/django-import-export/django-import-export/tree/2a59b55e6124b33dca7f48c12845c78130b20fd5"}
 
 
data/python/2ab9e84.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 171, "repo_owner": "qtile", "repo_name": "qtile", "head_branch": "lists-match", "workflow_name": "Run pre-commit", "workflow_filename": "pre_commit.yml", "workflow_path": ".github/workflows/pre_commit.yml", "contributor": "jwijenbergh", "sha_fail": "2ab9e843db39063cc9bc33b924cef535eb289894", "sha_success": "26c042804d308f4c1c5b5061ee10da1515129489", "workflow": "name: Run pre-commit\n\non:\n push:\n pull_request:\n\njobs:\n check:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v4\n - name: Install dependencies\n run: |\n sudo apt update\n sudo apt install --no-install-recommends libxkbcommon-dev\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "check/5_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nDefaulting to user installation because normal site-packages is not writeable\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 KB 6.3 MB/s eta 0:00:00\nCollecting nodeenv>=0.11.1\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nRequirement already satisfied: pyyaml>=5.1 in /usr/lib/python3/dist-packages (from pre-commit) (5.4.1)\nCollecting cfgv>=2.0.0\n Downloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nCollecting virtualenv>=20.10.0\n Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 24.4 MB/s eta 0:00:00\nCollecting identify>=1.0.0\n Downloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 KB 37.4 MB/s eta 0:00:00\nRequirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from nodeenv>=0.11.1->pre-commit) (59.6.0)\nCollecting filelock<4,>=3.12.2\n Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\nRequirement already satisfied: platformdirs<5,>=3.9.1 in /usr/local/lib/python3.10/dist-packages (from virtualenv>=20.10.0->pre-commit) (4.1.0)\nCollecting distlib<1,>=0.3.7\n Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 43.2 MB/s eta 0:00:00\nInstalling collected packages: distlib, nodeenv, identify, filelock, cfgv, virtualenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 pre-commit-3.6.0 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nargcomplete==3.2.1\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\npackaging==23.2\npipx==1.3.3\nplatformdirs==4.1.0\npre-commit==3.6.0\ntomli==2.0.1\nuserpath==1.9.1\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3||05c1fbbb63b353467651b511e6ac241ffd2d8e71749cfabfe5ee6bc8366d2d02\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\n##[endgroup]\nCache Size: ~36 MB (37431926 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/8e85cf73-4d99-4c40-a6e8-25c81b58e73c/cache.tzst -P -C /home/runner/work/qtile/qtile --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3||05c1fbbb63b353467651b511e6ac241ffd2d8e71749cfabfe5ee6bc8366d2d02\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[42mPassed\u001b[m\nmypy.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: mypy\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\nlibqtile/config.py:886: error: Statement is unreachable [unreachable]\nlibqtile/config.py:890: error: Statement is unreachable [unreachable]\nlibqtile/config.py:894: error: Statement is unreachable [unreachable]\nlibqtile/config.py:911: error: Statement is unreachable [unreachable]\nlibqtile/config.py:915: error: Statement is unreachable [unreachable]\nFound 5 errors in 1 file (checked 179 source files)\n\nvulture..................................................................\u001b[42mPassed\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/libqtile/config.py b/libqtile/config.py\nindex 9801b3ba..6a6163e2 100644\n--- a/libqtile/config.py\n+++ b/libqtile/config.py\n@@ -883,15 +883,15 @@ class Match:\n \n if title is not None:\n if isinstance(title, list): # type: ignore\n- title = convert_deprecated_list(title, \"title\")\n+ title = convert_deprecated_list(title, \"title\") # type: ignore\n self._rules[\"title\"] = title\n if wm_class is not None:\n if isinstance(wm_class, list): # type: ignore\n- wm_class = convert_deprecated_list(wm_class, \"wm_class\")\n+ wm_class = convert_deprecated_list(wm_class, \"wm_class\") # type: ignore\n self._rules[\"wm_class\"] = wm_class\n if wm_instance_class is not None:\n if isinstance(wm_instance_class, list): # type: ignore\n- wm_instance_class = convert_deprecated_list(\n+ wm_instance_class = convert_deprecated_list( # type: ignore\n wm_instance_class, \"wm_instance_class\"\n )\n self._rules[\"wm_instance_class\"] = wm_instance_class\n@@ -908,11 +908,11 @@ class Match:\n \n if role is not None:\n if isinstance(role, list): # type: ignore\n- role = convert_deprecated_list(role, \"role\")\n+ role = convert_deprecated_list(role, \"role\") # type: ignore\n self._rules[\"role\"] = role\n if wm_type is not None:\n if isinstance(wm_type, list): # type: ignore\n- wm_type = convert_deprecated_list(wm_type, \"wm_type\")\n+ wm_type = convert_deprecated_list(wm_type, \"wm_type\") # type: ignore\n self._rules[\"wm_type\"] = wm_type\n \n @staticmethod\n", "difficulty": 1, "changed_files": ["libqtile/config.py"], "commit_link": "https://github.com/qtile/qtile/tree/2ab9e843db39063cc9bc33b924cef535eb289894"}
 
 
data/python/2c06ffa.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 4, "repo_owner": "huggingface", "repo_name": "diffusers", "head_branch": "ipadapterfaceid", "workflow_name": "Run code quality checks", "workflow_filename": "pr_quality.yml", "workflow_path": ".github/workflows/pr_quality.yml", "contributor": "fabiorigano", "sha_fail": "2c06ffa4c9d2c37846c60ad75899b4d72f214ff9", "sha_success": "217d9d073981605acab5200fc841f20c798c1449", "workflow": "name: Run code quality checks\n\non:\n pull_request:\n branches:\n - main\n push:\n branches:\n - main\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }}\n cancel-in-progress: true\n\njobs:\n check_code_quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python\n uses: actions/setup-python@v4\n with:\n python-version: \"3.8\"\n - name: Install dependencies\n run: |\n python -m pip install --upgrade pip\n pip install .[quality]\n - name: Check quality\n run: |\n ruff check examples tests src utils scripts\n ruff format examples tests src utils scripts --check\n\n check_repository_consistency:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python\n uses: actions/setup-python@v4\n with:\n python-version: \"3.8\"\n - name: Install dependencies\n run: |\n python -m pip install --upgrade pip\n pip install .[quality]\n - name: Check quality\n run: |\n python utils/check_copies.py\n python utils/check_dummies.py\n make deps_table_check_updated\n", "logs": [{"step_name": "check_code_quality/5_Check quality.txt", "log": "##[group]Run ruff check examples tests src utils scripts\n\u001b[36;1mruff check examples tests src utils scripts\u001b[0m\n\u001b[36;1mruff format examples tests src utils scripts --check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/actions-runner/_work/_tool/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/actions-runner/_work/_tool/Python/3.8.18/x64/lib\n##[endgroup]\nexamples/community/ip_adapter_face_id.py:15:1: I001 [*] Import block is un-sorted or un-formatted\nFound 1 error.\n[*] 1 fixable with the `--fix` option.\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/examples/community/README.md b/examples/community/README.md\nindex f205f3b70..2fdbdb414 100755\n--- a/examples/community/README.md\n+++ b/examples/community/README.md\n@@ -3307,7 +3307,7 @@ pipeline = DiffusionPipeline.from_pretrained(\n torch_dtype=torch.float16,\n scheduler=noise_scheduler,\n vae=vae,\n- custom_pipeline=\"./forked/diffusers/examples/community/ip_adapter_face_id.py\"\n+ custom_pipeline=\"ip_adapter_face_id\"\n )\n pipeline.load_ip_adapter_face_id(\"h94/IP-Adapter-FaceID\", \"ip-adapter-faceid_sd15.bin\")\n pipeline.to(\"cuda\")\ndiff --git a/examples/community/ip_adapter_face_id.py b/examples/community/ip_adapter_face_id.py\nindex e3c5a2c84..d9325742c 100644\n--- a/examples/community/ip_adapter_face_id.py\n+++ b/examples/community/ip_adapter_face_id.py\n@@ -14,12 +14,12 @@\n \n import inspect\n from typing import Any, Callable, Dict, List, Optional, Union\n-from safetensors import safe_open\n \n import torch\n import torch.nn as nn\n import torch.nn.functional as F\n from packaging import version\n+from safetensors import safe_open\n from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection\n \n from diffusers.configuration_utils import FrozenDict\n@@ -27,20 +27,20 @@ from diffusers.image_processor import VaeImageProcessor\n from diffusers.loaders import FromSingleFileMixin, IPAdapterMixin, LoraLoaderMixin, TextualInversionLoaderMixin\n from diffusers.models import AutoencoderKL, UNet2DConditionModel\n from diffusers.models.attention_processor import FusedAttnProcessor2_0\n-from diffusers.models.lora import adjust_lora_scale_text_encoder, LoRALinearLayer\n+from diffusers.models.lora import LoRALinearLayer, adjust_lora_scale_text_encoder\n+from diffusers.pipelines.pipeline_utils import DiffusionPipeline\n+from diffusers.pipelines.stable_diffusion.pipeline_output import StableDiffusionPipelineOutput\n+from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker\n from diffusers.schedulers import KarrasDiffusionSchedulers\n from diffusers.utils import (\n- _get_model_file,\n USE_PEFT_BACKEND,\n+ _get_model_file,\n deprecate,\n logging,\n scale_lora_layers,\n unscale_lora_layers,\n )\n from diffusers.utils.torch_utils import randn_tensor\n-from diffusers.pipelines.pipeline_utils import DiffusionPipeline\n-from diffusers.pipelines.stable_diffusion.pipeline_output import StableDiffusionPipelineOutput\n-from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker\n \n \n logger = logging.get_logger(__name__) # pylint: disable=invalid-name\n@@ -555,7 +555,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(\n revision=revision,\n subfolder=subfolder,\n user_agent=user_agent,\n- )\n+ )\n if weight_name.endswith(\".safetensors\"):\n state_dict = {\"image_proj\": {}, \"ip_adapter\": {}}\n with safe_open(model_file, framework=\"pt\", device=\"cpu\") as f:\n@@ -1438,7 +1438,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(\n extra_step_kwargs = self.prepare_extra_step_kwargs(generator, eta)\n \n # 6.1 Add image embeds for IP-Adapter\n- added_cond_kwargs ={\"image_embeds\": image_embeds} if image_embeds is not None else None\n+ added_cond_kwargs = {\"image_embeds\": image_embeds} if image_embeds is not None else None\n \n # 6.2 Optionally get Guidance Scale Embedding\n timestep_cond = None\n", "difficulty": 0, "changed_files": ["examples/community/README.md", "examples/community/ip_adapter_face_id.py"], "commit_link": "https://github.com/huggingface/diffusers/tree/2c06ffa4c9d2c37846c60ad75899b4d72f214ff9"}
 
 
data/python/2ccb7d3.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 114, "repo_owner": "mikf", "repo_name": "gallery-dl", "head_branch": "steamgriddb", "workflow_name": "tests", "workflow_filename": "tests.yml", "workflow_path": ".github/workflows/tests.yml", "contributor": "the-blank-x", "sha_fail": "2ccb7d3bd3f071c6923ca6eb9baedd196665d769", "sha_success": "100966b122cd90ca139593cf8ff21fd0f777243a", "workflow": "name: tests\n\non:\n push:\n branches:\n - master\n pull_request:\n branches:\n - master\n\njobs:\n build:\n\n runs-on: ubuntu-20.04\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.5\", \"3.6\", \"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\", \"pypy3.9\"]\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Check file permissions\n run: |\n if [[ \"$(find ./gallery_dl -type f -not -perm 644)\" ]]; then exit 1; fi\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Install dependencies\n run: |\n pip install -r requirements.txt\n pip install flake8 youtube-dl\n\n - name: Install yt-dlp\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5)\n # don't install yt-dlp\n ;;\n 3.6|3.7)\n # install from PyPI\n pip install yt-dlp\n ;;\n *)\n # install from master\n pip install https://github.com/yt-dlp/yt-dlp/archive/refs/heads/master.tar.gz\n ;;\n esac\n\n - name: Lint with flake8\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5|3.6|3.7)\n flake8 --extend-exclude scripts/export_tests.py .\n ;;\n *)\n flake8 .\n ;;\n esac\n\n - name: Run tests\n run: |\n make test\n\n - name: Test autogeneration of man pages, bash/zsh/fish completion, etc\n run: |\n make\n", "logs": [{"step_name": "build (3.5)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.5\" in\n\u001b[36;1mcase \"3.5\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.5.10/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.6)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.6\" in\n\u001b[36;1mcase \"3.6\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.6.15/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.7)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.7\" in\n\u001b[36;1mcase \"3.7\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.8)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.8\" in\n\u001b[36;1mcase \"3.8\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.9\" in\n\u001b[36;1mcase \"3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.10)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.10\" in\n\u001b[36;1mcase \"3.10\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.11)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.11\" in\n\u001b[36;1mcase \"3.11\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.12)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.12\" in\n\u001b[36;1mcase \"3.12\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (pypy3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"pypy3.9\" in\n\u001b[36;1mcase \"pypy3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/PyPy/3.9.18/x64/bin/lib/pkgconfig\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gallery_dl/extractor/steamgriddb.py b/gallery_dl/extractor/steamgriddb.py\nindex 516c422b..1f803ffd 100644\n--- a/gallery_dl/extractor/steamgriddb.py\n+++ b/gallery_dl/extractor/steamgriddb.py\n@@ -173,7 +173,8 @@ class SteamgriddbGridsExtractor(SteamgriddbAssetsExtractor):\n example = \"https://www.steamgriddb.com/game/1234/grids\"\n valid_dimensions = (\"460x215\", \"920x430\", \"600x900\", \"342x482\", \"660x930\",\n \"512x512\", \"1024x1024\")\n- valid_styles = (\"alternate\", \"blurred\", \"no_logo\", \"material\", \"white_logo\")\n+ valid_styles = (\"alternate\", \"blurred\", \"no_logo\", \"material\",\n+ \"white_logo\")\n valid_file_types = (\"png\", \"jpeg\", \"jpg\", \"webp\")\n \n \n", "difficulty": 0, "changed_files": ["gallery_dl/extractor/steamgriddb.py"], "commit_link": "https://github.com/mikf/gallery-dl/tree/2ccb7d3bd3f071c6923ca6eb9baedd196665d769"}
 
 
data/python/2e41e78.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/2f0605c.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 85, "repo_owner": "django-import-export", "repo_name": "django-import-export", "head_branch": "fix-declared-model-fields-altering-export-order", "workflow_name": ".github/workflows/pre-commit.yml", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "cocorocho", "sha_fail": "2f0605c9ec79b7a675728cb525ad55b36ade2e93", "sha_success": "d1e35b26f1cfc1bd08b03ac606fa8bcd26dc91cb", "workflow": "on:\n pull_request:\n push:\n branches:\n - main\n\njobs:\n main:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: 3.x\n - uses: pre-commit/[email protected]\n - uses: pre-commit-ci/[email protected]\n if: always()\n", "logs": [{"step_name": "main/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.5.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.32-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.24.7-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.0.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.5.0-py2.py3-none-any.whl (203 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 203.7/203.7 kB 15.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.32-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 32.6 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 49.7 MB/s eta 0:00:00\nDownloading virtualenv-20.24.7-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 124.4 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 84.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.0.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 99.2 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.7 filelock-3.13.1 identify-2.5.32 nodeenv-1.8.0 platformdirs-4.0.0 pre-commit-3.5.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.24.7\n\n[notice] A new release of pip is available: 23.2.1 -> 23.3.1\n[notice] To update, run: pip install --upgrade pip\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.7\nfilelock==3.13.1\nidentify==2.5.32\nnodeenv==1.8.0\nplatformdirs==4.0.0\npre-commit==3.5.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.24.7\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.0/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nCache Size: ~15 MB (15834028 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/98029e33-8cfe-40ab-80a4-dbe66e8ec82f/cache.tzst -P -C /home/runner/work/django-import-export/django-import-export --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.0/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: flake8\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\n\u001b[1mimport_export/resources.py\u001b[m\u001b[36m:\u001b[m1360\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (101 > 88 characters)\n\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/import_export/resources.py b/import_export/resources.py\nindex 7d7ad475..99c3fe91 100644\n--- a/import_export/resources.py\n+++ b/import_export/resources.py\n@@ -1357,7 +1357,8 @@ class ModelDeclarativeMetaclass(DeclarativeMetaclass):\n continue\n \n if f.name in declared_fields:\n- # If model field is declared in `ModelResource`, remove it from `declared_fields`\n+ # If model field is declared in `ModelResource`,\n+ # remove it from `declared_fields`\n # to keep exact order of model fields\n field = declared_fields.pop(f.name)\n else:\n", "difficulty": 0, "changed_files": ["import_export/resources.py"], "commit_link": "https://github.com/django-import-export/django-import-export/tree/2f0605c9ec79b7a675728cb525ad55b36ade2e93"}
 
 
data/python/3054659.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/3137ef6.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 69, "repo_owner": "optuna", "repo_name": "optuna", "head_branch": "unifiy-implementation-of-fast-nondominated-sort", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "Alnusjaponica", "sha_fail": "3137ef65975fc93e9e82b130e545028223cef408", "sha_success": "a4cd1cd2d275d5163bcbbb4cb29b3cf378789e99", "workflow": "name: Checks\n\non:\n push:\n branches:\n - master\n pull_request: {}\n schedule:\n - cron: '0 23 * * SUN-THU'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.ref == 'refs/heads/master' && github.run_number || github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n if: (github.event_name == 'schedule' && github.repository == 'optuna/optuna') || (github.event_name != 'schedule')\n runs-on: ubuntu-latest\n\n steps:\n - uses: actions/checkout@v3\n\n - uses: actions/setup-python@v4\n with:\n python-version: 3.8\n\n - name: Install\n run: |\n python -m pip install -U pip\n pip install --progress-bar off -U .[checking]\n\n - name: Output installed packages\n run: |\n pip freeze --all\n - name: Output dependency tree\n run: |\n pip install pipdeptree\n pipdeptree\n\n - name: black\n run: black . --check --diff\n - name: flake8\n run: flake8 .\n - name: isort\n run: isort . --check --diff\n - name: mypy\n run: mypy .\n - name: blackdoc\n run: blackdoc . --check --diff\n", "logs": [{"step_name": "checks/9_isort.txt", "log": "##[group]Run isort . --check --diff\n\u001b[36;1misort . --check --diff\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nERROR: /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py Imports are incorrectly sorted and/or formatted.\n--- /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py:before\t2023-12-22 12:41:32.147471\n+++ /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py:after\t2023-12-22 12:42:07.977714\n@@ -1,13 +1,16 @@\n from __future__ import annotations\n \n from collections import defaultdict\n-from typing import List, Optional, Sequence\n+from typing import List\n+from typing import Optional\n+from typing import Sequence\n \n import numpy as np\n \n import optuna\n from optuna.study._study_direction import StudyDirection\n-from optuna.trial import FrozenTrial, TrialState\n+from optuna.trial import FrozenTrial\n+from optuna.trial import TrialState\n \n \n def _get_pareto_front_trials_2d(\nSkipped 30 files\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/optuna/study/_multi_objective.py b/optuna/study/_multi_objective.py\nindex 4e4a36557..18cc8bccf 100644\n--- a/optuna/study/_multi_objective.py\n+++ b/optuna/study/_multi_objective.py\n@@ -1,13 +1,16 @@\n from __future__ import annotations\n \n from collections import defaultdict\n-from typing import List, Optional, Sequence\n+from typing import List\n+from typing import Optional\n+from typing import Sequence\n \n import numpy as np\n \n import optuna\n from optuna.study._study_direction import StudyDirection\n-from optuna.trial import FrozenTrial, TrialState\n+from optuna.trial import FrozenTrial\n+from optuna.trial import TrialState\n \n \n def _get_pareto_front_trials_2d(\n", "difficulty": 0, "changed_files": ["optuna/study/_multi_objective.py"], "commit_link": "https://github.com/optuna/optuna/tree/3137ef65975fc93e9e82b130e545028223cef408"}
 
 
data/python/31c3e1d.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 104, "repo_owner": "deepchecks", "repo_name": "deepchecks", "head_branch": "main", "workflow_name": "Code Style", "workflow_filename": "lint.yml", "workflow_path": ".github/workflows/lint.yml", "contributor": "manijhariya", "sha_fail": "31c3e1dfba76d6e90b5580da2f80fe4b12e6cb5c", "sha_success": "dabfee7048e94ef6a7c8a89159a7ddb887eeacbb", "workflow": "# ----------------------------------------------------------------------------\n# Copyright (C) 2021-2023 Deepchecks (https://www.deepchecks.com)\n#\n# This file is part of Deepchecks.\n# Deepchecks is distributed under the terms of the GNU Affero General\n# Public License (version 3 or later).\n# You should have received a copy of the GNU Affero General Public License\n# along with Deepchecks. If not, see <http://www.gnu.org/licenses/>.\n# ----------------------------------------------------------------------------\n#\nname: Code Style\n\non:\n push:\n branches: [ main ]\n pull_request: \n branches: [ main, nlp-base]\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: true\n\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: \"true\"\n\njobs:\n pylint:\n runs-on: ubuntu-latest\n\n\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n\n - name: Run Pylint\n run: make pylint\n - name: Check import order\n uses: jamescurtin/isort-action@master\n with:\n sortPaths: deepchecks/\n docstring:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n - name: Run Pydocstyle\n run: make docstring\n\n license-check:\n runs-on: ubuntu-latest\n \n needs: docstring\n \n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n - name: Run License eye\n run: make license-check\n", "logs": [{"step_name": "pylint/5_Check import order.txt", "log": "##[group]Run jamescurtin/isort-action@master\nwith:\n sortPaths: deepchecks/\n isortVersion: latest\n configuration: --check-only --diff\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/ensure_python\n\u001b[36;1m$GITHUB_ACTION_PATH/bin/ensure_python\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/install_packages latest \n\u001b[36;1m$GITHUB_ACTION_PATH/bin/install_packages latest \u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Install isort\nInstalling latest version of isort\nCollecting isort[pipfile_deprecated_finder,requirements_deprecated_finder]\n Downloading isort-5.12.0-py3-none-any.whl (91 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 91.2/91.2 kB 4.2 MB/s eta 0:00:00\nCollecting pip-api (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 111.6/111.6 kB 8.3 MB/s eta 0:00:00\nCollecting pipreqs (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pipreqs-0.4.13-py2.py3-none-any.whl (33 kB)\nCollecting pip-shims>=0.5.2 (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip_shims-0.7.3-py2.py3-none-any.whl (35 kB)\nCollecting requirementslib (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading requirementslib-3.0.0-py2.py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: pip>=20 in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder]) (23.0.1)\nCollecting packaging (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder]) (58.1.0)\nCollecting wheel (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading wheel-0.42.0-py3-none-any.whl.metadata (2.2 kB)\nCollecting docopt (from pipreqs->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading docopt-0.6.2.tar.gz (25 kB)\n Installing build dependencies: started\n Installing build dependencies: finished with status 'done'\n Getting requirements to build wheel: started\n Getting requirements to build wheel: finished with status 'done'\n Preparing metadata (pyproject.toml): started\n Preparing metadata (pyproject.toml): finished with status 'done'\nCollecting yarg (from pipreqs->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading yarg-0.1.9-py2.py3-none-any.whl (19 kB)\nCollecting distlib>=0.2.8 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting pep517>=0.5.0 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pep517-0.13.1-py3-none-any.whl.metadata (4.1 kB)\nCollecting pip>=20 (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip-23.3.1-py3-none-any.whl.metadata (3.5 kB)\nCollecting platformdirs (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting plette[validation] (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading plette-0.4.4-py2.py3-none-any.whl (12 kB)\nCollecting pydantic (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pydantic-2.5.2-py3-none-any.whl.metadata (65 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 65.2/65.2 kB 8.9 MB/s eta 0:00:00\nCollecting requests (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting tomlkit>=0.5.3 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading tomlkit-0.12.3-py3-none-any.whl.metadata (2.7 kB)\nCollecting tomli>=1.1.0 (from pep517>=0.5.0->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting cerberus (from plette[validation]->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading Cerberus-1.3.5-py3-none-any.whl.metadata (6.0 kB)\nCollecting annotated-types>=0.4.0 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading annotated_types-0.6.0-py3-none-any.whl.metadata (12 kB)\nCollecting pydantic-core==2.14.5 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pydantic_core-2.14.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB)\nCollecting typing-extensions>=4.6.1 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading typing_extensions-4.8.0-py3-none-any.whl.metadata (3.0 kB)\nCollecting charset-normalizer<4,>=2 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (33 kB)\nCollecting idna<4,>=2.5 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading idna-3.6-py3-none-any.whl.metadata (9.9 kB)\nCollecting urllib3<3,>=1.21.1 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading urllib3-2.1.0-py3-none-any.whl.metadata (6.4 kB)\nCollecting certifi>=2017.4.17 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading certifi-2023.11.17-py3-none-any.whl.metadata (2.2 kB)\nDownloading requirementslib-3.0.0-py2.py3-none-any.whl (96 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 96.7/96.7 kB 8.0 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 10.1 MB/s eta 0:00:00\nDownloading pep517-0.13.1-py3-none-any.whl (19 kB)\nDownloading pip-23.3.1-py3-none-any.whl (2.1 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.1/2.1 MB 17.3 MB/s eta 0:00:00\nDownloading tomlkit-0.12.3-py3-none-any.whl (37 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 15.6 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pydantic-2.5.2-py3-none-any.whl (381 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 381.9/381.9 kB 21.0 MB/s eta 0:00:00\nDownloading pydantic_core-2.14.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.1/2.1 MB 29.9 MB/s eta 0:00:00\nDownloading requests-2.31.0-py3-none-any.whl (62 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 62.6/62.6 kB 18.9 MB/s eta 0:00:00\nUsing cached wheel-0.42.0-py3-none-any.whl (65 kB)\nDownloading annotated_types-0.6.0-py3-none-any.whl (12 kB)\nDownloading certifi-2023.11.17-py3-none-any.whl (162 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 162.5/162.5 kB 33.6 MB/s eta 0:00:00\nDownloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (142 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 142.3/142.3 kB 36.6 MB/s eta 0:00:00\nDownloading idna-3.6-py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.6/61.6 kB 19.5 MB/s eta 0:00:00\nDownloading typing_extensions-4.8.0-py3-none-any.whl (31 kB)\nDownloading urllib3-2.1.0-py3-none-any.whl (104 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.6/104.6 kB 29.6 MB/s eta 0:00:00\nDownloading Cerberus-1.3.5-py3-none-any.whl (30 kB)\nBuilding wheels for collected packages: docopt\n Building wheel for docopt (pyproject.toml): started\n Building wheel for docopt (pyproject.toml): finished with status 'done'\n Created wheel for docopt: filename=docopt-0.6.2-py2.py3-none-any.whl size=13706 sha256=29c03dca51f57ce29e3477c3d0f4958f94387b06a456f0fb38fa4f2f840b837c\n Stored in directory: /home/runner/.cache/pip/wheels/70/4a/46/1309fc853b8d395e60bafaf1b6df7845bdd82c95fd59dd8d2b\nSuccessfully built docopt\nInstalling collected packages: docopt, distlib, cerberus, wheel, urllib3, typing-extensions, tomlkit, tomli, platformdirs, pip, packaging, isort, idna, charset-normalizer, certifi, annotated-types, requests, pydantic-core, plette, pip-shims, pip-api, pep517, yarg, pydantic, requirementslib, pipreqs\n Attempting uninstall: pip\n Found existing installation: pip 23.0.1\n Uninstalling pip-23.0.1:\n Successfully uninstalled pip-23.0.1\nSuccessfully installed annotated-types-0.6.0 cerberus-1.3.5 certifi-2023.11.17 charset-normalizer-3.3.2 distlib-0.3.7 docopt-0.6.2 idna-3.6 isort-5.12.0 packaging-23.2 pep517-0.13.1 pip-23.3.1 pip-api-0.0.30 pip-shims-0.7.3 pipreqs-0.4.13 platformdirs-4.1.0 plette-0.4.4 pydantic-2.5.2 pydantic-core-2.14.5 requests-2.31.0 requirementslib-3.0.0 tomli-2.0.1 tomlkit-0.12.3 typing-extensions-4.8.0 urllib3-2.1.0 wheel-0.42.0 yarg-0.1.9\n##[endgroup]\n##[group]Run echo \"::add-matcher::$GITHUB_ACTION_PATH/isort-matcher.json\"\n\u001b[36;1mecho \"::add-matcher::$GITHUB_ACTION_PATH/isort-matcher.json\"\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/run_isort --check-only --diff deepchecks/\n\u001b[36;1m$GITHUB_ACTION_PATH/bin/run_isort --check-only --diff deepchecks/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nRunning isort --check-only --diff deepchecks/\n##[error]ERROR: /home/runner/work/deepchecks/deepchecks/deepchecks/nlp/utils/text_embeddings.py Imports are incorrectly sorted and/or formatted.\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/deepchecks/nlp/utils/text_embeddings.py b/deepchecks/nlp/utils/text_embeddings.py\nindex 758d3554b..85fc5ad47 100644\n--- a/deepchecks/nlp/utils/text_embeddings.py\n+++ b/deepchecks/nlp/utils/text_embeddings.py\n@@ -9,8 +9,8 @@\n # ----------------------------------------------------------------------------\n #\n \"\"\"Utils module for calculating embeddings for text.\"\"\"\n-import sys\n import re\n+import sys\n import warnings\n from itertools import islice\n from typing import Optional\n", "difficulty": 0, "changed_files": ["deepchecks/nlp/utils/text_embeddings.py"], "commit_link": "https://github.com/deepchecks/deepchecks/tree/31c3e1dfba76d6e90b5580da2f80fe4b12e6cb5c"}
 
 
data/python/385c14d.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 26, "repo_owner": "canonical", "repo_name": "cloud-init", "head_branch": "holman/dhcpcd", "workflow_name": "Lint Tests", "workflow_filename": "check_format.yml", "workflow_path": ".github/workflows/check_format.yml", "contributor": "holmanb", "sha_fail": "385c14d0ae500918cff5565ea836884bfaa2bfa5", "sha_success": "0193e09ca15ed70a351203ab6f21dc52a60d6253", "workflow": "name: Lint Tests\non:\n pull_request:\n push:\n branches:\n - main\n\nconcurrency:\n group: 'ci-${{ github.workflow }}-${{ github.ref }}'\n cancel-in-progress: true\ndefaults:\n run:\n shell: sh -ex {0}\n\njobs:\n check_format:\n strategy:\n fail-fast: false\n matrix:\n env: [ruff, mypy, pylint, black, isort]\n lint-with:\n - {tip-versions: false, os: ubuntu-20.04}\n - {tip-versions: true, os: ubuntu-latest}\n name: Check ${{ matrix.lint-with.tip-versions && 'tip-' || '' }}${{ matrix.env }}\n runs-on: ${{ matrix.lint-with.os }}\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n\n - name: Dependencies\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n\n - name: Print version\n run: python3 --version\n\n - name: Test\n if: ${{ !matrix.lint-with.tip-versions }}\n env:\n # matrix env: not to be confused w/environment variables or testenv\n TOXENV: ${{ matrix.env }}\n run: tox\n - name: Test (tip versions)\n if: matrix.lint-with.tip-versions\n continue-on-error: true\n env:\n TOXENV: tip-${{ matrix.env }}\n run: tox\n schema-format:\n strategy:\n fail-fast: false\n name: Check json format\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Test format\"\n run: |\n tools/check_json_format.sh cloudinit/config/schemas/schema-cloud-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/schema-network-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/versions.schema.cloud-config.json\n\n doc:\n strategy:\n fail-fast: false\n name: Check docs\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n - name: \"Install Python 3.10\"\n uses: actions/setup-python@v4\n with:\n python-version: '3.10.8'\n - name: \"Install dependencies\"\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n - name: \"Spellcheck\"\n run: |\n tox\n env:\n TOXENV: doc-spelling\n - name: \"Build docs\"\n env:\n TOXENV: doc\n run: |\n tox\n\n shell-lint:\n name: Shell Lint\n runs-on: ubuntu-22.04\n steps:\n - name: Checkout\n uses: actions/[email protected]\n\n - name: Install ShellCheck\n run: |\n sudo apt-get update\n sudo apt install shellcheck\n\n - name: Run ShellCheck\n run: |\n shellcheck ./tools/ds-identify\n\n check-cla-signers:\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/[email protected]\n\n - name: Check CLA signers file\n run: tools/check-cla-signers\n", "logs": [{"step_name": "Check pylint/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: pylint\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\npylint create: /home/runner/work/cloud-init/cloud-init/.tox/pylint\npylint installdeps: pylint===2.13.9, -r/home/runner/work/cloud-init/cloud-init/test-requirements.txt, -r/home/runner/work/cloud-init/cloud-init/integration-requirements.txt\npylint inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+77.g80f699fe.zip\npylint installed: adal==1.2.7,anyio==4.2.0,appdirs==1.4.4,applicationinsights==0.11.10,argcomplete==3.1.6,astroid==2.11.7,attrs==23.2.0,autopage==0.5.2,azure-cli-core==2.55.0,azure-cli-telemetry==1.1.0,azure-common==1.1.28,azure-core==1.29.6,azure-identity==1.15.0,azure-mgmt-compute==30.4.0,azure-mgmt-core==1.4.0,azure-mgmt-network==25.2.0,azure-mgmt-resource==23.0.1,bcrypt==4.1.2,boto3==1.34.14,botocore==1.34.14,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,circuitbreaker==1.4.0,cliff==4.4.0,cloud-init==23.4+77.g80f699fe,cmd2==2.4.3,configobj==5.0.8,coverage==7.4.0,cryptography==41.0.7,debtcollector==2.5.0,decorator==5.1.1,dill==0.3.7,distro==1.9.0,dogpile.cache==1.3.0,exceptiongroup==1.2.0,google-api-core==2.15.0,google-api-python-client==2.112.0,google-auth==2.26.1,google-auth-httplib2==0.2.0,googleapis-common-protos==1.62.0,httplib2==0.22.0,humanfriendly==10.0,ibm-cloud-sdk-core==3.18.2,ibm-platform-services==0.48.1,ibm-vpc==0.20.0,idna==3.6,importlib-metadata==7.0.1,importlib-resources==6.1.1,iniconfig==2.0.0,iso8601==2.1.0,isodate==0.6.1,isort==5.13.2,Jinja2==3.1.2,jmespath==1.0.1,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,keystoneauth1==5.4.0,knack==0.11.0,lazy-object-proxy==1.10.0,MarkupSafe==2.1.3,mccabe==0.7.0,msal==1.24.0b2,msal-extensions==1.0.0,msgpack==1.0.7,msrest==0.7.1,msrestazure==0.6.4,netaddr==0.10.1,netifaces==0.11.0,oauthlib==3.2.2,oci==2.118.0,openstacksdk==1.4.0,os-service-types==1.7.0,osc-lib==2.9.0,oslo.config==9.2.0,oslo.i18n==6.2.0,oslo.serialization==5.2.0,oslo.utils==6.3.0,packaging==23.2,paramiko==3.4.0,passlib==1.7.4,pbr==6.0.0,pkginfo==1.9.6,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pluggy==1.3.0,portalocker==2.8.2,prettytable==3.9.0,protobuf==3.19.6,psutil==5.9.7,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycloudlib==1!5.12.1,pycparser==2.21,pygments==2.17.2,PyJWT==2.8.0,pylint==2.13.9,PyNaCl==1.5.0,pyOpenSSL==23.3.0,pyparsing==2.4.7,pyperclip==1.8.2,pyserial==3.5,pytest==7.4.4,pytest-cov==4.1.0,pytest-mock==3.12.0,python-cinderclient==9.4.0,python-dateutil==2.8.2,python-keystoneclient==5.2.0,python-novaclient==18.4.0,python-openstackclient==6.4.0,pytz==2023.3.post1,PyYAML==6.0.1,qemu.qmp==0.0.3,referencing==0.32.1,requests==2.31.0,requests-oauthlib==1.3.1,requestsexceptions==1.4.0,responses==0.24.1,rfc3986==2.0.0,rpds-py==0.16.2,rsa==4.9,s3transfer==0.10.0,simplejson==3.19.2,six==1.16.0,sniffio==1.3.0,stevedore==5.1.0,tabulate==0.9.0,toml==0.10.0,tomli==2.0.1,typing-extensions==4.9.0,tzdata==2023.4,uritemplate==4.1.1,urllib3==2.1.0,wcwidth==0.2.12,wrapt==1.16.0,zipp==3.17.0\npylint run-test-pre: PYTHONHASHSEED='2276146897'\npylint run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py\n************* Module cloudinit.net.dhcp\ncloudinit/net/dhcp.py:575: [W0612(unused-variable), Dhcpcd.dhcp_discovery] Unused variable 'out'\ncloudinit/net/dhcp.py:575: [W0612(unused-variable), Dhcpcd.dhcp_discovery] Unused variable 'err'\n\n------------------------------------\nYour code has been rated at 10.00/10\n\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py (exited with code 4)\n___________________________________ summary ____________________________________\nERROR: pylint: commands failed\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Check black/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: black\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\nblack create: /home/runner/work/cloud-init/cloud-init/.tox/black\nblack installdeps: black===22.3.0\nblack inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+77.g80f699fe.zip\nblack installed: attrs==23.2.0,black==22.3.0,certifi==2023.11.17,charset-normalizer==3.3.2,click==8.1.7,cloud-init==23.4+77.g80f699fe,configobj==5.0.8,idna==3.6,importlib-resources==6.1.1,Jinja2==3.1.2,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,MarkupSafe==2.1.3,mypy-extensions==1.0.0,netifaces==0.11.0,oauthlib==3.2.2,pathspec==0.12.1,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pyserial==3.5,PyYAML==6.0.1,referencing==0.32.1,requests==2.31.0,rpds-py==0.16.2,six==1.16.0,tomli==2.0.1,typing-extensions==4.9.0,urllib3==2.1.0,zipp==3.17.0\nblack run-test-pre: PYTHONHASHSEED='3554362748'\nblack run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/black/bin/python -m black . --check\nwould reformat tests/unittests/net/test_dhcp.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 549 files would be left unchanged.\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/black/bin/python -m black . --check (exited with code 1)\n___________________________________ summary ____________________________________\nERROR: black: commands failed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/cloudinit/net/dhcp.py b/cloudinit/net/dhcp.py\nindex 1b8caee4f..6a4e5a3fa 100644\n--- a/cloudinit/net/dhcp.py\n+++ b/cloudinit/net/dhcp.py\n@@ -175,6 +175,11 @@ class DhcpClient(abc.ABC):\n client_name = \"\"\n max_wait = 5\n \n+ def __init__(self):\n+ self.dhcp_client_path = subp.which(self.client_name)\n+ if not self.dhcp_client_path:\n+ raise NoDHCPLeaseMissingDhclientError()\n+\n @classmethod\n def kill_dhcp_client(cls):\n subp.subp([\"pkill\", cls.client_name], rcs=[0, 1])\n@@ -200,14 +205,6 @@ class DhcpClient(abc.ABC):\n class IscDhclient(DhcpClient):\n client_name = \"dhclient\"\n \n- def __init__(self):\n- self.dhclient_path = subp.which(\"dhclient\")\n- if not self.dhclient_path:\n- LOG.debug(\n- \"Skip dhclient configuration: No dhclient command found.\"\n- )\n- raise NoDHCPLeaseMissingDhclientError()\n-\n @staticmethod\n def parse_dhcp_lease_file(lease_file: str) -> List[Dict[str, Any]]:\n \"\"\"Parse the given dhcp lease file returning all leases as dicts.\n@@ -298,7 +295,7 @@ class IscDhclient(DhcpClient):\n try:\n out, err = subp.subp(\n distro.build_dhclient_cmd(\n- self.dhclient_path,\n+ self.dhcp_client_path,\n lease_file,\n pid_file,\n interface,\n@@ -536,14 +533,6 @@ class IscDhclient(DhcpClient):\n class Dhcpcd:\n client_name = \"dhcpcd\"\n \n- def __init__(self):\n- self.dhclient_path = subp.which(\"dhcpcd\")\n- if not self.dhclient_path:\n- LOG.debug(\n- \"Skip dhclient configuration: No dhclient command found.\"\n- )\n- raise NoDHCPLeaseMissingDhclientError()\n-\n def dhcp_discovery(\n self,\n interface,\n@@ -572,9 +561,9 @@ class Dhcpcd:\n # /lib/dhcpcd/dhcpcd-hooks/ and pass each of those with the --nohook\n # argument to dhcpcd\n try:\n- out, err = subp.subp(\n+ subp.subp(\n [\n- \"dhcpcd\",\n+ self.client_name,\n \"--oneshot\", # get lease then exit\n \"--nobackground\", # don't fork\n \"--ipv4only\", # only attempt configuring ipv4\n@@ -584,6 +573,7 @@ class Dhcpcd:\n interface,\n ]\n )\n+ return self.parse_dhcp_lease_file(interface)\n except subp.ProcessExecutionError as error:\n LOG.debug(\n \"dhclient exited with code: %s stderr: %r stdout: %r\",\n@@ -592,7 +582,6 @@ class Dhcpcd:\n error.stdout,\n )\n raise NoDHCPLeaseError from error\n- return self.parse_dhcp_lease_file(interface)\n \n @staticmethod\n def parse_dhcpcd_lease(lease_dump: str, interface: str) -> List[dict]:\n@@ -656,7 +645,7 @@ class Dhcpcd:\n \n except subp.ProcessExecutionError as error:\n LOG.debug(\n- \"dhclient exited with code: %s stderr: %r stdout: %r\",\n+ \"dhcpcd exited with code: %s stderr: %r stdout: %r\",\n error.exit_code,\n error.stderr,\n error.stdout,\n@@ -667,12 +656,6 @@ class Dhcpcd:\n class Udhcpc(DhcpClient):\n client_name = \"udhcpc\"\n \n- def __init__(self):\n- self.udhcpc_path = subp.which(\"udhcpc\")\n- if not self.udhcpc_path:\n- LOG.debug(\"Skip udhcpc configuration: No udhcpc command found.\")\n- raise NoDHCPLeaseMissingUdhcpcError()\n-\n def dhcp_discovery(\n self,\n interface,\n@@ -702,7 +685,7 @@ class Udhcpc(DhcpClient):\n util.write_file(udhcpc_script, UDHCPC_SCRIPT, 0o755)\n \n cmd = [\n- self.udhcpc_path,\n+ self.dhcp_client_path,\n \"-O\",\n \"staticroutes\",\n \"-i\",\ndiff --git a/tests/unittests/net/test_dhcp.py b/tests/unittests/net/test_dhcp.py\nindex 2841a7244..fa8308068 100644\n--- a/tests/unittests/net/test_dhcp.py\n+++ b/tests/unittests/net/test_dhcp.py\n@@ -8,6 +8,7 @@ import pytest\n import responses\n \n from cloudinit.net.dhcp import (\n+ Dhcpcd,\n InvalidDHCPLeaseFileError,\n IscDhclient,\n NoDHCPLeaseError,\n@@ -438,7 +439,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n subp.ProcessExecutionError(exit_code=-5),\n ]\n \n- m_which.side_effect = [False, True]\n+ m_which.side_effect = [False, False, False]\n with pytest.raises(NoDHCPLeaseError):\n maybe_perform_dhcp_discovery(MockDistro())\n \n@@ -446,10 +447,6 @@ class TestDHCPDiscoveryClean(CiTestCase):\n \"DHCP client not found: dhclient\",\n self.logs.getvalue(),\n )\n- self.assertIn(\n- \"DHCP client not found: dhcpcd\",\n- self.logs.getvalue(),\n- )\n \n @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=None)\n def test_provided_nic_does_not_exist(self, m_fallback_nic):\n@@ -473,7 +470,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n maybe_perform_dhcp_discovery(MockDistro())\n \n self.assertIn(\n- \"Skip dhclient configuration: No dhclient command found.\",\n+ \"DHCP client not found: dhclient\",\n self.logs.getvalue(),\n )\n \n@@ -975,24 +972,6 @@ class TestUDHCPCDiscoveryClean(CiTestCase):\n with_logs = True\n maxDiff = None\n \n- @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n- def test_absent_udhcpc_command(self, m_fallback, m_which):\n- \"\"\"When dhclient doesn't exist in the OS, log the issue and no-op.\"\"\"\n- m_fallback.return_value = \"eth9\"\n- m_which.return_value = None # udhcpc isn't found\n-\n- distro = MockDistro()\n- distro.dhcp_client_priority = [Udhcpc]\n-\n- with pytest.raises(NoDHCPLeaseMissingDhclientError):\n- maybe_perform_dhcp_discovery(distro)\n-\n- self.assertIn(\n- \"Skip udhcpc configuration: No udhcpc command found.\",\n- self.logs.getvalue(),\n- )\n-\n @mock.patch(\"cloudinit.net.dhcp.is_ib_interface\", return_value=False)\n @mock.patch(\"cloudinit.net.dhcp.subp.which\", return_value=\"/sbin/udhcpc\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n@@ -1135,3 +1114,30 @@ class TestUDHCPCDiscoveryClean(CiTestCase):\n ),\n ]\n )\n+\n+\n+class TestDhcpcd:\n+ def test_parse_lease(self):\n+ lease = dedent(\n+ \"\"\"\n+ broadcast_address='192.168.15.255'\n+ dhcp_lease_time='3600'\n+ dhcp_message_type='5'\n+ dhcp_server_identifier='192.168.0.1'\n+ domain_name='us-east-2.compute.internal'\n+ domain_name_servers='192.168.0.2'\n+ host_name='ip-192-168-0-212'\n+ interface_mtu='9001'\n+ ip_address='192.168.0.212'\n+ network_number='192.168.0.0'\n+ routers='192.168.0.1'\n+ subnet_cidr='20'\n+ subnet_mask='255.255.240.0'\n+ \"\"\"\n+ )\n+ parsed_lease = Dhcpcd.parse_dhcpcd_lease(lease, \"eth0\")[0]\n+ assert \"eth0\" == parsed_lease[\"interface\"]\n+ assert \"192.168.15.255\" == parsed_lease[\"broadcast-address\"]\n+ assert \"192.168.0.212\" == parsed_lease[\"fixed-address\"]\n+ assert \"255.255.240.0\" == parsed_lease[\"subnet-mask\"]\n+ assert \"192.168.0.1\" == parsed_lease[\"routers\"]\n", "difficulty": 0, "changed_files": ["cloudinit/net/dhcp.py", "tests/unittests/net/test_dhcp.py"], "commit_link": "https://github.com/canonical/cloud-init/tree/385c14d0ae500918cff5565ea836884bfaa2bfa5"}
 
 
data/python/386bf6f.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 196, "repo_owner": "danielegrattarola", "repo_name": "spektral", "head_branch": "master", "workflow_name": "Run examples", "workflow_filename": "examples.yml", "workflow_path": ".github/workflows/examples.yml", "contributor": "danielegrattarola", "sha_fail": "386bf6f0815368b78261be43bf90e203dfe9c13f", "sha_success": "bf6071475028cf9711b8264e8952deb2aee2e32b", "workflow": "name: Run examples\n\non: [push, pull_request]\n\njobs:\n build:\n\n runs-on: ubuntu-latest\n\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.11\n uses: actions/setup-python@v2\n with:\n python-version: 3.11\n - name: Install dependencies\n run: |\n pip install ogb matplotlib\n - name: Install Spektral\n run: |\n pip install .\n - name: Just one epoch\n run: |\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/node_prediction/*.py\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/graph_prediction/*.py\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/other/*.py\n - name: Run all examples\n run: |\n cd examples/node_prediction/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n cd graph_prediction/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n cd other/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n", "logs": [{"step_name": "build/7_Run all examples.txt", "log": "##[group]Run cd examples/node_prediction/\n\u001b[36;1mcd examples/node_prediction/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\n\u001b[36;1mcd graph_prediction/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\n\u001b[36;1mcd other/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n##### citation_arma.py #####\n2024-01-21 16:14:04.697281: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:04.697333: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:04.698603: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:04.704823: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:05.946659: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nDownloading cora dataset.\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n arma_conv (ARMAConv) (None, 16) 91744 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n dropout (Dropout) (None, 16) 0 ['arma_conv[0][0]'] \n \n arma_conv_1 (ARMAConv) (None, 7) 231 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 91975 (359.28 KB)\nTrainable params: 91975 (359.28 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 0.1167 - acc: 0.1714\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 1s 938ms/step - loss: 0.1167 - acc: 0.1714 - val_loss: 0.3394 - val_acc: 0.3080\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 0.6695 - acc: 0.3360\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 12ms/step - loss: 0.6695 - acc: 0.3360\nDone.\nTest loss: 0.6695395708084106\nTest accuracy: 0.335999995470047\n##### citation_cheby.py #####\n2024-01-21 16:14:13.031828: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:13.031877: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:13.033233: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:13.039361: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:13.892745: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n dropout (Dropout) (None, 1433) 0 ['input_1[0][0]'] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n cheb_conv (ChebConv) (None, 16) 45856 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n dropout_1 (Dropout) (None, 16) 0 ['cheb_conv[0][0]'] \n \n cheb_conv_1 (ChebConv) (None, 7) 224 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 46080 (180.00 KB)\nTrainable params: 46080 (180.00 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 1.9901 - acc: 0.1357\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 492ms/step - loss: 1.9901 - acc: 0.1357 - val_loss: 1.9229 - val_acc: 0.2740\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9236 - acc: 0.2650\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 15ms/step - loss: 1.9236 - acc: 0.2650\nDone.\nTest loss: 1.9235953092575073\nTest accuracy: 0.26499998569488525\n##### citation_gat.py #####\n2024-01-21 16:14:17.870770: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:17.870818: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:17.872219: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:17.878280: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:18.689165: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nPre-processing node features\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n dropout (Dropout) (None, 1433) 0 ['input_1[0][0]'] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n gat_conv (GATConv) (None, 64) 91904 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n dropout_1 (Dropout) (None, 64) 0 ['gat_conv[0][0]'] \n \n gat_conv_1 (GATConv) (None, 7) 469 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 92373 (360.83 KB)\nTrainable params: 92373 (360.83 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 1.9498 - acc: 0.1429\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 2s 2s/step - loss: 1.9498 - acc: 0.1429 - val_loss: 1.9496 - val_acc: 0.2020\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9495 - acc: 0.2120\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 15ms/step - loss: 1.9495 - acc: 0.2120\nDone.\nTest loss: 1.9495328664779663\nTest accuracy: 0.21199998259544373\n##### citation_gat_custom.py #####\n2024-01-21 16:14:24.090507: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:24.090565: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:24.092222: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:24.098707: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:24.924070: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nPre-processing node features\nLoss tr: 1.9471, Acc tr: 0.3214, Loss va: 1.9493, Acc va: 0.1920, Loss te: 1.9494, Acc te: 0.1790\nImproved\nGAT (1 epochs)\nElapsed: 2.46s\n##### citation_gcn.py #####\n2024-01-21 16:14:30.891866: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:30.891915: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:30.893225: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:30.899400: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:31.735939: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nPre-processing node features\n\n1/1 [==============================] - ETA: 0s - loss: 1.9535 - acc: 0.1286\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 1s 1s/step - loss: 1.9535 - acc: 0.1286 - val_loss: 1.9469 - val_acc: 0.4000\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9470 - acc: 0.3910\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 9ms/step - loss: 1.9470 - acc: 0.3910\nDone.\nTest loss: 1.946979284286499\nTest accuracy: 0.390999972820282\n##### citation_gcn_custom.py #####\n2024-01-21 16:14:35.783612: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:35.783669: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:35.785139: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:35.791466: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:36.608646: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nPre-processing node features\nSpektral - GCN (200 epochs)\nElapsed: 2.85s\nFinal loss = 0.617151141166687\n##### citation_simple_gc.py #####\n2024-01-21 16:14:43.435717: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:43.435759: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:43.437049: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:43.443159: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:44.284258: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n gcn_conv (GCNConv) (None, 7) 10031 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 10031 (39.18 KB)\nTrainable params: 10031 (39.18 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 0.1008 - acc: 0.1357\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 355ms/step - loss: 0.1008 - acc: 0.1357 - val_loss: 0.2166 - val_acc: 0.6780\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 0.4094 - acc: 0.7070\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 12ms/step - loss: 0.4094 - acc: 0.7070\nDone.\nTest loss: 0.40943244099617004\nTest accuracy: 0.7070000171661377\n##### ogbn-arxiv_gcn.py #####\n2024-01-21 16:14:47.998765: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:47.998814: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:48.000133: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:48.006402: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:48.837666: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\nDownloading http://snap.stanford.edu/ogb/data/nodeproppred/arxiv.zip\n\n 0%| | 0/81 [00:00<?, ?it/s]\nDownloaded 0.00 GB: 0%| | 0/81 [00:00<?, ?it/s]\nDownloaded 0.00 GB: 1%| | 1/81 [00:00<01:13, 1.09it/s]\nDownloaded 0.00 GB: 1%| | 1/81 [00:01<01:13, 1.09it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.00 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:01<00:02, 23.06it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.06 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:02<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.06 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81/81 [00:03<00:00, 23.38it/s]\n\n 0%| | 0/1 [00:00<?, ?it/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1/1 [00:00<00:00, 21076.90it/s]\nExtracting dataset/arxiv.zip\nLoading necessary files...\nThis might take a while.\nProcessing graphs...\nSaving...\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 128)] 0 [] \n \n input_2 (InputLayer) [(None, 169343)] 0 [] \n \n gcn_conv (GCNConv) (None, 256) 33024 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n batch_normalization (Batch (None, 256) 1024 ['gcn_conv[0][0]'] \n Normalization) \n \n dropout (Dropout) (None, 256) 0 ['batch_normalization[0][0]'] \n \n gcn_conv_1 (GCNConv) (None, 256) 65792 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n batch_normalization_1 (Bat (None, 256) 1024 ['gcn_conv_1[0][0]'] \n chNormalization) \n \n dropout_1 (Dropout) (None, 256) 0 ['batch_normalization_1[0][0]'\n ] \n \n gcn_conv_2 (GCNConv) (None, 40) 10280 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 111144 (434.16 KB)\nTrainable params: 110120 (430.16 KB)\nNon-trainable params: 1024 (4.00 KB)\n__________________________________________________________________________________________________\nEp. 1 - Loss: 5.289 - Acc: 0.109 - Val acc: 0.177 - Test acc: 0.241\nEvaluating model.\nDone! - Test acc: 0.241\n##### custom_dataset.py #####\n2024-01-21 16:15:07.319425: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:15:07.319469: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:15:07.320948: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:15:07.327116: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:15:08.139878: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/data/utils.py:221: UserWarning: you are shuffling a 'MyDataset' object which is not a subclass of 'Sequence'; `shuffle` is not guaranteed to behave correctly. E.g., non-numpy array/tensor objects with view semantics may contain duplicates after shuffling.\n np.random.shuffle(a)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nEp. 1 - Loss: 0.729 - Acc: 0.694 - Val loss: 0.334 - Val acc: 0.880\nNew best val_loss 0.334\nDone. Test loss: 0.4293. Test acc: 0.84\n##### general_gnn.py #####\n2024-01-21 16:15:13.211838: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:15:13.211888: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:15:13.213185: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:15:13.219400: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:15:14.044203: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\nDownloading PROTEINS dataset.\n\n 0%| | 0.00/447k [00:00<?, ?B/s]\n 7%|\u2588\u2588\u258a | 32.0k/447k [00:00<00:01, 249kB/s]\n 21%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 96.0k/447k [00:00<00:00, 419kB/s]\n 50%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 224k/447k [00:00<00:00, 777kB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 447k/447k [00:00<00:00, 1.02MB/s]\nTraceback (most recent call last):\n File \"/home/runner/work/spektral/spektral/examples/graph_prediction/general_gnn.py\", line 39, in <module>\n data = TUDataset(\"PROTEINS\")\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 66, in __init__\n super().__init__(**kwargs)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/data/dataset.py\", line 118, in __init__\n self.graphs = self.read()\n ^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 128, in read\n [_normalize(xl_[:, None], \"ohe\") for xl_ in x_labs.T], -1\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 128, in <listcomp>\n [_normalize(xl_[:, None], \"ohe\") for xl_ in x_labs.T], -1\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 219, in _normalize\n fnorm = OneHotEncoder(sparse=False, categories=\"auto\")\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nTypeError: OneHotEncoder.__init__() got an unexpected keyword argument 'sparse'\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/spektral/datasets/tudataset.py b/spektral/datasets/tudataset.py\nindex 3877535..f26b0de 100644\n--- a/spektral/datasets/tudataset.py\n+++ b/spektral/datasets/tudataset.py\n@@ -216,7 +216,7 @@ def _normalize(x, norm=None):\n Apply one-hot encoding or z-score to a list of node features\n \"\"\"\n if norm == \"ohe\":\n- fnorm = OneHotEncoder(sparse=False, categories=\"auto\")\n+ fnorm = OneHotEncoder(sparse_output=False, categories=\"auto\")\n elif norm == \"zscore\":\n fnorm = StandardScaler()\n else:\n", "difficulty": 2, "changed_files": ["spektral/datasets/tudataset.py"], "commit_link": "https://github.com/danielegrattarola/spektral/tree/386bf6f0815368b78261be43bf90e203dfe9c13f"}
 
 
data/python/38a3024.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 165, "repo_owner": "prompt-toolkit", "repo_name": "python-prompt-toolkit", "head_branch": "master", "workflow_name": "test", "workflow_filename": "test.yaml", "workflow_path": ".github/workflows/test.yaml", "contributor": "auxsvr", "sha_fail": "38a30240c4e21de4c03322cfbf982119094bdf45", "sha_success": "55cde3eb0e1be72d9232e979a708f1a792339ad9", "workflow": "name: test\n\non:\n push: # any branch\n pull_request:\n branches: [master]\n\nenv:\n FORCE_COLOR: 1\n\njobs:\n test-ubuntu:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n\n steps:\n - uses: actions/checkout@v4\n - name: Setup Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n allow-prereleases: true\n - name: Install Dependencies\n run: |\n sudo apt remove python3-pip\n python -m pip install --upgrade pip\n python -m pip install . ruff typos coverage codecov mypy pytest readme_renderer types-contextvars asyncssh\n pip list\n - name: Ruff\n run: |\n ruff .\n ruff format --check .\n typos .\n - name: Tests\n run: |\n coverage run -m pytest\n - if: \"matrix.python-version != '3.7'\"\n name: Mypy\n # Check whether the imports were sorted correctly.\n # When this fails, please run ./tools/sort-imports.sh\n run: |\n mypy --strict src/prompt_toolkit --platform win32\n mypy --strict src/prompt_toolkit --platform linux\n mypy --strict src/prompt_toolkit --platform darwin\n - name: Validate README.md\n # Ensure that the README renders correctly (required for uploading to PyPI).\n run: |\n python -m readme_renderer README.rst > /dev/null\n - name: Run codecov\n run: |\n codecov\n", "logs": [{"step_name": "test-ubuntu (3.7)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.8)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.9)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.10)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.11)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.11.6/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.12)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/src/prompt_toolkit/application/application.py b/src/prompt_toolkit/application/application.py\nindex 6a8f0290..2dcb3130 100644\n--- a/src/prompt_toolkit/application/application.py\n+++ b/src/prompt_toolkit/application/application.py\n@@ -985,7 +985,7 @@ class Application(Generic[_AppResult]):\n # See whether a loop was installed already. If so, use that.\n # That's required for the input hooks to work, they are\n # installed using `set_event_loop`.\n- loop = asyncio.get_running_loop()\n+ loop = asyncio.get_event_loop()\n except RuntimeError:\n # No loop installed. Run like usual.\n return asyncio.run(coro)\n@@ -1545,7 +1545,7 @@ async def _do_wait_for_enter(wait_text: AnyFormattedText) -> None:\n \n @contextmanager\n def attach_winch_signal_handler(\n- handler: Callable[[], None]\n+ handler: Callable[[], None],\n ) -> Generator[None, None, None]:\n \"\"\"\n Attach the given callback as a WINCH signal handler within the context\ndiff --git a/src/prompt_toolkit/eventloop/inputhook.py b/src/prompt_toolkit/eventloop/inputhook.py\nindex 5731573f..a4c0eee6 100644\n--- a/src/prompt_toolkit/eventloop/inputhook.py\n+++ b/src/prompt_toolkit/eventloop/inputhook.py\n@@ -66,7 +66,7 @@ InputHook: TypeAlias = Callable[[InputHookContext], None]\n \n \n def new_eventloop_with_inputhook(\n- inputhook: Callable[[InputHookContext], None]\n+ inputhook: Callable[[InputHookContext], None],\n ) -> AbstractEventLoop:\n \"\"\"\n Create a new event loop with the given inputhook.\n@@ -77,7 +77,7 @@ def new_eventloop_with_inputhook(\n \n \n def set_eventloop_with_inputhook(\n- inputhook: Callable[[InputHookContext], None]\n+ inputhook: Callable[[InputHookContext], None],\n ) -> AbstractEventLoop:\n \"\"\"\n Create a new event loop with the given inputhook, and activate it.\ndiff --git a/src/prompt_toolkit/patch_stdout.py b/src/prompt_toolkit/patch_stdout.py\nindex 46619cac..528bec7f 100644\n--- a/src/prompt_toolkit/patch_stdout.py\n+++ b/src/prompt_toolkit/patch_stdout.py\n@@ -203,6 +203,13 @@ class StdoutProxy:\n \"\"\"\n \n def write_and_flush() -> None:\n+ # Ensure that autowrap is enabled before calling `write`.\n+ # XXX: On Windows, the `Windows10_Output` enables/disables VT\n+ # terminal processing for every flush. It turns out that this\n+ # causes autowrap to be reset (disabled) after each flush. So,\n+ # we have to enable it again before writing text.\n+ self._output.enable_autowrap()\n+\n if self.raw:\n self._output.write_raw(text)\n else:\n", "difficulty": 0, "changed_files": ["src/prompt_toolkit/application/application.py", "src/prompt_toolkit/eventloop/inputhook.py", "src/prompt_toolkit/patch_stdout.py"], "commit_link": "https://github.com/prompt-toolkit/python-prompt-toolkit/tree/38a30240c4e21de4c03322cfbf982119094bdf45"}
 
 
data/python/3dd8e44.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 128, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "device-agnostic-testing", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "statelesshz", "sha_fail": "3dd8e4404a0ce2e29db4911dc2cd7e94755be631", "sha_success": "c9a13fe9a890cfb48ee6c7b1c6abd64c8fb432da", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\ntests/deepspeed/test_deepspeed.py:15:1: I001 [*] Import block is un-sorted or un-formatted\ntests/fsdp/test_fsdp.py:16:1: I001 [*] Import block is un-sorted or un-formatted\nFound 2 errors.\n[*] 2 fixable with the `--fix` option.\nmake: *** [Makefile:16: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/tests/deepspeed/test_deepspeed.py b/tests/deepspeed/test_deepspeed.py\nindex 035965c..612f234 100644\n--- a/tests/deepspeed/test_deepspeed.py\n+++ b/tests/deepspeed/test_deepspeed.py\n@@ -36,9 +36,9 @@ from accelerate.test_utils.testing import (\n AccelerateTestCase,\n TempDirTestCase,\n execute_subprocess_async,\n- require_non_cpu,\n require_deepspeed,\n require_multi_device,\n+ require_non_cpu,\n slow,\n )\n from accelerate.test_utils.training import RegressionDataset\ndiff --git a/tests/fsdp/test_fsdp.py b/tests/fsdp/test_fsdp.py\nindex d939991..c494f5e 100644\n--- a/tests/fsdp/test_fsdp.py\n+++ b/tests/fsdp/test_fsdp.py\n@@ -28,9 +28,9 @@ from accelerate.test_utils.testing import (\n AccelerateTestCase,\n TempDirTestCase,\n execute_subprocess_async,\n- require_non_cpu,\n require_fsdp,\n require_multi_device,\n+ require_non_cpu,\n slow,\n )\n from accelerate.utils.constants import (\n", "difficulty": 0, "changed_files": ["tests/deepspeed/test_deepspeed.py", "tests/fsdp/test_fsdp.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/3dd8e4404a0ce2e29db4911dc2cd7e94755be631"}
 
 
data/python/3ed7a88.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/3fcfe29.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 110, "repo_owner": "mikf", "repo_name": "gallery-dl", "head_branch": "master", "workflow_name": "tests", "workflow_filename": "tests.yml", "workflow_path": ".github/workflows/tests.yml", "contributor": "bun-dev", "sha_fail": "3fcfe2925441e86f02d700a3bd080ee7d4c85153", "sha_success": "c51a3c67884331c409a05779bcec57ecd402cd18", "workflow": "name: tests\n\non:\n push:\n branches:\n - master\n pull_request:\n branches:\n - master\n\njobs:\n build:\n\n runs-on: ubuntu-20.04\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.5\", \"3.6\", \"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\", \"pypy3.9\"]\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Check file permissions\n run: |\n if [[ \"$(find ./gallery_dl -type f -not -perm 644)\" ]]; then exit 1; fi\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Install dependencies\n run: |\n pip install -r requirements.txt\n pip install flake8 youtube-dl\n\n - name: Install yt-dlp\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5)\n # don't install yt-dlp\n ;;\n 3.6|3.7)\n # install from PyPI\n pip install yt-dlp\n ;;\n *)\n # install from master\n pip install https://github.com/yt-dlp/yt-dlp/archive/refs/heads/master.tar.gz\n ;;\n esac\n\n - name: Lint with flake8\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5|3.6|3.7)\n flake8 --extend-exclude scripts/export_tests.py .\n ;;\n *)\n flake8 .\n ;;\n esac\n\n - name: Run tests\n run: |\n make test\n\n - name: Test autogeneration of man pages, bash/zsh/fish completion, etc\n run: |\n make\n", "logs": [{"step_name": "build (3.5)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.5\" in\n\u001b[36;1mcase \"3.5\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.5.10/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.6)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.6\" in\n\u001b[36;1mcase \"3.6\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.6.15/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.7)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.7\" in\n\u001b[36;1mcase \"3.7\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.8)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.8\" in\n\u001b[36;1mcase \"3.8\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.9\" in\n\u001b[36;1mcase \"3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.10)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.10\" in\n\u001b[36;1mcase \"3.10\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.11)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.11\" in\n\u001b[36;1mcase \"3.11\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.12)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.12\" in\n\u001b[36;1mcase \"3.12\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (pypy3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"pypy3.9\" in\n\u001b[36;1mcase \"pypy3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/PyPy/3.9.18/x64/bin/lib/pkgconfig\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gallery_dl/extractor/shimmie2.py b/gallery_dl/extractor/shimmie2.py\nindex a9ed615b..299b2394 100644\n--- a/gallery_dl/extractor/shimmie2.py\n+++ b/gallery_dl/extractor/shimmie2.py\n@@ -223,7 +223,7 @@ class Shimmie2TagExtractor(Shimmie2Extractor):\n break\n \n tags, dimensions, size, ext = extr(\n- \"title=\\\"\", \"\\\"\").split(\" // \")\n+ \"title=\\\"\", \"\\\"\").split(\" // \")\n width, _, height = dimensions.partition(\"x\")\n md5 = extr(\"/_thumbs/\", \"/\")\n \n", "difficulty": 0, "changed_files": ["gallery_dl/extractor/shimmie2.py"], "commit_link": "https://github.com/mikf/gallery-dl/tree/3fcfe2925441e86f02d700a3bd080ee7d4c85153"}
 
 
data/python/434321a.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/43dd59c.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 9, "repo_owner": "yt-dlp", "repo_name": "yt-dlp", "head_branch": "master", "workflow_name": "Quick Test", "workflow_filename": "quick-test.yml", "workflow_path": ".github/workflows/quick-test.yml", "contributor": "Kenshin9977", "sha_fail": "43dd59c3137df77f5dd22cef4cb7bedfe9f6b12e", "sha_success": "0bedc0ee590f45a6229a4b78f83297f8c07392bb", "workflow": "name: Quick Test\non: [push, pull_request]\npermissions:\n contents: read\n\njobs:\n tests:\n name: Core Test\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python 3.11\n uses: actions/setup-python@v4\n with:\n python-version: '3.11'\n - name: Install test requirements\n run: pip install pytest pycryptodomex\n - name: Run tests\n run: |\n python3 -m yt_dlp -v || true\n ./devscripts/run_tests.sh core\n flake8:\n name: Linter\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n - name: Install flake8\n run: pip install flake8\n - name: Make lazy extractors\n run: python devscripts/make_lazy_extractors.py\n - name: Run flake8\n run: flake8 .\n", "logs": [{"step_name": "Core Test/5_Run tests.txt", "log": "##[group]Run python3 -m yt_dlp -v || true\n\u001b[36;1mpython3 -m yt_dlp -v || true\u001b[0m\n\u001b[36;1m./devscripts/run_tests.sh core\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.6/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib\n##[endgroup]\n[debug] Command-line config: ['-v']\n[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8 (No ANSI), error utf-8 (No ANSI), screen utf-8 (No ANSI)\n[debug] yt-dlp version [email protected] [b634ba742] (source)\n[debug] Lazy loading extractors is disabled\n[debug] Git HEAD: f8abf45\n[debug] Python 3.11.6 (CPython x86_64 64bit) - Linux-6.2.0-1015-azure-x86_64-with-glibc2.35 (OpenSSL 3.0.2 15 Mar 2022, glibc 2.35)\n[debug] exe versions: none\n[debug] Optional libraries: Cryptodome-3.19.0, sqlite3-3.37.2\n[debug] Proxy map: {}\n[debug] Request Handlers: urllib\n[debug] Loaded 1890 extractors\n\nUsage: yt-dlp [OPTIONS] URL [URL...]\n\nyt-dlp: error: You must provide at least one URL.\nType yt-dlp --help to see a list of all options.\n============================= test session starts ==============================\nplatform linux -- Python 3.11.6, pytest-7.4.3, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.11.6/x64/bin/python3\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yt-dlp/yt-dlp\nconfigfile: setup.cfg\ncollecting ... collected 6320 items / 5705 deselected / 615 selected\n\ntest/test_InfoExtractor.py::TestInfoExtractor::test_download_json PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_extract_jwplayer_data_realworld PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_html_search_meta PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_html_search_regex PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_ie_key PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_opengraph PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_f4m_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_html5_media_entries PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_ism_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_m3u8_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_mpd_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_xspf PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_response_with_expected_status_returns_content PASSED [ 2%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_search_json_ld_realworld PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_audio_only_extractor_format_selection PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_default_format_spec PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_filtering PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_not_available PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_audio PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_audio_exts PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_issue_10083 PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_string_ops PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_video PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_invalid_format_specs PASSED [ 4%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_prefer_free_formats PASSED [ 4%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_youtube_format_selection PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_add_extra_info PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_add_headers_cookie PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_do_not_override_ie_key_in_url_transparent PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_format_note PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_header_cookies PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_ignoreerrors_for_playlist_with_url_transparent_iterable_entries PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_infojson_cookies PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_match_filter PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_playlist_items_selection PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_postprocessors PASSED [ 6%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_prepare_outtmpl_and_filename PASSED [ 6%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_subtitles PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_get_cookie_header PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_get_cookies_for_url PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_keep_session_cookies PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_malformed_cookies PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_strip_httponly_prefix PASSED [ 7%]\ntest/test_aes.py::TestAES::test_cbc_decrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_cbc_encrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ctr_decrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ctr_encrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_decrypt_text PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ecb_decrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_ecb_encrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_encrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_gcm_decrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_key_expansion PASSED [ 8%]\ntest/test_aes.py::TestAES::test_pad_block PASSED [ 8%]\ntest/test_all_urls.py::TestAllURLsMatching::test_facebook_matching PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_keywords PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_no_duplicated_ie_names PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_no_duplicates PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_pbs PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_soundcloud_not_matching_sets PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_tumblr PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_vimeo_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_channel_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_feeds PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_playlist_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_search_matching PASSED [ 11%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_user_matching PASSED [ 11%]\ntest/test_cache.py::TestCache::test_cache PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_etree_fromstring PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_etree_fromstring_doctype PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_expanduser PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_passthrough PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_unquote PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_unquote_plus PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_urlencode PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_struct_unpack PASSED [ 12%]\ntest/test_config.py::TestConfig::test_config__ENVIRON_DEFAULTS_sanity PASSED [ 12%]\ntest/test_config.py::TestConfig::test_config_all_environ_values PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_default_expected_locations PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_default_grouping PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_early_exit_commandline PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_early_exit_files PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_override_commandline PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_override_files PASSED [ 13%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_derive_key PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_v11 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_mac_derive_key PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_mac_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_windows_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_get_desktop_environment PASSED [ 15%]\ntest/test_cookies.py::TestCookies::test_pbkdf2_sha1 PASSED [ 15%]\ntest/test_cookies.py::TestCookies::test_safari_cookie_parsing PASSED [ 15%]\ntest/test_cookies.py::TestLenientSimpleCookie::test_lenient_parsing PASSED [ 15%]\ntest/test_cookies.py::TestLenientSimpleCookie::test_parsing PASSED [ 15%]\ntest/test_downloader_external.py::TestHttpieFD::test_make_cmd PASSED [ 15%]\ntest/test_downloader_external.py::TestAxelFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestWgetFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestCurlFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestAria2cFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestFFmpegFD::test_make_cmd SKIPPED [ 16%]\ntest/test_downloader_http.py::TestHttpFD::test_chunked FAILED [ 16%]\ntest/test_downloader_http.py::TestHttpFD::test_regular FAILED [ 17%]\ntest/test_execution.py::TestExecution::test_cmdline_umlauts PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_import PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_lazy_extractors PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_main_exec PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_module_exec PASSED [ 17%]\ntest/test_jsinterp.py::TestJSInterpreter::test_add PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_array_access PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_assignments PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_basic PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_bitwise_operators_overflow PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_bitwise_operators_typecast PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_builtins PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_calc PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_call PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_catch PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_char_code_at PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_comma PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_comments SKIPPED (Not...) [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_date PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_div PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_empty_return PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_exp PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_finally PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_break PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_continue PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_try PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_if PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_literal_list PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_mod PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_morespace PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_mul PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_negative PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_nested_try PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_null PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_object PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_operators PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_packed SKIPPED (Not i...) [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_parens PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_precedence PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_quotes PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_regex PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_replace SKIPPED (Not ...) [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_return_function PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_strange_chars PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_sub PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_switch PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_switch_default PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_try PASSED [ 25%]\ntest/test_jsinterp.py::TestJSInterpreter::test_undefined PASSED [ 25%]\ntest/test_jsinterp.py::TestJSInterpreter::test_void PASSED [ 25%]\ntest/test_netrc.py::TestNetRc::test_netrc_present PASSED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_verify_cert[Urllib] PASSED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_verify_cert[Requests] SKIPPED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_ssl_error[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_ssl_error[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_percent_encode[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_percent_encode[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_remove_dot_segments[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_remove_dot_segments[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unicode_path_redirection[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unicode_path_redirection[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_raise_http_error[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_raise_http_error[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_response_url[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_response_url[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_request_cookie_header[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_request_cookie_header[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect_loop[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect_loop[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_incompleteread[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_incompleteread[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_cookies[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_cookies[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_headers[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_headers[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_timeout[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_timeout[Requests] SKIPPED [ 30%]\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 691, in process_request_thread\n self.finish_request(request, client_address)\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 361, in finish_request\n self.RequestHandlerClass(request, client_address, self)\ntest/test_networking.py::TestHTTPRequestHandler::test_source_address[Urllib] PASSED [ 30%]\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 755, in __init__\n self.handle()\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/http/server.py\", line 436, in handle\n self.handle_one_request()\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/http/server.py\", line 425, in handle_one_request\n self.wfile.flush() #actually send the response if not already done.\n ^^^^^^^^^^^^^^^^^^\nValueError: I/O operation on closed file.\n----------------------------------------\ntest/test_networking.py::TestHTTPRequestHandler::test_source_address[Requests] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip_trailing_garbage[Urllib] PASSED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip_trailing_garbage[Requests] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_brotli[Urllib] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_brotli[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_deflate[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_deflate[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_multiple_encodings[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_multiple_encodings[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unsupported_encoding[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unsupported_encoding[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_read[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_read[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_noproxy[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_noproxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_allproxy[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_allproxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy_with_idn[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy_with_idn[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_nopass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_nopass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_nopass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_nopass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_pass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_pass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_pass[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_pass[Requests] SKIPPED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_file_urls[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_http_error_returns_content[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_verify_cert_error_text[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req0-method can't contain control characters-<lambda>-Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req1-URL can't contain control characters-<lambda>-Urllib] PASSED [ 36%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req2-Invalid header name-None-Urllib] PASSED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError0] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError1] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError2] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError3] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-ProxyError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-CertificateVerifyError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-SSLError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError0] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError1] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError4] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError2] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-SSLError-None] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None0] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None1] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None2] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None3] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None4] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-IncompleteRead-3 bytes read, 4 more expected] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-IncompleteRead-3 bytes read, 5 more expected] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-http-False-handler_kwargs0] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-https-False-handler_kwargs1] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-data-False-handler_kwargs2] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-ftp-False-handler_kwargs3] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-file-UnsupportedRequest-handler_kwargs4] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-file-False-handler_kwargs5] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Requests-http-False-handler_kwargs6] SKIPPED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Requests-https-False-handler_kwargs7] SKIPPED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[NoCheckRH-http-False-handler_kwargs8] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[ValidationRH-http-UnsupportedRequest-handler_kwargs9] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_no_proxy[Urllib-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_no_proxy[Requests-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Urllib-all-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Urllib-unrelated-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Requests-all-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Requests-unrelated-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[NoCheckRH-all-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[HTTPSupportedRH-all-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[HTTPSupportedRH-no-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-http-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-https-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks4-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks4a-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks5-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks5h-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks-UnsupportedRequest] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-http-False] SKIPPED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-https-False] SKIPPED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks4-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks4a-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks5-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks5h-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[NoCheckRH-http-False] PASSED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[HTTPSupportedRH-http-UnsupportedRequest] PASSED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[Urllib] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[HTTPSupportedRH] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[Requests] SKIPPED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-//example.com] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-example.com] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-127.0.0.1] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-/a/b/c] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-//example.com] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-example.com] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-127.0.0.1] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-/a/b/c] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions0-AssertionError] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions1-False] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions2-AssertionError] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions3-False] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions4-AssertionError] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions5-UnsupportedRequest] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions6-AssertionError] SKIPPED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions7-False] SKIPPED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions8-False] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions9-AssertionError] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions10-UnsupportedRequest] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[NoCheckRH-extensions11-False] PASSED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[NoCheckRH-extensions12-False] PASSED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_request_type PASSED [ 48%]\ntest/test_networking.py::TestRequestDirector::test_handler_operations PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_send PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_unsupported_handlers PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_unexpected_error PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_preference PASSED [ 49%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_opener PASSED [ 49%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[http://127.0.0.1:8080-expected0] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[-expected1] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[None-expected2] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_request PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_extract_basic_auth PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_sanitize_url PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_file_urls_error PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_legacy_server_connect_error PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-__noproxy__-None] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[no-127.0.0.1,foo.bar-127.0.0.1,foo.bar] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-example.com-http://example.com] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-//example.com-http://example.com] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-socks5://example.com-socks5h://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-socks://example.com-socks4://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-socks4://example.com-socks4://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[unrelated-/bad/proxy-/bad/proxy] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy_header PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_header PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_build_handler_params PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params0] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params1] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params2] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params3] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_urllib_file_urls PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_opt_prefer_urllib PASSED [ 53%]\ntest/test_networking.py::TestRequest::test_query PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_method PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_request_helpers PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_headers PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_data_type PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_content_length_header PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_content_type_header PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_update_req PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_proxies PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_extensions PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_copy PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_url PASSED [ 55%]\ntest/test_networking.py::TestResponse::test_reason[custom-200-custom] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[None-404-Not Found] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[-403-Forbidden] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[None-999-None] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_headers PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_get_header PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_compat PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_select_proxy PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[socks5h://example.com-expected0] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[***example.com:5555-expected1] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[***127.0.0.1:1080-expected2] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[socks4a://:pa%[email protected]] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_unknown PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_load_certifi SKIPPED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[GET-303-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-303-HEAD] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[PUT-303-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-301-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-301-HEAD] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-302-GET] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-302-HEAD] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[PUT-302-PUT] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-308-POST] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-307-POST] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-308-HEAD] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-307-HEAD] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers0-supported_encodings0-expected0] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers1-supported_encodings1-expected1] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers2-supported_encodings2-expected2] PASSED [ 60%]\ntest/test_networking_utils.py::TestInstanceStoreMixin::test_mixin PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_http_error[HTTPError] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_http_error[<lambda>] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_redirect_http_error[HTTPError] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_redirect_http_error[<lambda>] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_compat_http_error PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_compat_http_error_autoclose PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_incomplete_read_error PASSED [ 61%]\ntest/test_plugins.py::TestPlugins::test_directories_containing_plugins PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_extractor_classes PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_importing_zipped_module PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_postprocessor_classes PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_field_to_template PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_format_to_regex PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_metadatafromfield PASSED [ 63%]\ntest/test_postprocessors.py::TestConvertThumbnail::test_escaping PASSED [ 63%]\ntest/test_postprocessors.py::TestExec::test_parse_cmd PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_CommonCase PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_NoZeroDurationChunkAtVideoEnd PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_NoZeroDurationChunkAtVideoStart PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_QuotesAtEnd PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_QuotesAtStart PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_RunsOfQuotes PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CanGetThroughUnaltered PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithAdjacentCuts PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithAdjacentSponsors PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithCutHidingSponsor PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithOverlappingCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithOverlappingSponsors PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingSponsors PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingSponsorsAndCuts PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsorCutInTheMiddle PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsors PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsorsAndCuts PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChaptersAfterLastCut PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChaptersAfterLastSponsor PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutCoincidesWithChapters PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutEndsAtChapterEnd PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutOverlapsMultipleChapters PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutStartsAtChapterStart PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsAtVideoBoundaries PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsOverlapChaptersAtVideoBoundaries PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsWithinSomeChaptersAndOverlappingOthers PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_EverythingCut PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_EverythingSponsored PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_OverlappingSponsorsDifferentTitlesAfterCut PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SingleTinyChapterIsPreserved PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SmallestSponsorInTheOverlapGetsNamed PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorBlockChapters PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorCoincidesWithChapters PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorEndsAtChapterEnd PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorOverlapsMultipleChapters PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorStartsAtChapterStart PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsAtVideoBoundaries PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsNoLongerOverlapAfterCut PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsOverlapChaptersAtVideoBoundaries PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsStillOverlapAfterCut PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsWithinSomeChaptersAndOverlappingOthers PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChapterAtTheStartPrependedToTheNext PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersInTheOriginalArePreserved PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersResultingFromCutsAreIgnored PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersResultingFromSponsorOverlapAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsOverlapsAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsPrependedToTheNextSponsor PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_UniqueNamesForOverlappingSponsors PASSED [ 71%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_no_auth[Urllib-http] PASSED [ 71%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_no_auth[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_auth[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_auth[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_ipv4_target[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_ipv4_target[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_domain_target[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_domain_target[Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv4_client_source_address[Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv4_client_source_address[Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[91-Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[91-Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[92-Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[92-Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[93-Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[93-Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv6_socks4_proxy[Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv6_socks4_proxy[Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_timeout[Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_timeout[Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_no_auth[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_no_auth[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_user_pass[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_user_pass[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv4_target[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv4_target[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_domain_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_domain_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_domain_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_domain_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_ip_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_ip_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv6_destination[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv6_destination[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv6_socks5_proxy[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv6_socks5_proxy[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv4_client_source_address[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv4_client_source_address[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[1-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[1-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[2-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[2-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[3-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[3-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[4-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[4-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[5-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[5-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[6-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[6-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[7-Urllib-http] PASSED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[7-Requests-http] SKIPPED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[8-Urllib-http] PASSED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[8-Requests-http] SKIPPED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_timeout[Urllib-http] PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_LazyList PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_LazyList_laziness PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_Popen_windows_escaping SKIPPED (O...) [ 81%]\ntest/test_utils.py::TestUtil::test_age_restricted PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_args_to_str PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_base_url PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_caesar PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_clean_html PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_clean_podcast_url PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_bool_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_valueless_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_date_from_str PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_daterange PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_datetime_from_str PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_detect_exe_version PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_determine_ext PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_determine_file_encoding PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_dfxp2srt PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_dict_get PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_encode_base_n PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_encode_compat_str PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_escape_rfc3986 PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_expand_path PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_extract_attributes PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_extract_basic_auth PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_find_xpath_attr PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_fix_xml_ampersands PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_float_or_none PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_format_bytes PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_compatible_ext PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_element_by_attribute PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_element_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_html_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_html_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_text_and_html_by_tag PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_html_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_html_by_class PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_get_elements_text_and_html_by_attribute PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_hide_login_info PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_http_header_dict PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_int_or_none PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_intlist_to_bytes PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_iri_to_uri PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_is_html PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_common_constructors PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_edgecases PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_malformed PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_realworld PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_template_literal PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_js_to_json_vars_strings PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_limit_length PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_locked_file PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_lowercase_escape PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_match_str PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_merge_dicts PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_mimetype2ext PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_month_by_name PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_multipart_encode PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_normalize_url PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_ohdave_rsa_encrypt PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_ordered_set PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_paged_list PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_age_limit PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_bitrate PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_codecs PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_count PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_dfxp_time_expr PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_duration PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_filesize PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_iso8601 PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_resolution PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_pkcs1pad PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_prepend_extension PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_read_batch_urls PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_dot_segments PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_end PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_quotes PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_start PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_render_table PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_replace_extension PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_rot47 PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_filename PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_filename_restricted PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_ids PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_path PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_url PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_shell_quote PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_smuggle_url PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_str_to_int PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_strip_jsonp PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_strip_or_none PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_subtitles_filename PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_timeconvert PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_traverse_obj PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_try_call PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unescape_html PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unified_dates PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unified_timestamps PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_update_url_query PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_uppercase_escape PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_url_basename PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_url_or_none PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_urlencode_postdata PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_urljoin PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_urshift PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_variadic PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_version_tuple PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_attr PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_element PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_text PASSED [ 99%]\ntest/test_utils.py::TestUtil::test_xpath_with_ns PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_arg PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_eq PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_shortarg PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_shortarg_eq PASSED [ 99%]\ntest/test_youtube_misc.py::TestYoutubeMisc::test_youtube_extract PASSED [100%]\n\n=================================== FAILURES ===================================\n___________________________ TestHttpFD.test_chunked ____________________________\n\nself = <test.test_downloader_http.TestHttpFD testMethod=test_chunked>\n\n def test_chunked(self):\n> self.download_all({\n 'http_chunk_size': 1000,\n })\n\ntest/test_downloader_http.py:100: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \ntest/test_downloader_http.py:94: in download_all\n self.download(params, ep)\ntest/test_downloader_http.py:86: in download\n self.assertTrue(downloader.real_download(filename, {\nyt_dlp/downloader/http.py:365: in real_download\n return download()\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\n def download():\n data_len = ctx.data.headers.get('Content-length')\n \n if ctx.data.headers.get('Content-encoding'):\n # Content-encoding is present, Content-length is not reliable anymore as we are\n # doing auto decompression. (See: https://github.com/yt-dlp/yt-dlp/pull/6176)\n data_len = None\n \n # Range HTTP header may be ignored/unsupported by a webserver\n # (e.g. extractor/scivee.py, extractor/bambuser.py).\n # However, for a test we still would like to download just a piece of a file.\n # To achieve this we limit data_len to _TEST_FILE_SIZE and manually control\n # block size when downloading a file.\n if is_test and (data_len is None or int(data_len) > self._TEST_FILE_SIZE):\n data_len = self._TEST_FILE_SIZE\n \n if data_len is not None:\n data_len = int(data_len) + ctx.resume_len\n min_data_len = self.params.get('min_filesize')\n max_data_len = self.params.get('max_filesize')\n if min_data_len is not None and data_len < min_data_len:\n self.to_screen(\n f'\\r[download] File is smaller than min-filesize ({data_len} bytes < {min_data_len} bytes). Aborting.')\n return False\n if max_data_len is not None and data_len > max_data_len:\n self.to_screen(\n f'\\r[download] File is larger than max-filesize ({data_len} bytes > {max_data_len} bytes). Aborting.')\n return False\n \n byte_counter = 0 + ctx.resume_len\n block_size = ctx.block_size\n start = time.time()\n \n # measure time over whole while-loop, so slow_down() and best_block_size() work together properly\n now = None # needed for slow_down() in the first loop run\n before = start # start measuring\n \n def retry(e):\n close_stream()\n ctx.resume_len = (byte_counter if ctx.tmpfilename == '-'\n else os.path.getsize(encodeFilename(ctx.tmpfilename)))\n raise RetryDownload(e)\n \n while True:\n try:\n # Download and write\n data_block = ctx.data.read(block_size if not is_test else min(block_size, data_len - byte_counter))\n except TransportError as err:\n retry(err)\n \n byte_counter += len(data_block)\n \n # exit loop when download is finished\n if len(data_block) == 0:\n break\n \n # Open destination file just in time\n if ctx.stream is None:\n try:\n ctx.stream, ctx.tmpfilename = self.sanitize_open(\n ctx.tmpfilename, ctx.open_mode)\n assert ctx.stream is not None\n ctx.filename = self.undo_temp_name(ctx.tmpfilename)\n self.report_destination(ctx.filename)\n except OSError as err:\n self.report_error('unable to open for writing: %s' % str(err))\n return False\n \n if self.params.get('xattr_set_filesize', False) and data_len is not None:\n try:\n write_xattr(ctx.tmpfilename, 'user.ytdl.filesize', str(data_len).encode())\n except (XAttrUnavailableError, XAttrMetadataError) as err:\n self.report_error('unable to set filesize xattr: %s' % str(err))\n \n try:\n ctx.stream.write(data_block)\n except OSError as err:\n self.to_stderr('\\n')\n self.report_error('unable to write data: %s' % str(err))\n return False\n \n # Apply rate limit\n self.slow_down(start, now, byte_counter - ctx.resume_len)\n \n # end measuring of one loop run\n now = time.time()\n after = now\n \n # Adjust block size\n if not self.params.get('noresizebuffer', False):\n block_size = self.best_block_size(after - before, len(data_block))\n \n before = after\n \n # Progress message\n> speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\nE AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n\nyt_dlp/downloader/http.py:296: AttributeError\n___________________________ TestHttpFD.test_regular ____________________________\n\nself = <test.test_downloader_http.TestHttpFD testMethod=test_regular>\n\n def test_regular(self):\n> self.download_all({})\n\ntest/test_downloader_http.py:97: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \ntest/test_downloader_http.py:94: in download_all\n self.download(params, ep)\ntest/test_downloader_http.py:86: in download\n self.assertTrue(downloader.real_download(filename, {\nyt_dlp/downloader/http.py:365: in real_download\n return download()\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\n def download():\n data_len = ctx.data.headers.get('Content-length')\n \n if ctx.data.headers.get('Content-encoding'):\n # Content-encoding is present, Content-length is not reliable anymore as we are\n # doing auto decompression. (See: https://github.com/yt-dlp/yt-dlp/pull/6176)\n data_len = None\n \n # Range HTTP header may be ignored/unsupported by a webserver\n # (e.g. extractor/scivee.py, extractor/bambuser.py).\n # However, for a test we still would like to download just a piece of a file.\n # To achieve this we limit data_len to _TEST_FILE_SIZE and manually control\n # block size when downloading a file.\n if is_test and (data_len is None or int(data_len) > self._TEST_FILE_SIZE):\n data_len = self._TEST_FILE_SIZE\n \n if data_len is not None:\n data_len = int(data_len) + ctx.resume_len\n min_data_len = self.params.get('min_filesize')\n max_data_len = self.params.get('max_filesize')\n if min_data_len is not None and data_len < min_data_len:\n self.to_screen(\n f'\\r[download] File is smaller than min-filesize ({data_len} bytes < {min_data_len} bytes). Aborting.')\n return False\n if max_data_len is not None and data_len > max_data_len:\n self.to_screen(\n f'\\r[download] File is larger than max-filesize ({data_len} bytes > {max_data_len} bytes). Aborting.')\n return False\n \n byte_counter = 0 + ctx.resume_len\n block_size = ctx.block_size\n start = time.time()\n \n # measure time over whole while-loop, so slow_down() and best_block_size() work together properly\n now = None # needed for slow_down() in the first loop run\n before = start # start measuring\n \n def retry(e):\n close_stream()\n ctx.resume_len = (byte_counter if ctx.tmpfilename == '-'\n else os.path.getsize(encodeFilename(ctx.tmpfilename)))\n raise RetryDownload(e)\n \n while True:\n try:\n # Download and write\n data_block = ctx.data.read(block_size if not is_test else min(block_size, data_len - byte_counter))\n except TransportError as err:\n retry(err)\n \n byte_counter += len(data_block)\n \n # exit loop when download is finished\n if len(data_block) == 0:\n break\n \n # Open destination file just in time\n if ctx.stream is None:\n try:\n ctx.stream, ctx.tmpfilename = self.sanitize_open(\n ctx.tmpfilename, ctx.open_mode)\n assert ctx.stream is not None\n ctx.filename = self.undo_temp_name(ctx.tmpfilename)\n self.report_destination(ctx.filename)\n except OSError as err:\n self.report_error('unable to open for writing: %s' % str(err))\n return False\n \n if self.params.get('xattr_set_filesize', False) and data_len is not None:\n try:\n write_xattr(ctx.tmpfilename, 'user.ytdl.filesize', str(data_len).encode())\n except (XAttrUnavailableError, XAttrMetadataError) as err:\n self.report_error('unable to set filesize xattr: %s' % str(err))\n \n try:\n ctx.stream.write(data_block)\n except OSError as err:\n self.to_stderr('\\n')\n self.report_error('unable to write data: %s' % str(err))\n return False\n \n # Apply rate limit\n self.slow_down(start, now, byte_counter - ctx.resume_len)\n \n # end measuring of one loop run\n now = time.time()\n after = now\n \n # Adjust block size\n if not self.params.get('noresizebuffer', False):\n block_size = self.best_block_size(after - before, len(data_block))\n \n before = after\n \n # Progress message\n> speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\nE AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n\nyt_dlp/downloader/http.py:296: AttributeError\n=========================== short test summary info ============================\nSKIPPED [1] test/test_downloader_external.py:115: ffmpeg not found\nSKIPPED [1] test/test_jsinterp.py:111: Not implemented\nSKIPPED [1] test/test_jsinterp.py:373: Not implemented\nSKIPPED [1] test/test_jsinterp.py:335: Not implemented\nSKIPPED [1] test/test_networking.py:309: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:320: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:338: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:350: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:365: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:372: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:382: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:393: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:448: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:480: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:486: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:492: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:509: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:525: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:541: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:549: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:555: brotli support is not installed\nSKIPPED [1] test/test_networking.py:566: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:576: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:586: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:597: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:607: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:637: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:663: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:673: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:681: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:719: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:725: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:732: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:739: Requests request handler is not available\nSKIPPED [11] test/test_networking.py:824: Requests request handler is not available\nSKIPPED [8] test/test_networking.py:855: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:996: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:1005: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:1010: Requests request handler is not available\nSKIPPED [6] test/test_networking.py:1019: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:1028: Requests request handler is not available\nSKIPPED [4] test/test_networking.py:1033: Requests request handler is not available\nSKIPPED [5] test/test_networking.py:1038: Requests request handler is not available\nSKIPPED [1] test/test_networking_utils.py:96: certifi is not installed\nSKIPPED [1] test/test_socks.py:266: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:274: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:284: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:292: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:301: Requests request handler is not available\nSKIPPED [3] test/test_socks.py:311: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:323: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:332: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:342: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:350: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:363: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:371: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:379: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:388: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:397: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:405: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:416: Requests request handler is not available\nSKIPPED [8] test/test_socks.py:425: Requests request handler is not available\nSKIPPED [1] test/test_utils.py:2398: Only relevant on Windows\nFAILED test/test_downloader_http.py::TestHttpFD::test_chunked - AttributeError: 'HttpFD' object has no attribute 'calc_speed'\nFAILED test/test_downloader_http.py::TestHttpFD::test_regular - AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n=== 2 failed, 509 passed, 104 skipped, 5705 deselected in 120.03s (0:02:00) ====\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/yt_dlp/downloader/http.py b/yt_dlp/downloader/http.py\nindex f5237443e..a1f95db29 100644\n--- a/yt_dlp/downloader/http.py\n+++ b/yt_dlp/downloader/http.py\n@@ -11,6 +11,7 @@\n )\n from ..utils import (\n ContentTooShortError,\n+ FormatProgressInfos,\n RetryManager,\n ThrottledDownload,\n XAttrMetadataError,\n@@ -293,11 +294,11 @@ def retry(e):\n before = after\n \n # Progress message\n- speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\n+ speed = FormatProgressInfos.calc_speed(start, now, byte_counter - ctx.resume_len)\n if ctx.data_len is None:\n eta = None\n else:\n- eta = self.calc_eta(start, time.time(), ctx.data_len - ctx.resume_len, byte_counter - ctx.resume_len)\n+ eta = FormatProgressInfos.calc_eta(start, time.time(), ctx.data_len - ctx.resume_len, byte_counter - ctx.resume_len)\n \n self._hook_progress({\n 'status': 'downloading',\ndiff --git a/yt_dlp/downloader/rtmp.py b/yt_dlp/downloader/rtmp.py\nindex 0e0952599..9373f3585 100644\n--- a/yt_dlp/downloader/rtmp.py\n+++ b/yt_dlp/downloader/rtmp.py\n@@ -5,6 +5,7 @@\n \n from .common import FileDownloader\n from ..utils import (\n+ FormatProgressInfos,\n Popen,\n check_executable,\n encodeArgument,\n@@ -50,8 +51,8 @@ def run_rtmpdump(args):\n resume_percent = percent\n resume_downloaded_data_len = downloaded_data_len\n time_now = time.time()\n- eta = self.calc_eta(start, time_now, 100 - resume_percent, percent - resume_percent)\n- speed = self.calc_speed(start, time_now, downloaded_data_len - resume_downloaded_data_len)\n+ eta = FormatProgressInfos.calc_eta(start, time_now, 100 - resume_percent, percent - resume_percent)\n+ speed = FormatProgressInfos.calc_speed(start, time_now, downloaded_data_len - resume_downloaded_data_len)\n data_len = None\n if percent > 0:\n data_len = int(downloaded_data_len * 100 / percent)\n", "difficulty": 2, "changed_files": ["yt_dlp/downloader/http.py", "yt_dlp/downloader/rtmp.py"], "commit_link": "https://github.com/yt-dlp/yt-dlp/tree/43dd59c3137df77f5dd22cef4cb7bedfe9f6b12e"}
 
 
data/python/4410203.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 96, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "error_logging_callback", "workflow_name": "Smoketest", "workflow_filename": "smoketest.yaml", "workflow_path": ".github/workflows/smoketest.yaml", "contributor": "bmosaicml", "sha_fail": "4410203c56984c613d23f29a81ecd1b96c57b1ee", "sha_success": "e4eb7eeac2e76c3287bdd650407e4067ad3a7625", "workflow": "name: Smoketest\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n smoketest:\n runs-on: ubuntu-20.04\n timeout-minutes: 10\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .\n python -m pip install pytest==7.2.1 pytest_codeblocks==0.16.1\n - name: Run checks\n run: |\n pytest tests/test_smoketest.py\n", "logs": [{"step_name": "smoketest (3.8)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}, {"step_name": "smoketest (3.9)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}, {"step_name": "smoketest (3.10)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}], "diff": "diff --git a/composer/callbacks/eval_output_logging_callback.py b/composer/callbacks/eval_output_logging_callback.py\nindex 48e4f76d..65cbdebb 100644\n--- a/composer/callbacks/eval_output_logging_callback.py\n+++ b/composer/callbacks/eval_output_logging_callback.py\n@@ -7,7 +7,7 @@ import hashlib\n import os\n import random\n import shutil\n-import time\n+import tempfile\n from typing import Callable, Optional\n \n from torch.utils.data import DataLoader\n@@ -58,6 +58,8 @@ class EvalOutputLogging(Callback):\n self.hash = hashlib.sha256()\n self.destination_file = None\n \n+ # with tempfile.NamedTemporaryFile\n+ # tmp_dir =\n def _write_tables_to_output_dir(self, state: State):\n try:\n import pandas as pd\n@@ -66,16 +68,9 @@ class EvalOutputLogging(Callback):\n conda_package='pandas',\n conda_channel='conda-forge') from e\n # write tmp files\n- self.hash.update((str(time.time()) + str(random.randint(0, 1_000_000))).encode('utf-8'))\n- tmp_dir = os.getcwd() + '/' + self.hash.hexdigest()\n-\n- if not os.path.exists(tmp_dir):\n- with dist.local_rank_zero_download_and_wait(tmp_dir):\n- if dist.get_local_rank() == 0:\n- os.mkdir(tmp_dir)\n \n full_df = pd.DataFrame()\n- file_name = f'eval-outputs-ba{state.timestamp.batch.value}.tsv'\n+ upload_file_name = f'eval-outputs-ba{state.timestamp.batch.value}.tsv'\n \n for benchmark in self.table:\n cols, rows = self.table[benchmark]\n@@ -84,18 +79,14 @@ class EvalOutputLogging(Callback):\n df['benchmark'] = benchmark\n full_df = pd.concat([full_df, df], ignore_index=True)\n \n- with dist.local_rank_zero_download_and_wait(f'{tmp_dir}/{file_name}'):\n- if dist.get_local_rank() == 0:\n- with open(f'{tmp_dir}/{file_name}', 'wb') as f:\n- full_df.to_csv(f, sep='\\t', index=False)\n+ tmp_file = ''\n+ with tempfile.NamedTemporaryFile('wb') as f:\n+ full_df.to_csv(f, sep='\\t', index=False)\n+ tmp_file = f.name\n \n # copy/upload tmp files\n- _write(destination_path=f'{self.output_directory}/{file_name}', src_file=f'{tmp_dir}/{file_name}')\n- os.remove(f'{tmp_dir}/{file_name}')\n- self.destination_file = f'{self.output_directory}/{file_name}'\n-\n- # delete tmp files\n- os.rmdir(tmp_dir)\n+ _write(destination_path=f'{self.output_directory}/{upload_file_name}', src_file=tmp_file)\n+ self.destination_file = f'{self.output_directory}/{upload_file_name}'\n \n def _prep_response_cache(self, state, cache):\n benchmark = state.dataloader_label\ndiff --git a/composer/datasets/in_context_learning_evaluation.py b/composer/datasets/in_context_learning_evaluation.py\nindex e1566509..1a53f75c 100644\n--- a/composer/datasets/in_context_learning_evaluation.py\n+++ b/composer/datasets/in_context_learning_evaluation.py\n@@ -10,7 +10,6 @@ import random\n from typing import TYPE_CHECKING, Any, Dict, List, Optional, Union\n \n import torch\n-import transformers\n from torch.utils.data import DataLoader, Dataset\n from tqdm import tqdm\n \n", "difficulty": 1, "changed_files": ["composer/callbacks/eval_output_logging_callback.py", "composer/datasets/in_context_learning_evaluation.py"], "commit_link": "https://github.com/mosaicml/composer/tree/4410203c56984c613d23f29a81ecd1b96c57b1ee"}
 
 
data/python/44b56e0.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 127, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "main", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "PhilJd", "sha_fail": "44b56e01683771fb4ca583f9ea57c67dcee8e779", "sha_success": "90bb882770fafc8db2fabc7fe58f446e12b55d77", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\ndoc-builder style src/accelerate docs/source --max_len 119 --check_only\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/bin/doc-builder\", line 8, in <module>\n sys.exit(main())\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/doc_builder/commands/doc_builder_cli.py\", line 47, in main\n args.func(args)\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/doc_builder/commands/style.py\", line 28, in style_command\n raise ValueError(f\"{len(changed)} files should be restyled!\")\nValueError: 1 files should be restyled!\nmake: *** [Makefile:17: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/src/accelerate/big_modeling.py b/src/accelerate/big_modeling.py\nindex 5fd2bd5..6a2392a 100644\n--- a/src/accelerate/big_modeling.py\n+++ b/src/accelerate/big_modeling.py\n@@ -73,7 +73,9 @@ def init_empty_weights(include_buffers: bool = None):\n \n Any model created under this context manager has no weights. As such you can't do something like\n `model.to(some_device)` with it. To load weights inside your empty model, see [`load_checkpoint_and_dispatch`].\n- Make sure to overwrite the default device_map param, otherwise dispatch is not called.\n+ Make sure to overwrite the default device_map param for [`load_checkpoint_and_dispatch`], otherwise dispatch is not\n+ called.\n+\n </Tip>\n \"\"\"\n if include_buffers is None:\n@@ -480,7 +482,7 @@ def load_checkpoint_and_dispatch(\n \n To have Accelerate compute the most optimized `device_map` automatically, set `device_map=\"auto\"`. For more\n information about each option see [here](../concept_guides/big_model_inference#designing-a-device-map).\n- Defaults to None, which means `dispatch_model` will not be called.\n+ Defaults to None, which means [`dispatch_model`] will not be called.\n max_memory (`Dict`, *optional*):\n A dictionary device identifier to maximum memory. Will default to the maximum memory available for each GPU\n and the available CPU RAM if unset.\n", "difficulty": 0, "changed_files": ["src/accelerate/big_modeling.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/44b56e01683771fb4ca583f9ea57c67dcee8e779"}
 
 
data/python/4541644.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 41, "repo_owner": "beetbox", "repo_name": "beets", "head_branch": "lb", "workflow_name": "Check code formatting", "workflow_filename": "formatting_check.yml", "workflow_path": ".github/workflows/formatting_check.yml", "contributor": "arsaboo", "sha_fail": "454164496177fd8b9d6aad4f106e68e816becb6c", "sha_success": "0ed65564b30e002d7f5884e32132098542c3c7dd", "workflow": "name: formatting_check\nrun-name: Check code formatting\non: [push, pull_request]\njobs:\n formatting_check:\n runs-on: ubuntu-latest\n steps:\n - name: Install dependencies\n uses: actions/checkout@v3\n\n - name: Run formatting check\n uses: paolorechia/[email protected]\n with:\n tox_env: \"format_check\"\n", "logs": [{"step_name": "formatting_check/3_Run formatting check.txt", "log": "##[group]Run paolorechia/[email protected]\nwith:\n tox_env: format_check\n##[endgroup]\nstdout: Defaulting to user installation because normal site-packages is not writeable\n\nstdout: Collecting tox\n\nstdout: Downloading tox-4.11.4-py3-none-any.whl (153 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 KB 5.2 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting pyproject-api>=1.6.1\n\nstdout: Downloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\n\nstdout: Collecting colorama>=0.4.6\n\nstdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\n\nstdout: Collecting filelock>=3.12.3\n\nstdout: Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\n\nstdout: Requirement already satisfied: tomli>=2.0.1 in /usr/local/lib/python3.10/dist-packages (from tox) (2.0.1)\n\nstdout: Collecting pluggy>=1.3\n\nstdout: Downloading pluggy-1.3.0-py3-none-any.whl (18 kB)\n\nstdout: Collecting chardet>=5.2\n\nstdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 KB 9.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Requirement already satisfied: packaging>=23.1 in /usr/local/lib/python3.10/dist-packages (from tox) (23.2)\n\nstdout: Collecting virtualenv>=20.24.3\n\nstdout: Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 18.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting cachetools>=5.3.1\n\nstdout: Downloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\n\nstdout: Requirement already satisfied: platformdirs>=3.10 in /usr/local/lib/python3.10/dist-packages (from tox) (4.1.0)\n\nstdout: Collecting distlib<1,>=0.3.7\n\nstdout: Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 30.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Installing collected packages: distlib, pyproject-api, pluggy, filelock, colorama, chardet, cachetools, virtualenv, tox\n\nstdout: Successfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\nReceived 6565789 of 6565789 (100.0%), 7.4 MBs/sec\nCache Size: ~6 MB (6565789 B)\n[command]/usr/bin/tar -z -xf /home/runner/work/_temp/2ecbc63f-4679-45d0-8b2e-0acac9db60db/cache.tgz -P -C /home/runner/work/beets/beets\nCache restored successfully\nHit cache on tox\nstdout: format_check: recreate env because python changed virtualenv version='20.24.5'->'20.25.0'\n\nstdout: format_check: remove tox env folder /home/runner/work/beets/beets/.tox/format_check\n\nstdout: format_check: install_deps> python -I -m pip install black isort\n\nstdout: format_check: commands[0]> isort beets beetsplug test --check\n\nstderr: ERROR: /home/runner/work/beets/beets/beetsplug/listenbrainz.py Imports are incorrectly sorted and/or formatted.\n\nstdout: format_check: exit 1 (0.57 seconds) /home/runner/work/beets/beets> isort beets beetsplug test --check pid=1921\n\nstdout: format_check: FAIL code 1 (4.17=setup[3.60]+cmd[0.57] seconds)\n evaluation failed :( (4.22 seconds)\n\nchild process exited with code 1\n##[warning]The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/\n##[error]Tox failed\n"}], "diff": "diff --git a/beetsplug/listenbrainz.py b/beetsplug/listenbrainz.py\nindex bb5020902..a5595642f 100644\n--- a/beetsplug/listenbrainz.py\n+++ b/beetsplug/listenbrainz.py\n@@ -4,13 +4,15 @@ import datetime\n \n import musicbrainzngs\n import requests\n+\n from beets import config, ui\n from beets.plugins import BeetsPlugin\n from beetsplug.lastimport import process_tracks\n \n \n class ListenBrainzPlugin(BeetsPlugin):\n- \"\"\" A Beets plugin for interacting with ListenBrainz.\"\"\"\n+ \"\"\"A Beets plugin for interacting with ListenBrainz.\"\"\"\n+\n data_source = \"ListenBrainz\"\n ROOT = \"http://api.listenbrainz.org/1/\"\n \n@@ -101,7 +103,6 @@ class ListenBrainzPlugin(BeetsPlugin):\n else:\n return None\n \n-\n def get_tracks_from_listens(self, listens):\n \"\"\"Returns a list of tracks from a list of listens.\"\"\"\n tracks = []\n", "difficulty": 0, "changed_files": ["beetsplug/listenbrainz.py"], "commit_link": "https://github.com/beetbox/beets/tree/454164496177fd8b9d6aad4f106e68e816becb6c"}
 
 
data/python/483a1db.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 139, "repo_owner": "burnash", "repo_name": "gspread", "head_branch": "simplify-get-records", "workflow_name": "lint_python", "workflow_filename": "main.yaml", "workflow_path": ".github/workflows/main.yaml", "contributor": "burnash", "sha_fail": "483a1db177fd4c33b68d122640c32034d5dbb8ae", "sha_success": "74980b8fba2d73c8446a2664103c52a9c914a4ca", "workflow": "name: lint_python\non:\n push:\n branches:\n - \"**\" # run all branches\n tags-ignore:\n - \"*\" # ignore all tags, release.yaml will trigger the CI\n pull_request: # run on all pull requests\n\nconcurrency:\n cancel-in-progress: true\n group: group-${{ github.ref_name }}\n\njobs:\n lint_python:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n python: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.x\"]\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v5\n with:\n python-version: ${{ matrix.python }}\n - run: pip install -U pip\n - run: pip install -U bandit mypy pyupgrade safety tox setuptools\n - run: bandit --recursive --skip B105,B110,B311,B605,B607 --exclude ./.tox .\n if: ${{ matrix.python >= '3.8' }}\n - run: tox -e lint\n - run: tox -e py\n - run: shopt -s globstar && pyupgrade --py3-only **/*.py # --py36-plus\n - run: safety check -i 42559 -i 62044 # pip <= 20.1.1, we upgrade it just above\n - run: tox -e build\n - run: tox -e doc\n", "logs": [{"step_name": "lint_python (3.8)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:11.002645 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:38.759397 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.18 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1934\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (11.96=setup[10.78]+cmd[1.18] seconds)\n evaluation failed :( (12.11 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.9)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:12.019448 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:43.591179 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.31 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1903\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (12.96=setup[11.66]+cmd[1.31] seconds)\n evaluation failed :( (13.20 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.10)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:10.826736 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:40.023159 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.43 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1907\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (13.37=setup[11.94]+cmd[1.43] seconds)\n evaluation failed :( (13.62 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.11)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:12.077791 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:39.355057 +0000\n@@ -517,13 +517,14 @@\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.19 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1935\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (12.92=setup[11.72]+cmd[1.19] seconds)\n evaluation failed :( (12.98 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.x)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:11.804783 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:49.027795 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (2.35 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1932\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (15.07=setup[12.72]+cmd[2.35] seconds)\n evaluation failed :( (15.37 seconds)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gspread/worksheet.py b/gspread/worksheet.py\nindex d6268ea..8b4f551 100644\n--- a/gspread/worksheet.py\n+++ b/gspread/worksheet.py\n@@ -519,9 +519,10 @@ class Worksheet:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n", "difficulty": 0, "changed_files": ["gspread/worksheet.py"], "commit_link": "https://github.com/burnash/gspread/tree/483a1db177fd4c33b68d122640c32034d5dbb8ae"}
 
 
data/python/4921af9.json DELETED
The diff for this file is too large to render. See raw diff
 
data/python/4b7e472.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 94, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "composer_lora", "workflow_name": "Code Quality Checks", "workflow_filename": "code-quality.yaml", "workflow_path": ".github/workflows/code-quality.yaml", "contributor": "dakinggg", "sha_fail": "4b7e4724b2f92741fb094421ebfa60be5bfb18d9", "sha_success": "09346e7f3367ae17409d2a153ec02db1869e09e0", "workflow": "name: Code Quality Checks\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n code-quality:\n runs-on: ubuntu-20.04\n timeout-minutes: 10\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n pip_deps:\n - \"[dev]\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .${{ matrix.pip_deps }}\n - name: Run checks\n run: |\n pre-commit run --all-files\n", "logs": [{"step_name": "code-quality (3.8, [dev])/5_Run checks.txt", "log": "##[group]Run pre-commit run --all-files\n\u001b[36;1mpre-commit run --all-files\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n[INFO] Initializing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Initializing environment for https://github.com/google/yapf.\n[INFO] Initializing environment for https://github.com/google/yapf:toml.\n[INFO] Initializing environment for https://github.com/pycqa/isort.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle:toml.\n[INFO] Initializing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/kynan/nbstripout.\n[INFO] Initializing environment for local:[email protected].\n[INFO] Initializing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Installing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/google/yapf.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pycqa/isort.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/kynan/nbstripout.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for local.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\nruff.....................................................................Passed\nyapf.....................................................................Failed\n- hook id: yapf\n- files were modified by this hook\n\nReformatting composer/datasets/imagenet.py\nReformatting tests/loggers/test_cometml_logger.py\nReformatting tests/algorithms/test_gradient_clipping.py\nReformatting tests/fixtures/__init__.py\nReformatting composer/callbacks/runtime_estimator.py\nReformatting composer/__init__.py\nReformatting composer/trainer/mosaic_fsdp.py\nReformatting composer/models/efficientnetb0/__init__.py\nReformatting composer/callbacks/export_for_inference.py\nReformatting composer/algorithms/cutmix/__init__.py\nReformatting composer/profiler/profiler_action.py\nReformatting tests/datasets/test_mnist.py\nReformatting composer/utils/inference.py\nReformatting tests/trainer/test_predict.py\nReformatting composer/algorithms/randaugment/__init__.py\nReformatting tests/test_notebooks.py\nReformatting composer/algorithms/alibi/__init__.py\nReformatting composer/models/tasks/__init__.py\nReformatting composer/utils/object_store/gcs_object_store.py\nReformatting composer/models/unet/unet.py\nReformatting composer/utils/object_store/object_store.py\nReformatting composer/loggers/remote_uploader_downloader.py\nReformatting composer/utils/eval_client/local_eval_client.py\nReformatting composer/devices/__init__.py\nReformatting tests/utils/test_file_helpers.py\nReformatting composer/algorithms/ema/__init__.py\nReformatting tests/algorithms/test_cutmix.py\nReformatting tests/fixtures/autouse_fixtures.py\nReformatting composer/_version.py\nReformatting composer/core/time.py\nReformatting tests/common/models.py\nReformatting composer/utils/device.py\nReformatting tests/test_passes.py\nReformatting tests/datasets/test_synthetic_data.py\nReformatting tests/utils/__init__.py\nReformatting composer/trainer/__init__.py\nReformatting tests/utils/test_module_surgery.py\nReformatting composer/models/unet/__init__.py\nReformatting tests/callbacks/test_inference.py\nReformatting composer/loggers/wandb_logger.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/_bert.py\nReformatting tests/algorithms/test_channels_last.py\nReformatting tests/common/__init__.py\nReformatting tests/test_split_batch.py\nReformatting composer/algorithms/factorize/factorize_core.py\nReformatting composer/utils/string_enum.py\nReformatting composer/profiler/profiler_schedule.py\nReformatting tests/models/test_bert.py\nReformatting tests/algorithms/test_blurpool_layers.py\nReformatting composer/devices/device_gpu.py\nReformatting docs/source/doctest_cleanup.py\nReformatting composer/datasets/utils.py\nReformatting docs/source/tables/utils.py\nReformatting composer/cli/launcher.py\nReformatting tests/utils/test_dist.py\nReformatting composer/loggers/mosaicml_logger.py\nReformatting composer/profiler/system_profiler.py\nReformatting composer/optim/scheduler.py\nReformatting composer/callbacks/threshold_stopper.py\nReformatting tests/models/test_hf_model.py\nReformatting tests/utils/eval_client/__init__.py\nReformatting composer/core/data_spec.py\nReformatting composer/models/timm/model.py\nReformatting composer/utils/module_surgery.py\nReformatting composer/models/efficientnetb0/efficientnets.py\nReformatting composer/datasets/in_context_learning_evaluation.py\nReformatting tests/callbacks/test_optimizer_monitor.py\nReformatting composer/callbacks/health_checker.py\nReformatting composer/core/__init__.py\nReformatting composer/loggers/progress_bar_logger.py\nReformatting composer/models/gpt2/__init__.py\nReformatting composer/algorithms/gated_linear_units/__init__.py\nReformatting composer/models/bert/__init__.py\nReformatting tests/algorithms/test_factorize_core.py\nReformatting composer/models/resnet_cifar/__init__.py\nReformatting tests/test_precision.py\nReformatting composer/algorithms/randaugment/randaugment.py\nReformatting composer/algorithms/mixup/mixup.py\nReformatting composer/algorithms/stochastic_depth/stochastic_layers.py\nReformatting composer/devices/device_mps.py\nReformatting composer/utils/object_store/mlflow_object_store.py\nReformatting composer/core/callback.py\nReformatting composer/cli/__init__.py\nReformatting composer/utils/eval_client/lambda_eval_client.py\nReformatting composer/models/resnet_cifar/model.py\nReformatting tests/optim/__init__.py\nReformatting composer/utils/auto_log_hparams.py\nReformatting tests/utils/object_store/test_integration_gs_object_store.py\nReformatting composer/algorithms/factorize/factorize.py\nReformatting tests/callbacks/test_mlperf_callback.py\nReformatting composer/callbacks/activation_monitor.py\nReformatting composer/core/engine.py\nReformatting tests/test_engine.py\nReformatting composer/models/classify_mnist/__init__.py\nReformatting tests/utils/eval_client/test_mcli_eval_client.py\nReformatting tests/utils/test_retrying.py\nReformatting tests/utils/object_store/test_uc_object_store.py\nReformatting composer/algorithms/squeeze_excite/squeeze_excite.py\nReformatting tests/profiler/test_json_trace_handler.py\nReformatting composer/models/base.py\nReformatting tests/callbacks/test_callbacks.py\nReformatting tests/models/test_composer_model.py\nReformatting composer/algorithms/warnings.py\nReformatting tests/trainer/test_fsdp_checkpoint.py\nReformatting tests/loggers/test_progress_bar_logger.py\nReformatting composer/metrics/metrics.py\nReformatting composer/loggers/mlflow_logger.py\nReformatting docs/source/conf.py\nReformatting composer/algorithms/mixup/__init__.py\nReformatting composer/models/unet/model.py\nReformatting composer/models/resnet/__init__.py\nReformatting composer/trainer/dist_strategy.py\nReformatting composer/functional/__init__.py\nReformatting composer/utils/misc.py\nReformatting composer/algorithms/cutout/__init__.py\nReformatting tests/datasets/test_ffcv_utils.py\nReformatting tests/algorithms/test_low_precision_groupnorm.py\nReformatting composer/algorithms/cutout/cutout.py\nReformatting tests/loggers/test_mlflow_logger.py\nReformatting tests/common/markers.py\nReformatting tests/trainer/test_checkpoint.py\nReformatting tests/utils/test_autolog_hparams.py\nReformatting tests/test_time.py\nReformatting composer/algorithms/low_precision_layernorm/__init__.py\nReformatting composer/utils/object_store/uc_object_store.py\nReformatting composer/utils/eval_client/__init__.py\nReformatting composer/algorithms/stochastic_depth/stochastic_depth.py\nReformatting tests/utils/test_dynamic_import.py\nReformatting composer/algorithms/layer_freezing/layer_freezing.py\nReformatting tests/utils/object_store/test_gs_object_store.py\nReformatting docs/source/tables/update_alg_tables.py\nReformatting tests/utils/test_batch_helpers.py\nReformatting composer/trainer/_deepspeed.py\nReformatting composer/utils/reproducibility.py\nReformatting composer/loggers/logger_destination.py\nReformatting tests/callbacks/test_threshold_stopper.py\nReformatting tests/algorithms/test_squeeze_excite.py\nReformatting composer/loggers/tensorboard_logger.py\nReformatting tests/trainer/test_trainer_eval.py\nReformatting tests/loggers/test_file_logger.py\nReformatting tests/utils/object_store/test_s3_object_store.py\nReformatting tests/algorithms/test_selective_backprop.py\nReformatting tests/datasets/test_cifar.py\nReformatting composer/algorithms/weight_standardization/weight_standardization.py\nReformatting composer/callbacks/memory_monitor.py\nReformatting composer/utils/batch_helpers.py\nReformatting tests/metrics/test_current_metrics.py\nReformatting composer/core/evaluator.py\nReformatting composer/algorithms/blurpool/blurpool_layers.py\nReformatting composer/core/types.py\nReformatting composer/algorithms/gyro_dropout/__init__.py\nReformatting tests/trainer/test_scale_schedule.py\nReformatting composer/models/deeplabv3/model.py\nReformatting composer/algorithms/gated_linear_units/gated_linear_unit_layers.py\nReformatting composer/algorithms/augmix/__init__.py\nReformatting composer/algorithms/channels_last/channels_last.py\nReformatting composer/models/unet/_layers.py\nReformatting composer/utils/__init__.py\nReformatting composer/core/state.py\nReformatting composer/algorithms/seq_length_warmup/__init__.py\nReformatting docs/source/tables/generate_cost_graphs.py\nReformatting composer/algorithms/label_smoothing/label_smoothing.py\nReformatting tests/utils/object_store/test_libcloud_object_store.py\nReformatting composer/algorithms/alibi/alibi.py\nReformatting composer/algorithms/progressive_resizing/__init__.py\nReformatting tests/models/test_gpt2.py\nReformatting tests/__init__.py\nReformatting tests/loggers/test_mosaicml_logger.py\nReformatting tests/metrics/__init__.py\nReformatting tests/test_smoketest.py\nReformatting composer/datasets/brats.py\nReformatting composer/algorithms/low_precision_groupnorm/low_precision_groupnorm.py\nReformatting tests/trainer/test_ddp_sync_strategy.py\nReformatting tests/trainer/test_trainer.py\nReformatting composer/profiler/torch_profiler.py\nReformatting composer/utils/object_store/sftp_object_store.py\nReformatting composer/algorithms/selective_backprop/__init__.py\nReformatting composer/algorithms/low_precision_layernorm/low_precision_layernorm.py\nReformatting tests/callbacks/test_memory_monitor.py\nReformatting tests/datasets/test_add_dataset_transform.py\nReformatting composer/algorithms/weight_standardization/__init__.py\nReformatting composer/algorithms/ghost_batchnorm/__init__.py\nReformatting composer/loggers/console_logger.py\nReformatting tests/test_docs.py\nReformatting tests/algorithms/test_alibi.py\nReformatting composer/__main__.py\nReformatting docker/pillow_stub/setup.py\nReformatting composer/models/resnet_cifar/resnets.py\nReformatting composer/datasets/c4.py\nReformatting tests/loggers/test_wandb_logger.py\nReformatting composer/algorithms/no_op_model/no_op_model.py\nReformatting tests/algorithms/test_torch_export.py\nReformatting tests/common/state.py\nReformatting composer/models/initializers.py\nReformatting composer/models/gpt2/model.py\nReformatting tests/metrics/test_miou.py\nReformatting tests/metrics/test_map.py\nReformatting docs/source/tables/__init__.py\nReformatting tests/test_state.py\nReformatting composer/loss/loss.py\nReformatting composer/callbacks/generate.py\nReformatting tests/metrics/metric_setter_callback.py\nReformatting composer/algorithms/channels_last/__init__.py\nReformatting composer/utils/retrying.py\nReformatting composer/algorithms/stochastic_depth/__init__.py\nReformatting composer/core/algorithm.py\nReformatting composer/models/vit_small_patch16/__init__.py\nReformatting tests/optim/test_scheduler.py\nReformatting composer/models/efficientnetb0/model.py\nReformatting composer/models/__init__.py\nReformatting composer/algorithms/factorize/__init__.py\nReformatting composer/callbacks/lr_monitor.py\nReformatting composer/loggers/__init__.py\nReformatting composer/callbacks/__init__.py\nReformatting composer/callbacks/image_visualizer.py\nReformatting composer/utils/eval_client/mosaicml_lambda_eval_client.py\nReformatting tests/test_simple_nlp.py\nReformatting tests/utils/eval_client/test_local_eval_client.py\nReformatting composer/metrics/nlp.py\nReformatting tests/trainer/test_fsdp_param_groups.py\nReformatting tests/algorithms/test_stochastic_depth.py\nReformatting composer/utils/fx_utils.py\nReformatting composer/algorithms/squeeze_excite/__init__.py\nReformatting examples/profiler_demo.py\nReformatting tests/algorithms/test_progressive_resizing.py\nReformatting tests/callbacks/test_runtime_estimator.py\nReformatting composer/loss/__init__.py\nReformatting tests/common/events.py\nReformatting composer/algorithms/gyro_dropout/gyro_dropout.py\nReformatting tests/callbacks/test_nan_monitor.py\nReformatting composer/models/timm/__init__.py\nReformatting composer/algorithms/utils/augmentation_common.py\nReformatting composer/profiler/__init__.py\nReformatting composer/utils/eval_client/eval_client.py\nReformatting composer/utils/object_store/s3_object_store.py\nReformatting tests/datasets/__init__.py\nReformatting composer/models/bert/model.py\nReformatting tests/algorithms/test_colout.py\nReformatting composer/utils/dist.py\nReformatting composer/loggers/cometml_logger.py\nReformatting composer/utils/object_store/oci_object_store.py\nReformatting composer/algorithms/no_op_model/__init__.py\nReformatting tests/trainer/__init__.py\nReformatting docs/source/tables/update_model_tables.py\nReformatting composer/utils/import_helpers.py\nReformatting tests/datasets/test_segmentation_transforms.py\nReformatting tests/loggers/__init__.py\nReformatting tests/algorithms/test_ghost_batchnorm.py\nReformatting tests/test_device.py\nReformatting composer/utils/file_helpers.py\nReformatting tests/loggers/test_slack_logger.py\nReformatting composer/devices/device.py\nReformatting composer/algorithms/__init__.py\nReformatting composer/callbacks/speed_monitor.py\nReformatting tests/algorithms/test_required_on_load.py\nReformatting composer/utils/object_store/libcloud_object_store.py\nReformatting tests/datasets/test_in_context_learning_datasets.py\nReformatting tests/callbacks/__init__.py\nReformatting tests/callbacks/test_system_metrics_monitor.py\nReformatting tests/test_cli.py\nReformatting tests/algorithms/test_low_precision_layernorm.py\nReformatting setup.py\nReformatting composer/metrics/__init__.py\nReformatting composer/callbacks/mlperf.py\nReformatting composer/loggers/logger.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/__init__.py\nReformatting tests/callbacks/test_loggers_across_callbacks.py\nReformatting composer/algorithms/swa/swa.py\nReformatting tests/utils/test_iter_helpers.py\nReformatting examples/custom_models.py\nReformatting tests/trainer/test_dataspec.py\nReformatting tests/metrics/test_nlp_metrics.py\nReformatting examples/segmentation/train_deeplabv3_ade20k.py\nReformatting tests/algorithms/test_weight_standardization.py\nReformatting examples/imagenet/train_resnet_imagenet1k.py\nReformatting composer/algorithms/progressive_resizing/progressive_resizing.py\nReformatting docker/generate_build_matrix.py\nReformatting composer/datasets/__init__.py\nReformatting composer/algorithms/cutmix/cutmix.py\nReformatting tests/utils/test_inference.py\nReformatting composer/devices/device_cpu.py\nReformatting composer/core/serializable.py\nReformatting tests/utils/object_store/test_mlflow_object_store.py\nReformatting composer/callbacks/early_stopper.py\nReformatting tests/common/compare.py\nReformatting composer/callbacks/optimizer_monitor.py\nReformatting tests/algorithms/test_factorize_algorithm.py\nReformatting composer/algorithms/sam/__init__.py\nReformatting docs/source/doctest_fixtures.py\nReformatting composer/loss/utils.py\nReformatting composer/profiler/trace_handler.py\nReformatting tests/algorithms/__init__.py\nReformatting tests/callbacks/test_health_checker.py\nReformatting composer/trainer/_scaler.py\nReformatting tests/algorithms/algorithm_settings.py\nReformatting composer/devices/device_tpu.py\nReformatting composer/models/deeplabv3/__init__.py\nReformatting tests/algorithms/test_blurpool.py\nReformatting tests/algorithms/test_layer_freezing.py\nReformatting tests/algorithms/test_augmentations_functional.py\nReformatting composer/algorithms/augmix/augmix.py\nReformatting composer/profiler/json_trace_handler.py\nReformatting tests/fixtures/fixtures.py\nReformatting tests/callbacks/test_early_stopper.py\nReformatting composer/algorithms/blurpool/__init__.py\nReformatting composer/profiler/utils.py\nReformatting composer/algorithms/selective_backprop/selective_backprop.py\nReformatting composer/datasets/lm_dataset.py\nReformatting composer/callbacks/checkpoint_saver.py\nReformatting tests/utils/object_store/object_store_settings.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/utils.py\nReformatting tests/callbacks/test_speed_monitor.py\nReformatting tests/models/test_mmdet_model.py\nReformatting composer/algorithms/gradient_clipping/gradient_clipping.py\nReformatting .github/mcli/mcli_pytest.py\nReformatting tests/algorithms/test_algorithms_train.py\nReformatting composer/datasets/ade20k.py\nReformatting composer/algorithms/colout/colout.py\nReformatting composer/models/mmdetection.py\nReformatting composer/algorithms/ghost_batchnorm/ghost_batchnorm.py\nReformatting composer/loggers/file_logger.py\nReformatting composer/algorithms/label_smoothing/__init__.py\nReformatting tests/datasets/test_dataset_utils.py\nReformatting composer/algorithms/utils/__init__.py\nReformatting tests/algorithms/test_label_smoothing.py\nReformatting composer/datasets/synthetic.py\nReformatting tests/loggers/test_in_memory_logger.py\nReformatting composer/models/efficientnetb0/_layers.py\nReformatting composer/algorithms/utils/augmentation_primitives.py\nReformatting tests/algorithms/test_cutout.py\nReformatting tests/loggers/test_console_logger.py\nReformatting tests/utils/object_store/__init__.py\nReformatting composer/algorithms/gradient_clipping/__init__.py\nReformatting tests/utils/eval_client/test_lambda_eval_client.py\nReformatting tests/callbacks/test_image_visualizer.py\nReformatting tests/algorithms/test_mixup.py\nReformatting composer/utils/object_store/__init__.py\nReformatting composer/profiler/marker.py\nReformatting tests/conftest.py\nReformatting composer/cli/__main__.py\nReformatting tests/utils/object_store/test_oci_object_store.py\nReformatting tests/algorithms/test_gated_linear_units.py\nReformatting composer/callbacks/system_metrics_monitor.py\nReformatting composer/algorithms/blurpool/blurpool.py\nReformatting tests/models/test_efficientnet.py\nReformatting tests/callbacks/test_generate.py\nReformatting tests/utils/test_fx_utils.py\nReformatting composer/loggers/in_memory_logger.py\nReformatting .github/bin/gen_docker_matrix.py\nReformatting tests/algorithms/test_sam.py\nReformatting composer/optim/decoupled_weight_decay.py\nReformatting composer/core/event.py\nReformatting tests/algorithms/test_seq_length_warmup.py\nReformatting composer/algorithms/ema/ema.py\nReformatting tests/utils/object_store/test_object_store.py\nReformatting composer/models/tasks/classification.py\nReformatting composer/algorithms/layer_freezing/__init__.py\nReformatting composer/core/passes.py\nReformatting composer/utils/iter_helpers.py\nReformatting tests/test_events.py\nReformatting examples/checkpoint_with_wandb.py\nReformatting composer/models/resnet/model.py\nReformatting composer/core/precision.py\nReformatting docs/source/tables/update_methods_overview.py\nReformatting tests/callbacks/test_activation_monitor.py\nReformatting tests/trainer/test_ddp.py\nReformatting composer/callbacks/utils.py\nReformatting tests/algorithms/test_ema.py\nReformatting tests/profiler/test_profiler.py\nReformatting composer/algorithms/sam/sam.py\nReformatting tests/trainer/test_fsdp.py\nReformatting composer/models/classify_mnist/model.py\nReformatting composer/algorithms/seq_length_warmup/seq_length_warmup.py\nReformatting composer/datasets/cifar.py\nReformatting composer/trainer/mosaic_fsdp_utils.py\nReformatting tests/callbacks/callback_settings.py\nReformatting composer/trainer/_scale_schedule.py\nReformatting tests/test_full_nlp.py\nReformatting composer/algorithms/factorize/factorize_modules.py\nReformatting composer/models/vit_small_patch16/model.py\nReformatting composer/devices/device_hpu.py\nReformatting examples/gyro_dropout_example.py\nReformatting composer/algorithms/low_precision_groupnorm/__init__.py\nReformatting composer/utils/checkpoint.py\nReformatting tests/common/datasets.py\nReformatting composer/algorithms/swa/__init__.py\nReformatting tests/algorithms/test_algorithm_resumption.py\nReformatting tests/utils/test_string_enum.py\nReformatting tests/algorithms/test_factorized_modules.py\nReformatting tests/algorithms/algorithm_test_template.py\nReformatting composer/datasets/mnist.py\nReformatting composer/models/huggingface.py\nReformatting composer/callbacks/free_outputs.py\nReformatting tests/test_loss.py\nReformatting composer/algorithms/colout/__init__.py\nReformatting composer/metrics/map.py\nReformatting composer/utils/collect_env.py\nReformatting composer/profiler/profiler.py\nReformatting tests/test_docker.py\nReformatting tests/loggers/test_remote_uploader_downloader.py\nReformatting tests/algorithms/test_gyro_dropout.py\nReformatting tests/datasets/test_streaming_datasets_train.py\nReformatting composer/algorithms/gated_linear_units/gated_linear_units.py\nReformatting composer/datasets/ffcv_utils.py\nReformatting composer/callbacks/nan_monitor.py\nReformatting scripts/ffcv/create_ffcv_datasets.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/_gpt2.py\nReformatting tests/profiler/__init__.py\nReformatting composer/loggers/slack_logger.py\nReformatting composer/optim/__init__.py\nReformatting tests/loggers/test_logger.py\nReformatting composer/trainer/meta_safe_apply.py\nReformatting composer/profiler/json_trace_merger.py\nReformatting composer/trainer/trainer.py\n\nisort....................................................................Passed\npydocstyle...............................................................Passed\ncheck for added large files..............................................Passed\ncheck python ast.........................................................Passed\ncheck builtin type constructor use.......................................Passed\ncheck for case conflicts.................................................Passed\ncheck docstring is first.................................................Passed\ncheck that executables have shebangs.....................................Passed\ncheck json...............................................................Passed\ncheck that scripts with shebangs are executable..........................Passed\npretty format json.......................................................Passed\ncheck for merge conflicts................................................Passed\ncheck for broken symlinks................................................Passed\ncheck toml...............................................................Passed\ncheck vcs permalinks.....................................................Passed\ncheck xml................................................................Passed\ncheck yaml...............................................................Passed\ndebug statements (python)................................................Passed\ndetect destroyed symlinks................................................Passed\nfix double quoted strings................................................Passed\nfix end of files.........................................................Passed\nfix utf-8 byte order marker..............................................Passed\nmixed line ending........................................................Passed\ntrim trailing whitespace.................................................Passed\nInsert license in comments...............................................Passed\nnbstripout...............................................................Passed\npyright..................................................................Passed\nsecret scan..............................................................Passed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/composer/models/huggingface.py b/composer/models/huggingface.py\nindex 2498c5f3..d7fb0404 100644\n--- a/composer/models/huggingface.py\n+++ b/composer/models/huggingface.py\n@@ -169,13 +169,13 @@ class HuggingFaceModel(ComposerModel):\n def state_dict(self, *args, **kwargs) -> Dict[str, Any]:\n \"\"\"Returns the state dict of the model.\"\"\"\n full_state_dict = super().state_dict(*args, **kwargs)\n- \n+\n if self.peft_filter_state_dict_trainable:\n- full_state_dict = filter_state_dict_peft(full_state_dict, self.model.peft_config[self.model.active_adapter], False)\n+ full_state_dict = filter_state_dict_peft(full_state_dict, self.model.peft_config[self.model.active_adapter],\n+ False)\n \n return full_state_dict\n \n-\n @staticmethod\n def load_huggingface_tokenizer_from_saved_state(\n hf_state: Dict[str, Any],\n@@ -803,7 +803,10 @@ def write_huggingface_pretrained_from_composer_checkpoint(\n else:\n torch.save(weights_state_dict, Path(output_folder) / 'pytorch_model.bin')\n \n-def filter_state_dict_peft(state_dict: Dict[str, Any], peft_config: 'PeftConfig', remove_adapter_names: bool = True) -> Dict[str, Any]:\n+\n+def filter_state_dict_peft(state_dict: Dict[str, Any],\n+ peft_config: 'PeftConfig',\n+ remove_adapter_names: bool = True) -> Dict[str, Any]:\n # Filtering copied from https://github.com/huggingface/peft/blob/4186c9b104644fd247a4cc0dc2dfc1ede4665204/src/peft/utils/save_and_load.py#L68C1-L86C116\n bias = peft_config.bias\n if bias == 'none':\n@@ -821,7 +824,7 @@ def filter_state_dict_peft(state_dict: Dict[str, Any], peft_config: 'PeftConfig'\n else:\n raise NotImplementedError\n to_return = {k: v for k, v in to_return.items() if (('lora_' in k and 'default' in k) or ('bias' in k))}\n- \n+\n if remove_adapter_names:\n to_return = {k.replace(f'.default', ''): v for k, v in to_return.items()}\n- return to_return\n\\n+ return to_return\ndiff --git a/tests/models/test_hf_model.py b/tests/models/test_hf_model.py\nindex d7b7081d..0cf13fb3 100644\n--- a/tests/models/test_hf_model.py\n+++ b/tests/models/test_hf_model.py\n@@ -1223,6 +1223,7 @@ def test_peft_init_not_installed(tiny_gpt2_model, gpt2_peft_config):\n from composer.models import HuggingFaceModel\n _ = HuggingFaceModel(tiny_gpt2_model, peft_config=gpt2_peft_config)\n \n+\n @pytest.mark.parametrize('just_lora', [True, False])\n def test_peft_trains_and_loads(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path, just_lora):\n pytest.importorskip('peft')\n@@ -1283,6 +1284,7 @@ def test_peft_metadata(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config):\n \n assert loaded_peft_config == gpt2_peft_config\n \n+\n @pytest.mark.parametrize('just_lora', [True, False])\n def test_peft_write_hf_from_composer(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path, just_lora):\n peft = pytest.importorskip('peft')\n@@ -1369,11 +1371,14 @@ def test_peft_fsdp_trains(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config\n for p1, p2 in zip(trainer.state.model.parameters(), load_trainer.state.model.parameters()):\n torch.testing.assert_close(p1, p2)\n \n+\n def test_filtered_state_dict(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path):\n pytest.importorskip('peft')\n \n- hf_model = HuggingFaceModel(tiny_gpt2_model, tokenizer=tiny_gpt2_tokenizer, peft_config=gpt2_peft_config, peft_filter_state_dict_trainable=True)\n+ hf_model = HuggingFaceModel(tiny_gpt2_model,\n+ tokenizer=tiny_gpt2_tokenizer,\n+ peft_config=gpt2_peft_config,\n+ peft_filter_state_dict_trainable=True)\n state_dict = hf_model.state_dict()\n \n assert len(state_dict.keys()) == 4\n-\n", "difficulty": 2, "changed_files": ["composer/models/huggingface.py", "tests/models/test_hf_model.py"], "commit_link": "https://github.com/mosaicml/composer/tree/4b7e4724b2f92741fb094421ebfa60be5bfb18d9"}
 
 
data/python/4d5898b.json DELETED
@@ -1 +0,0 @@
1
- {"language": "Python", "id": 27, "repo_owner": "canonical", "repo_name": "cloud-init", "head_branch": "holman/dhcpcd", "workflow_name": "Lint Tests", "workflow_filename": "check_format.yml", "workflow_path": ".github/workflows/check_format.yml", "contributor": "holmanb", "sha_fail": "4d5898b8a73c93e1ed4434744c2fa7c3f7fbd501", "sha_success": "2bd296ef8983a716a7b8d107571181333f093109", "workflow": "name: Lint Tests\non:\n pull_request:\n push:\n branches:\n - main\n\nconcurrency:\n group: 'ci-${{ github.workflow }}-${{ github.ref }}'\n cancel-in-progress: true\ndefaults:\n run:\n shell: sh -ex {0}\n\njobs:\n check_format:\n strategy:\n fail-fast: false\n matrix:\n env: [ruff, mypy, pylint, black, isort]\n lint-with:\n - {tip-versions: false, os: ubuntu-20.04}\n - {tip-versions: true, os: ubuntu-latest}\n name: Check ${{ matrix.lint-with.tip-versions && 'tip-' || '' }}${{ matrix.env }}\n runs-on: ${{ matrix.lint-with.os }}\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n\n - name: Dependencies\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n\n - name: Print version\n run: python3 --version\n\n - name: Test\n if: ${{ !matrix.lint-with.tip-versions }}\n env:\n # matrix env: not to be confused w/environment variables or testenv\n TOXENV: ${{ matrix.env }}\n run: tox\n - name: Test (tip versions)\n if: matrix.lint-with.tip-versions\n continue-on-error: true\n env:\n TOXENV: tip-${{ matrix.env }}\n run: tox\n schema-format:\n strategy:\n fail-fast: false\n name: Check json format\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Test format\"\n run: |\n tools/check_json_format.sh cloudinit/config/schemas/schema-cloud-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/schema-network-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/versions.schema.cloud-config.json\n\n doc:\n strategy:\n fail-fast: false\n name: Check docs\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n - name: \"Install Python 3.10\"\n uses: actions/setup-python@v4\n with:\n python-version: '3.10.8'\n - name: \"Install dependencies\"\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n - name: \"Spellcheck\"\n run: |\n tox\n env:\n TOXENV: doc-spelling\n - name: \"Build docs\"\n env:\n TOXENV: doc\n run: |\n tox\n\n shell-lint:\n name: Shell Lint\n runs-on: ubuntu-22.04\n steps:\n - name: Checkout\n uses: actions/[email protected]\n\n - name: Install ShellCheck\n run: |\n sudo apt-get update\n sudo apt install shellcheck\n\n - name: Run ShellCheck\n run: |\n shellcheck ./tools/ds-identify\n\n check-cla-signers:\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/[email protected]\n\n - name: Check CLA signers file\n run: tools/check-cla-signers\n", "logs": [{"step_name": "Check pylint/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: pylint\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\npylint create: /home/runner/work/cloud-init/cloud-init/.tox/pylint\npylint installdeps: pylint===2.13.9, -r/home/runner/work/cloud-init/cloud-init/test-requirements.txt, -r/home/runner/work/cloud-init/cloud-init/integration-requirements.txt\npylint inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+88.g0e8fe220.zip\npylint installed: adal==1.2.7,anyio==4.2.0,appdirs==1.4.4,applicationinsights==0.11.10,argcomplete==3.1.6,astroid==2.11.7,attrs==23.2.0,autopage==0.5.2,azure-cli-core==2.56.0,azure-cli-telemetry==1.1.0,azure-common==1.1.28,azure-core==1.29.6,azure-identity==1.15.0,azure-mgmt-compute==30.4.0,azure-mgmt-core==1.4.0,azure-mgmt-network==25.2.0,azure-mgmt-resource==23.0.1,bcrypt==4.1.2,boto3==1.34.17,botocore==1.34.17,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,circuitbreaker==1.4.0,cliff==4.5.0,cloud-init==23.4+88.g0e8fe220,cmd2==2.4.3,configobj==5.0.8,coverage==7.4.0,cryptography==41.0.7,debtcollector==2.5.0,decorator==5.1.1,dill==0.3.7,distro==1.9.0,dogpile.cache==1.3.0,exceptiongroup==1.2.0,google-api-core==2.15.0,google-api-python-client==2.113.0,google-auth==2.26.2,google-auth-httplib2==0.2.0,googleapis-common-protos==1.62.0,httplib2==0.22.0,humanfriendly==10.0,ibm-cloud-sdk-core==3.18.2,ibm-platform-services==0.49.0,ibm-vpc==0.20.0,idna==3.6,importlib-metadata==7.0.1,importlib-resources==6.1.1,iniconfig==2.0.0,iso8601==2.1.0,isodate==0.6.1,isort==5.13.2,Jinja2==3.1.3,jmespath==1.0.1,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,keystoneauth1==5.5.0,knack==0.11.0,lazy-object-proxy==1.10.0,MarkupSafe==2.1.3,mccabe==0.7.0,msal==1.24.0b2,msal-extensions==1.0.0,msgpack==1.0.7,msrest==0.7.1,msrestazure==0.6.4,netaddr==0.10.1,netifaces==0.11.0,oauthlib==3.2.2,oci==2.118.2,openstacksdk==1.4.0,os-service-types==1.7.0,osc-lib==3.0.0,oslo.config==9.3.0,oslo.i18n==6.2.0,oslo.serialization==5.3.0,oslo.utils==6.3.0,packaging==23.2,paramiko==3.4.0,passlib==1.7.4,pbr==6.0.0,pkginfo==1.9.6,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pluggy==1.3.0,portalocker==2.8.2,prettytable==3.9.0,protobuf==3.19.6,psutil==5.9.7,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycloudlib==1!5.12.1,pycparser==2.21,pygments==2.17.2,PyJWT==2.8.0,pylint==2.13.9,PyNaCl==1.5.0,pyOpenSSL==23.3.0,pyparsing==2.4.7,pyperclip==1.8.2,pyserial==3.5,pytest==7.4.4,pytest-cov==4.1.0,pytest-mock==3.12.0,python-cinderclient==9.4.0,python-dateutil==2.8.2,python-keystoneclient==5.3.0,python-novaclient==18.4.0,python-openstackclient==6.4.0,pytz==2023.3.post1,PyYAML==6.0.1,qemu.qmp==0.0.3,referencing==0.32.1,requests==2.31.0,requests-oauthlib==1.3.1,requestsexceptions==1.4.0,responses==0.24.1,rfc3986==2.0.0,rpds-py==0.17.1,rsa==4.9,s3transfer==0.10.0,simplejson==3.19.2,six==1.16.0,sniffio==1.3.0,stevedore==5.1.0,tabulate==0.9.0,toml==0.10.0,tomli==2.0.1,typing-extensions==4.9.0,tzdata==2023.4,uritemplate==4.1.1,urllib3==2.1.0,wcwidth==0.2.13,wrapt==1.16.0,zipp==3.17.0\npylint run-test-pre: PYTHONHASHSEED='1345909868'\npylint run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py\n************* Module cloudinit.net.dhcp\ncloudinit/net/dhcp.py:538: [W0235(useless-super-delegation), Dhcpcd.__init__] Useless super delegation in method '__init__'\n\n------------------------------------\nYour code has been rated at 10.00/10\n\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py (exited with code 4)\n___________________________________ summary ____________________________________\nERROR: pylint: commands failed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/cloudinit/distros/__init__.py b/cloudinit/distros/__init__.py\nindex 5c891f26e..c02166ad2 100644\n--- a/cloudinit/distros/__init__.py\n+++ b/cloudinit/distros/__init__.py\n@@ -170,6 +170,7 @@ class Distro(persistence.CloudInitPickleMixin, metaclass=abc.ABCMeta):\n self._runner = helpers.Runners(paths)\n self.package_managers: List[PackageManager] = []\n self._dhcp_client = None\n+ self._fallback_interface = None\n \n def _unpickle(self, ci_pkl_version: int) -> None:\n \"\"\"Perform deserialization fixes for Distro.\"\"\"\n@@ -1274,6 +1275,13 @@ class Distro(persistence.CloudInitPickleMixin, metaclass=abc.ABCMeta):\n \"/bin/true\",\n ] + ([\"-cf\", config_file, interface] if config_file else [interface])\n \n+ @property\n+ def fallback_interface(self):\n+ \"\"\"Determine the network interface used during local network config.\"\"\"\n+ if self._fallback_interface is None:\n+ self._fallback_interface = net.find_fallback_nic()\n+ return self._fallback_interface\n+\n \n def _apply_hostname_transformations_to_url(url: str, transformations: list):\n \"\"\"\ndiff --git a/cloudinit/net/dhcp.py b/cloudinit/net/dhcp.py\nindex 9c94b3f92..7b2d41df1 100644\n--- a/cloudinit/net/dhcp.py\n+++ b/cloudinit/net/dhcp.py\n@@ -19,8 +19,6 @@ import configobj\n \n from cloudinit import subp, temp_utils, util\n from cloudinit.net import (\n- find_fallback_nic,\n- get_devicelist,\n get_ib_interface_hwaddr,\n get_interface_mac,\n is_ib_interface,\n@@ -98,17 +96,8 @@ def maybe_perform_dhcp_discovery(distro, nic=None, dhcp_log_func=None):\n from the dhclient discovery if run, otherwise an empty list is\n returned.\n \"\"\"\n- if nic is None:\n- nic = find_fallback_nic()\n- if nic is None:\n- LOG.debug(\"Skip dhcp_discovery: Unable to find fallback nic.\")\n- raise NoDHCPLeaseInterfaceError()\n- elif nic not in get_devicelist():\n- LOG.debug(\n- \"Skip dhcp_discovery: nic %s not found in get_devicelist.\", nic\n- )\n- raise NoDHCPLeaseInterfaceError()\n- return distro.dhcp_client.dhcp_discovery(nic, dhcp_log_func, distro)\n+ interface = nic or distro.fallback_interface\n+ return distro.dhcp_client.dhcp_discovery(interface, dhcp_log_func, distro)\n \n \n def networkd_parse_lease(content):\n@@ -185,7 +174,7 @@ class DhcpClient(abc.ABC):\n def parse_static_routes(routes: str) -> List[Tuple[str, str]]:\n return []\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n return {}\n \n \n@@ -218,15 +207,15 @@ class IscDhclient(DhcpClient):\n dhcp_leases.append(dict(lease_options))\n return dhcp_leases\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n \"\"\"Get the most recent lease from the ephemeral phase as a dict.\n \n Return a dict of dhcp options. The dict contains key value\n pairs from the most recent lease.\n \n- @param distro: a distro object - not used in this class, but required\n- for function signature compatibility with other classes\n- that require a distro object\n+ @param interface: an interface string - not used in this class, but\n+ required for function signature compatibility with other classes\n+ that require a distro object\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\n@@ -355,7 +344,7 @@ class IscDhclient(DhcpClient):\n )\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n- lease = self.get_newest_lease(distro)\n+ lease = self.get_newest_lease(interface)\n if lease:\n return lease\n raise InvalidDHCPLeaseFileError()\n@@ -535,9 +524,6 @@ class IscDhclient(DhcpClient):\n class Dhcpcd(DhcpClient):\n client_name = \"dhcpcd\"\n \n- def __init__(self):\n- super().__init__()\n-\n def dhcp_discovery(\n self,\n interface: str,\n@@ -547,12 +533,12 @@ class Dhcpcd(DhcpClient):\n \"\"\"Run dhcpcd on the interface without scripts/filesystem artifacts.\n \n @param interface: Name of the network interface on which to send a\n- dhcp request\n+ dhcp request\n @param dhcp_log_func: A callable accepting the client output and\n- error streams.\n+ error streams.\n @param distro: a distro object for network interface manipulation\n @return: dict of lease options representing the most recent dhcp lease\n- parsed from the dhclient.lease file\n+ parsed from the dhclient.lease file\n \"\"\"\n LOG.debug(\"Performing a dhcp discovery on %s\", interface)\n \n@@ -580,7 +566,7 @@ class Dhcpcd(DhcpClient):\n )\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n- lease = self.get_newest_lease(distro)\n+ lease = self.get_newest_lease(interface)\n if lease:\n return lease\n raise NoDHCPLeaseError(\"No lease found\")\n@@ -644,12 +630,10 @@ class Dhcpcd(DhcpClient):\n # for compatibility we still expect a list of leases\n return lease\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n- \"\"\"Return a lease\n-\n- Return a list of dicts of dhcp options. Each dict contains key value\n- pairs a specific lease in order from oldest to newest.\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n+ \"\"\"Return a dict of dhcp options.\n \n+ @param interface: which interface to dump the lease from\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\n@@ -659,11 +643,11 @@ class Dhcpcd(DhcpClient):\n [\n \"dhcpcd\",\n \"--dumplease\",\n- distro.fallback_interface,\n+ interface,\n ],\n rcs=[0, 1],\n ).stdout,\n- distro.fallback_interface,\n+ interface,\n )\n \n except subp.ProcessExecutionError as error:\n@@ -781,17 +765,17 @@ class Udhcpc(DhcpClient):\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n \n- return self.get_newest_lease(distro)\n+ return self.get_newest_lease(interface)\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n \"\"\"Get the most recent lease from the ephemeral phase as a dict.\n \n Return a dict of dhcp options. The dict contains key value\n pairs from the most recent lease.\n \n- @param distro: a distro object - not used in this class, but required\n- for function signature compatibility with other classes\n- that require a distro object\n+ @param interface: an interface name - not used in this class, but\n+ required for function signature compatibility with other classes\n+ that require a distro object\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\ndiff --git a/cloudinit/sources/__init__.py b/cloudinit/sources/__init__.py\nindex c207b5ed6..f9650268a 100644\n--- a/cloudinit/sources/__init__.py\n+++ b/cloudinit/sources/__init__.py\n@@ -610,13 +610,11 @@ class DataSource(CloudInitPickleMixin, metaclass=abc.ABCMeta):\n @property\n def fallback_interface(self):\n \"\"\"Determine the network interface used during local network config.\"\"\"\n- if self._fallback_interface is None:\n- self._fallback_interface = net.find_fallback_nic()\n- if self._fallback_interface is None:\n- LOG.warning(\n- \"Did not find a fallback interface on %s.\", self.cloud_name\n- )\n- return self._fallback_interface\n+ if self.distro.fallback_interface is None:\n+ LOG.warning(\n+ \"Did not find a fallback interface on %s.\", self.cloud_name\n+ )\n+ return self.distro.fallback_interface\n \n @property\n def platform_type(self):\ndiff --git a/tests/unittests/net/test_dhcp.py b/tests/unittests/net/test_dhcp.py\nindex ffa3eab17..9af5f93aa 100644\n--- a/tests/unittests/net/test_dhcp.py\n+++ b/tests/unittests/net/test_dhcp.py\n@@ -372,20 +372,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n with_logs = True\n ib_address_prefix = \"00:00:00:00:00:00:00:00:00:00:00:00\"\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n- def test_no_fallback_nic_found(self, m_fallback_nic):\n- \"\"\"Log and do nothing when nic is absent and no fallback is found.\"\"\"\n- m_fallback_nic.return_value = None # No fallback nic found\n-\n- with pytest.raises(NoDHCPLeaseInterfaceError):\n- maybe_perform_dhcp_discovery(MockDistro())\n-\n- self.assertIn(\n- \"Skip dhcp_discovery: Unable to find fallback nic.\",\n- self.logs.getvalue(),\n- )\n-\n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=\"eth9\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\", return_value=\"eth9\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n @mock.patch(\"cloudinit.net.dhcp.subp.subp\")\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n@@ -406,7 +393,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n self.logs.getvalue(),\n )\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=\"eth9\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\", return_value=\"eth9\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n @mock.patch(\"cloudinit.net.dhcp.subp.subp\")\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n@@ -435,17 +422,6 @@ class TestDHCPDiscoveryClean(CiTestCase):\n self.logs.getvalue(),\n )\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=None)\n- def test_provided_nic_does_not_exist(self, m_fallback_nic):\n- \"\"\"When the provided nic doesn't exist, log a message and no-op.\"\"\"\n- with pytest.raises(NoDHCPLeaseInterfaceError):\n- maybe_perform_dhcp_discovery(MockDistro(), \"idontexist\")\n-\n- self.assertIn(\n- \"Skip dhcp_discovery: nic idontexist not found in get_devicelist.\",\n- self.logs.getvalue(),\n- )\n-\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n def test_absent_dhclient_command(self, m_fallback, m_which):\ndiff --git a/tests/unittests/sources/test_cloudstack.py b/tests/unittests/sources/test_cloudstack.py\nindex 8ee04b3e1..b33c5e492 100644\n--- a/tests/unittests/sources/test_cloudstack.py\n+++ b/tests/unittests/sources/test_cloudstack.py\n@@ -244,6 +244,13 @@ class TestCloudStackHostname(CiTestCase):\n )\n )\n \n+ self.patches.enter_context(\n+ mock.patch(\n+ \"cloudinit.distros.net.find_fallback_nic\",\n+ return_value=\"eth0\",\n+ )\n+ )\n+\n self.patches.enter_context(\n mock.patch(\n MOD_PATH\n@@ -304,7 +311,7 @@ class TestCloudStackHostname(CiTestCase):\n )\n \n ds = DataSourceCloudStack(\n- {}, MockDistro(), helpers.Paths({\"run_dir\": self.tmp})\n+ {}, ubuntu.Distro(\"\", {}, {}), helpers.Paths({\"run_dir\": self.tmp})\n )\n ds._fallback_interface = \"eth0\"\n with mock.patch(MOD_PATH + \".util.load_file\"):\ndiff --git a/tests/unittests/sources/test_ec2.py b/tests/unittests/sources/test_ec2.py\nindex 5f60ad737..32fed89da 100644\n--- a/tests/unittests/sources/test_ec2.py\n+++ b/tests/unittests/sources/test_ec2.py\n@@ -10,6 +10,7 @@ import requests\n import responses\n \n from cloudinit import helpers\n+from cloudinit.distros import ubuntu\n from cloudinit.sources import DataSourceEc2 as ec2\n from tests.unittests import helpers as test_helpers\n \n@@ -342,9 +343,11 @@ class TestEc2(test_helpers.ResponsesTestCase):\n p.start()\n self.addCleanup(p.stop)\n \n- def _setup_ds(self, sys_cfg, platform_data, md, md_version=None):\n+ def _setup_ds(\n+ self, sys_cfg, platform_data, md, md_version=None, distro=None\n+ ):\n self.uris = []\n- distro = mock.MagicMock()\n+ distro = distro or mock.MagicMock()\n distro.get_tmp_exec_path = self.tmp_dir\n paths = helpers.Paths({\"run_dir\": self.tmp})\n if sys_cfg is None:\n@@ -846,7 +849,7 @@ class TestEc2(test_helpers.ResponsesTestCase):\n \n @mock.patch(\"cloudinit.net.ephemeral.EphemeralIPv6Network\")\n @mock.patch(\"cloudinit.net.ephemeral.EphemeralIPv4Network\")\n- @mock.patch(\"cloudinit.net.find_fallback_nic\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\")\n @mock.patch(\"cloudinit.net.ephemeral.maybe_perform_dhcp_discovery\")\n @mock.patch(\"cloudinit.sources.DataSourceEc2.util.is_FreeBSD\")\n def test_ec2_local_performs_dhcp_on_non_bsd(\n@@ -873,6 +876,7 @@ class TestEc2(test_helpers.ResponsesTestCase):\n platform_data=self.valid_platform_data,\n sys_cfg={\"datasource\": {\"Ec2\": {\"strict_id\": False}}},\n md={\"md\": DEFAULT_METADATA},\n+ distro=ubuntu.Distro(\"\", {}, {}),\n )\n \n ret = ds.get_data()\ndiff --git a/tests/unittests/sources/test_init.py b/tests/unittests/sources/test_init.py\nindex 44d63b816..b3764945d 100644\n--- a/tests/unittests/sources/test_init.py\n+++ b/tests/unittests/sources/test_init.py\n@@ -6,6 +6,7 @@ import os\n import stat\n \n from cloudinit import importer, util\n+from cloudinit.distros import ubuntu\n from cloudinit.event import EventScope, EventType\n from cloudinit.helpers import Paths\n from cloudinit.sources import (\n@@ -73,7 +74,7 @@ class TestDataSource(CiTestCase):\n def setUp(self):\n super(TestDataSource, self).setUp()\n self.sys_cfg = {\"datasource\": {\"_undef\": {\"key1\": False}}}\n- self.distro = \"distrotest\" # generally should be a Distro object\n+ self.distro = ubuntu.Distro(\"\", {}, {})\n self.paths = Paths({})\n self.datasource = DataSource(self.sys_cfg, self.distro, self.paths)\n \n@@ -201,7 +202,7 @@ class TestDataSource(CiTestCase):\n for log in expected_logs:\n self.assertIn(log, logs)\n \n- @mock.patch(\"cloudinit.sources.net.find_fallback_nic\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\")\n def test_fallback_interface_is_discovered(self, m_get_fallback_nic):\n \"\"\"The fallback_interface is discovered via find_fallback_nic.\"\"\"\n m_get_fallback_nic.return_value = \"nic9\"\n@@ -221,7 +222,7 @@ class TestDataSource(CiTestCase):\n @mock.patch(\"cloudinit.sources.net.find_fallback_nic\")\n def test_wb_fallback_interface_is_cached(self, m_get_fallback_nic):\n \"\"\"The fallback_interface is cached and won't be rediscovered.\"\"\"\n- self.datasource._fallback_interface = \"nic10\"\n+ self.datasource.distro._fallback_interface = \"nic10\"\n self.assertEqual(\"nic10\", self.datasource.fallback_interface)\n m_get_fallback_nic.assert_not_called()\n \ndiff --git a/tests/unittests/sources/test_openstack.py b/tests/unittests/sources/test_openstack.py\nindex 127123cff..97cc8c94e 100644\n--- a/tests/unittests/sources/test_openstack.py\n+++ b/tests/unittests/sources/test_openstack.py\n@@ -338,7 +338,7 @@ class TestOpenStackDataSource(test_helpers.ResponsesTestCase):\n ds_os_local = ds.DataSourceOpenStackLocal(\n settings.CFG_BUILTIN, distro, helpers.Paths({\"run_dir\": self.tmp})\n )\n- ds_os_local._fallback_interface = \"eth9\" # Monkey patch for dhcp\n+ distro.fallback_interface = \"eth9\" # Monkey patch for dhcp\n m_dhcp.return_value = {\n \"interface\": \"eth9\",\n \"fixed-address\": \"192.168.2.9\",\n", "difficulty": 0, "changed_files": ["cloudinit/distros/__init__.py", "cloudinit/net/dhcp.py", "cloudinit/sources/__init__.py", "tests/unittests/net/test_dhcp.py", "tests/unittests/sources/test_cloudstack.py", "tests/unittests/sources/test_ec2.py", "tests/unittests/sources/test_init.py", "tests/unittests/sources/test_openstack.py"], "commit_link": "https://github.com/canonical/cloud-init/tree/4d5898b8a73c93e1ed4434744c2fa7c3f7fbd501"}