Delete data/python
Browse filesremoved files before adding new filtered
This view is limited to 50 files because it contains too many changes.
See raw diff
- data/python/028ad1e.json +0 -1
- data/python/02d7747.json +0 -0
- data/python/034b195.json +0 -1
- data/python/03669a5.json +0 -1
- data/python/06bc6b6.json +0 -1
- data/python/077f6aa.json +0 -1
- data/python/07bc10c.json +0 -1
- data/python/0828c8d.json +0 -0
- data/python/08c47c4.json +0 -1
- data/python/0b08b8e.json +0 -0
- data/python/0b5c0d2.json +0 -0
- data/python/0b93d2d.json +0 -0
- data/python/0d26cc1.json +0 -1
- data/python/0d367ce.json +0 -0
- data/python/0f71221.json +0 -1
- data/python/102f918.json +0 -1
- data/python/1168b92.json +0 -1
- data/python/124065b.json +0 -1
- data/python/16a0c04.json +0 -1
- data/python/1afe2c9.json +0 -1
- data/python/1b22eb7.json +0 -1
- data/python/1e48195.json +0 -1
- data/python/212ff27.json +0 -1
- data/python/2201be2.json +0 -1
- data/python/2327ca6.json +0 -0
- data/python/2a104bf.json +0 -1
- data/python/2a59b55.json +0 -1
- data/python/2ab9e84.json +0 -1
- data/python/2c06ffa.json +0 -1
- data/python/2ccb7d3.json +0 -1
- data/python/2e41e78.json +0 -0
- data/python/2f0605c.json +0 -1
- data/python/3054659.json +0 -0
- data/python/3137ef6.json +0 -1
- data/python/31c3e1d.json +0 -1
- data/python/385c14d.json +0 -1
- data/python/386bf6f.json +0 -1
- data/python/38a3024.json +0 -1
- data/python/3dd8e44.json +0 -1
- data/python/3ed7a88.json +0 -0
- data/python/3fcfe29.json +0 -1
- data/python/434321a.json +0 -0
- data/python/43dd59c.json +0 -1
- data/python/4410203.json +0 -1
- data/python/44b56e0.json +0 -1
- data/python/4541644.json +0 -1
- data/python/483a1db.json +0 -1
- data/python/4921af9.json +0 -0
- data/python/4b7e472.json +0 -1
- data/python/4d5898b.json +0 -1
data/python/028ad1e.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 129, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "AjayP13-patch-1", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "AjayP13", "sha_fail": "028ad1efee2c41691d78e5a4de90ebd6f8236cad", "sha_success": "9ff59024aee19e24948401eb8cc7057602592b49", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\nsrc/accelerate/utils/__init__.py:1:1: I001 [*] Import block is un-sorted or un-formatted\nsrc/accelerate/utils/modeling.py:15:1: I001 [*] Import block is un-sorted or un-formatted\nFound 2 errors.\n[*] 2 fixable with the `--fix` option.\nmake: *** [Makefile:16: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/src/accelerate/utils/__init__.py b/src/accelerate/utils/__init__.py\nindex ddf794a..7179afb 100644\n--- a/src/accelerate/utils/__init__.py\n+++ b/src/accelerate/utils/__init__.py\n@@ -59,7 +59,6 @@ from .imports import (\n is_comet_ml_available,\n is_cuda_available,\n is_datasets_available,\n- is_peft_available,\n is_deepspeed_available,\n is_dvclive_available,\n is_fp8_available,\n@@ -70,6 +69,7 @@ from .imports import (\n is_msamp_available,\n is_npu_available,\n is_pandas_available,\n+ is_peft_available,\n is_rich_available,\n is_sagemaker_available,\n is_tensorboard_available,\n@@ -81,7 +81,6 @@ from .imports import (\n is_xpu_available,\n )\n from .modeling import (\n- is_peft_model,\n calculate_maximum_sizes,\n check_device_map,\n check_tied_parameters_in_config,\n@@ -96,6 +95,7 @@ from .modeling import (\n get_mixed_precision_context_manager,\n id_tensor_storage,\n infer_auto_device_map,\n+ is_peft_model,\n load_checkpoint_in_model,\n load_offloaded_weights,\n load_state_dict,\ndiff --git a/src/accelerate/utils/modeling.py b/src/accelerate/utils/modeling.py\nindex 802b13c..03d3a39 100644\n--- a/src/accelerate/utils/modeling.py\n+++ b/src/accelerate/utils/modeling.py\n@@ -30,7 +30,7 @@ import torch.nn as nn\n from ..state import AcceleratorState\n from .constants import SAFE_WEIGHTS_NAME, WEIGHTS_NAME\n from .dataclasses import AutocastKwargs, CustomDtype, DistributedType\n-from .imports import is_mps_available, is_npu_available, is_xpu_available, is_peft_available\n+from .imports import is_mps_available, is_npu_available, is_peft_available, is_xpu_available\n from .offload import load_offloaded_weight, offload_weight, save_offload_index\n from .tqdm import is_tqdm_available, tqdm\n \n", "difficulty": 0, "changed_files": ["src/accelerate/utils/__init__.py", "src/accelerate/utils/modeling.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/028ad1efee2c41691d78e5a4de90ebd6f8236cad"}
|
|
|
|
data/python/02d7747.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/034b195.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 157, "repo_owner": "modin-project", "repo_name": "modin", "head_branch": "merge_without_to_pandas", "workflow_name": "fuzzy", "workflow_filename": "fuzzydata-test.yml", "workflow_path": ".github/workflows/fuzzydata-test.yml", "contributor": "arunjose696", "sha_fail": "034b1956a0914b739c3e5fc2e19ffcfee0f08c02", "sha_success": "ac24182c5553424d2d00a7e573b5aa289c79d955", "workflow": "name: fuzzy\non:\n pull_request:\n paths:\n # NOTE: keep these paths in sync with the paths that trigger the CI Github\n # Actions in .github/workflows/ci.yml\n - .github/workflows/**\n - '!.github/workflows/push-to-master.yml'\n - asv_bench/**\n - modin/**\n - requirements/**\n - scripts/**\n - environment-dev.yml\n - requirements-dev.txt\n - setup.cfg\n - setup.py\n - versioneer.py\nconcurrency:\n # Cancel other jobs in the same branch. We don't care whether CI passes\n # on old commits.\n group: ${{ github.workflow }}-${{ github.ref }}\n cancel-in-progress: ${{ startsWith(github.ref, 'refs/pull/') }}\nenv:\n MODIN_GITHUB_CI: true\njobs:\n test-fuzzydata:\n runs-on: ubuntu-latest\n defaults:\n run:\n shell: bash -l {0}\n strategy:\n matrix:\n python-version: [\"3.9\"]\n engine: [\"ray\", \"dask\"]\n steps:\n - uses: actions/checkout@v3\n - uses: ./.github/actions/mamba-env\n with:\n environment-file: environment-dev.yml\n python-version: ${{matrix.python-version}}\n - name: test-fuzzydata (engine ${{matrix.engine}}, python ${{matrix.python-version}})\n run: python -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-${{matrix.engine}}/run.log --log-file-level=INFO\n env:\n MODIN_ENGINE: ${{matrix.engine}}\n - uses: actions/upload-artifact@v3\n if: success() || failure()\n with:\n name: fuzzydata-test-workflow-${{matrix.engine}}\n path: /tmp/fuzzydata-test-wf-${{matrix.engine}}/* # Must match output dir in test_fuzzydata.py\n if-no-files-found: error\n", "logs": [{"step_name": "test-fuzzydata (3.9, ray)/4_test-fuzzydata (engine ray, python 3.9).txt", "log": "##[group]Run python -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-ray/run.log --log-file-level=INFO\n\u001b[36;1mpython -m pytest modin/experimental/fuzzydata/test/test_fuzzydata.py -Wignore::UserWarning --log-file=/tmp/fuzzydata-test-wf-ray/run.log --log-file-level=INFO\u001b[0m\nshell: /usr/bin/bash -l {0}\nenv:\n MODIN_GITHUB_CI: true\n INPUT_RUN_POST: true\n CONDA: /usr/share/miniconda3\n CONDA_PKGS_DIR: /home/runner/conda_pkgs_dir\n MODIN_ENGINE: ray\n##[endgroup]\n============================= test session starts ==============================\nplatform linux -- Python 3.9.18, pytest-7.4.4, pluggy-1.3.0\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/modin/modin\nconfigfile: setup.cfg\nplugins: benchmark-4.0.0, xdist-3.5.0, anyio-4.2.0, Faker-22.1.0, cov-4.1.0\ncollected 1 item\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py F [100%]\n\n=================================== FAILURES ===================================\n________________________ test_fuzzydata_sample_workflow ________________________\n\n def test_fuzzydata_sample_workflow():\n # Workflow Generation Options\n wf_name = str(uuid.uuid4())[:8] # Unique name for the generated workflow\n num_versions = 10 # Number of unique CSV files to generate\n cols = 33 # Columns in Base Artifact\n rows = 1000 # Rows in Base Artifact\n bfactor = 1.0 # Branching Factor - 0.1 is linear, 10.0 is star-like\n exclude_ops = [\"groupby\"] # In-Memory groupby operations cause issue #4287\n matfreq = 2 # How many operations to chain before materialization\n \n engine = Engine.get().lower()\n \n # Create Output Directory for Workflow Data\n base_out_directory = (\n f\"/tmp/fuzzydata-test-wf-{engine}/\" # Must match corresponding github-action\n )\n if os.path.exists(base_out_directory):\n shutil.rmtree(base_out_directory)\n output_directory = f\"{base_out_directory}/{wf_name}/\"\n os.makedirs(output_directory, exist_ok=True)\n \n # Start Workflow Generation\n> workflow = generate_workflow(\n workflow_class=ModinWorkflow,\n name=wf_name,\n num_versions=num_versions,\n base_shape=(cols, rows),\n out_directory=output_directory,\n bfactor=bfactor,\n exclude_ops=exclude_ops,\n matfreq=matfreq,\n wf_options={\"modin_engine\": engine},\n )\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py:47: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/generator.py:381: in generate_workflow\n raise e\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/generator.py:367: in generate_workflow\n wf.execute_current_operation(next_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/workflow.py:197: in execute_current_operation\n raise e\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/workflow.py:168: in execute_current_operation\n new_artifact = self.current_operation.execute(new_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/core/operation.py:172: in execute\n result = self.materialize(new_label)\n/usr/share/miniconda3/envs/modin/lib/python3.9/site-packages/fuzzydata/clients/pandas.py:114: in materialize\n new_df = eval(self.code)\n<string>:1: in <module>\n ???\nmodin/logging/logger_decorator.py:129: in run_and_log\n return obj(*args, **kwargs)\nmodin/pandas/dataframe.py:1301: in merge\n query_compiler=self._query_compiler.merge(\nmodin/logging/logger_decorator.py:129: in run_and_log\n return obj(*args, **kwargs)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <modin.core.storage_formats.pandas.query_compiler.PandasQueryCompiler object at 0x7f9fccfaedf0>\nright = <modin.core.storage_formats.pandas.query_compiler.PandasQueryCompiler object at 0x7f9fccfae940>\nkwargs = {'copy': True, 'how': 'inner', 'indicator': False, 'left_index': False, ...}\nhow = 'inner', left_index = False, right_index = False, sort = False\nmap_func = <function PandasQueryCompiler.merge.<locals>.map_func at 0x7f9fd4036dc0>\nnew_columns = Index(['urhjs__ipv4_public', '19caY__sha1', 'uqdyN__bothify',\n 'Fpmv9__credit_card_number', 'cTFks__ipv4_private... 'NN7sX__day_of_month', 'Z2GvM__safe_color_name', 'RZGm2__pyfloat',\n 'JW1cO__address'],\n dtype='object')\nnew_dtypes = ModinDtypes:\n\tvalue type: <class 'modin.core.dataframe.pandas.metadata.dtypes.DtypesDescriptor'>;\n\tvalue:\n\tDtypesDescr...graph', '7tHsJ__sentence'];\n\tschema is known: False;\n\thas parent df: False;\n\tcolumns order: None;\n\tknow all names: True\n\n def merge(self, right, **kwargs):\n how = kwargs.get(\"how\", \"inner\")\n on = kwargs.get(\"on\", None)\n left_on = kwargs.get(\"left_on\", None)\n right_on = kwargs.get(\"right_on\", None)\n left_index = kwargs.get(\"left_index\", False)\n right_index = kwargs.get(\"right_index\", False)\n sort = kwargs.get(\"sort\", False)\n \n if how in [\"left\", \"inner\"] and left_index is False and right_index is False:\n kwargs[\"sort\"] = False\n \n def should_keep_index(left, right):\n keep_index = False\n if left_on is not None and right_on is not None:\n keep_index = any(\n o in left.index.names\n and o in right_on\n and o in right.index.names\n for o in left_on\n )\n elif on is not None:\n keep_index = any(\n o in left.index.names and o in right.index.names for o in on\n )\n return keep_index\n \n def map_func(\n left, right, *axis_lengths, kwargs=kwargs, **service_kwargs\n ): # pragma: no cover\n df = pandas.merge(left, right, **kwargs)\n \n if kwargs[\"how\"] == \"left\":\n partition_idx = service_kwargs[\"partition_idx\"]\n if len(axis_lengths):\n if not should_keep_index(left, right):\n # Doesn't work for \"inner\" case, since the partition sizes of the\n # left dataframe may change\n start = sum(axis_lengths[:partition_idx])\n stop = sum(axis_lengths[: partition_idx + 1])\n \n df.index = pandas.RangeIndex(start, stop)\n \n return df\n \n # Want to ensure that these are python lists\n if left_on is not None and right_on is not None:\n left_on = list(left_on) if is_list_like(left_on) else [left_on]\n right_on = list(right_on) if is_list_like(right_on) else [right_on]\n elif on is not None:\n on = list(on) if is_list_like(on) else [on]\n \n new_columns = None\n new_dtypes = None\n if self._modin_frame.has_materialized_columns:\n if left_on is None and right_on is None:\n if on is None:\n on = [c for c in self.columns if c in right.columns]\n _left_on, _right_on = on, on\n else:\n if left_on is None or right_on is None:\n raise MergeError(\n \"Must either pass only 'on' or 'left_on' and 'right_on', not combination of them.\"\n )\n _left_on, _right_on = left_on, right_on\n \n try:\n new_columns, left_renamer, right_renamer = join_columns(\n self.columns,\n right.columns,\n _left_on,\n _right_on,\n kwargs.get(\"suffixes\", (\"_x\", \"_y\")),\n )\n except NotImplementedError:\n # This happens when one of the keys to join is an index level. Pandas behaviour\n # is really complicated in this case, so we're not computing resulted columns for now.\n pass\n else:\n # renamers may contain columns from 'index', so trying to merge index and column dtypes here\n right_index_dtypes = (\n right.index.dtypes\n if isinstance(right.index, pandas.MultiIndex)\n else pandas.Series(\n [right.index.dtype], index=[right.index.name]\n )\n )\n right_dtypes = pandas.concat([right.dtypes, right_index_dtypes])[\n right_renamer.keys()\n ].rename(right_renamer)\n \n left_index_dtypes = (\n self._modin_frame._index_cache.maybe_get_dtypes()\n )\n left_dtypes = (\n ModinDtypes.concat(\n [self._modin_frame._dtypes, left_index_dtypes]\n )\n .lazy_get(left_renamer.keys())\n .set_index(list(left_renamer.values()))\n )\n new_dtypes = ModinDtypes.concat([left_dtypes, right_dtypes])\n \n new_self = self.__constructor__(\n self._modin_frame.broadcast_apply_full_axis(\n axis=1,\n func=map_func,\n enumerate_partitions=how == \"left\",\n other=right._modin_frame,\n # We're going to explicitly change the shape across the 1-axis,\n # so we want for partitioning to adapt as well\n keep_partitioning=False,\n num_splits=merge_partitioning(\n self._modin_frame, right._modin_frame, axis=1\n ),\n new_columns=new_columns,\n sync_labels=False,\n dtypes=new_dtypes,\n pass_axis_lengths_to_partitions=how == \"left\",\n )\n )\n \n # Here we want to understand whether we're joining on a column or on an index level.\n # It's cool if indexes are already materialized so we can easily check that, if not\n # it's fine too, we can also decide that by columns, which tend to be already\n # materialized quite often compared to the indexes.\n keep_index = False\n if self._modin_frame.has_materialized_index:\n> keep_index = should_keep_index(self, right_pandas)\nE NameError: name 'right_pandas' is not defined\n\nmodin/core/storage_formats/pandas/query_compiler.py:643: NameError\n----------------------------- Captured stderr call -----------------------------\n2024-01-10 15:52:55,130\tINFO worker.py:1664 -- Started a local Ray instance. View the dashboard at \u001b[1m\u001b[32m127.0.0.1:8265 \u001b[39m\u001b[22m\n------------------------------ Captured log call -------------------------------\nINFO fuzzydata.core.workflow:workflow.py:53 Creating new Workflow cc54e466\nINFO fuzzydata.core.generator:generator.py:72 Generating base df with 1000 rows and 33 columns\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=True; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO fuzzydata.core.generator:generator.py:301 Selected Artifact: Artifact(label=artifact_0), initializing operation chain\nINFO fuzzydata.core.generator:generator.py:339 Chaining Operation: sample\nINFO fuzzydata.core.generator:generator.py:70 Generating right-merge df df with 998 rows and 5 columns\nINFO modin.logger.default:partition_manager.py:89 Using sequential splitting in '.from_pandas()' because of some of the conditions are False: enough_elements=False; all_numeric_types=False; async_mode_on=False\nINFO fuzzydata.core.generator:generator.py:339 Chaining Operation: merge\nINFO fuzzydata.core.generator:generator.py:365 Executing current operation list: {'sources': ['artifact_0', 'artifact_1'], 'new_label': None, 'op_list': [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]}\nERROR fuzzydata.core.workflow:workflow.py:192 Could not execute Operation: [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]\nERROR fuzzydata.core.generator:generator.py:376 Error during generation, stopping...\nERROR fuzzydata.core.generator:generator.py:377 Was trying to execute operation: {'sources': ['artifact_0', 'artifact_1'], 'new_label': None, 'op_list': [{'op': 'sample', 'args': {'frac': 0.73}}, {'op': 'merge', 'args': {'key_col': 'uqdyN__bothify'}}]} on soruce artifact(s): [Artifact(label=artifact_0), Artifact(label=artifact_1)]\nERROR fuzzydata.core.generator:generator.py:378 Was trying to execute code: self.sources[0].table.sample(frac=0.73).merge(self.sources[1].table, on=\"uqdyN__bothify\")\nERROR fuzzydata.core.generator:generator.py:379 Writing out all files to /tmp/fuzzydata-test-wf-ray//cc54e466/\n=============================== warnings summary ===============================\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google.cloud')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\nmodin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow\n DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n=========================== short test summary info ============================\nFAILED modin/experimental/fuzzydata/test/test_fuzzydata.py::test_fuzzydata_sample_workflow - NameError: name 'right_pandas' is not defined\n======================== 1 failed, 6 warnings in 11.54s ========================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/modin/core/storage_formats/pandas/query_compiler.py b/modin/core/storage_formats/pandas/query_compiler.py\nindex 32b379994..7ca5bb8f1 100644\n--- a/modin/core/storage_formats/pandas/query_compiler.py\n+++ b/modin/core/storage_formats/pandas/query_compiler.py\n@@ -640,19 +640,19 @@ class PandasQueryCompiler(BaseQueryCompiler):\n # materialized quite often compared to the indexes.\n keep_index = False\n if self._modin_frame.has_materialized_index:\n- keep_index = should_keep_index(self, right_pandas)\n+ keep_index = should_keep_index(self, right)\n else:\n # Have to trigger columns materialization. Hope they're already available at this point.\n if left_on is not None and right_on is not None:\n keep_index = any(\n- o not in right_pandas.columns\n+ o not in right.columns\n and o in left_on\n and o not in self.columns\n for o in right_on\n )\n elif on is not None:\n keep_index = any(\n- o not in right_pandas.columns and o not in self.columns\n+ o not in right.columns and o not in self.columns\n for o in on\n )\n \n", "difficulty": 2, "changed_files": ["modin/core/storage_formats/pandas/query_compiler.py"], "commit_link": "https://github.com/modin-project/modin/tree/034b1956a0914b739c3e5fc2e19ffcfee0f08c02"}
|
|
|
|
data/python/03669a5.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 153, "repo_owner": "mikel-brostrom", "repo_name": "yolo_tracking", "head_branch": "centroid-asso-support", "workflow_name": "CI CPU testing", "workflow_filename": "ci.yml", "workflow_path": ".github/workflows/ci.yml", "contributor": "mikel-brostrom", "sha_fail": "03669a5d72130c57575bedd657b82c601f08a982", "sha_success": "0948605abb9d6d962450f1bbf9a0b9c96c429b29", "workflow": "# name of the workflow, what it is doing (optional)\nname: CI CPU testing\n\n# events that trigger the workflow (required)\non:\n push:\n branches: [master, CIdebug]\n pull_request:\n # pull request where master is target\n branches: [master]\n\nenv:\n # Directory of PyPi package to be tested\n PACKAGE_DIR: boxmot\n # Minimum acceptable test coverage\n # Increase as you add more tests to increase coverage\n COVERAGE_FAIL_UNDER: 29\n\n# the workflow that gets triggerd\njobs:\n build:\n runs-on: ${{ matrix.os }}\n strategy:\n fail-fast: false\n matrix:\n os: [ubuntu-latest] # skip windows-latest for\n python-version: ['3.8', '3.9', '3.10']\n #model: ['yolov8n', 'yolo_nas_s', yolox_n] # yolo models to test\n #tracking-methods: ['deepocsort', 'ocsort', 'botsort', 'strongsort', 'bytetrack'] # tracking methods to test\n\n # Timeout: https://stackoverflow.com/a/59076067/4521646\n timeout-minutes: 50\n steps:\n\n - uses: actions/checkout@v4 # Check out the repository\n - uses: actions/setup-python@v5 # Prepare environment with python 3.9\n with:\n python-version: ${{ matrix.python-version }}\n cache: 'pip' # caching pip dependencies\n - name: Install requirements\n shell: bash # for Windows compatibility\n run: |\n python -m pip install --upgrade pip setuptools wheel\n pip install -e . pytest pytest-cov --extra-index-url https://download.pytorch.org/whl/cpu\n python --version\n pip --version\n pip list\n\n - name: Tests all tracking options\n shell: bash # for Windows compatibility\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n run: |\n # deepocsort fro all supported yolo models\n python examples/track.py --tracking-method deepocsort --source $IMG --imgsz 320\n python examples/track.py --yolo-model yolo_nas_s --tracking-method deepocsort --source $IMG --imgsz 320\n # python examples/track.py --yolo-model yolox_n --tracking-method deepocsort --source $IMG --imgsz 320\n\n # hybridsort\n python examples/track.py --tracking-method hybridsort --source $IMG --imgsz 320\n\n # botsort\n python examples/track.py --tracking-method botsort --source $IMG --imgsz 320\n\n # strongsort\n python examples/track.py --tracking-method strongsort --source $IMG --imgsz 320\n\n # ocsort\n python examples/track.py --tracking-method ocsort --source $IMG --imgsz 320\n\n # bytetrack\n python examples/track.py --tracking-method bytetrack --source $IMG --imgsz 320\n\n - name: Pytest tests # after tracking options as this does not download models\n shell: bash # for Windows compatibility\n run: |\n\n # needed in TFLite export\n wget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\n tar -zxvf flatc.tar.gz\n sudo chmod +x flatc\n sudo mv flatc /usr/bin/\n\n pytest --cov=$PACKAGE_DIR --cov-report=html -v tests\n coverage report --fail-under=$COVERAGE_FAIL_UNDER\n\n - name: Tests exported reid models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n\n # test exported reid model\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17.torchscript --source $IMG --imgsz 320\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17.onnx --source $IMG --imgsz 320\n #python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17_saved_model/osnet_x0_25_msmt17_float16.tflite --source $IMG --imgsz 320\n python examples/track.py --reid-model examples/weights/osnet_x0_25_msmt17_openvino_model --source $IMG --imgsz 320\n\n - name: Test tracking with seg models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n # tracking with SEG models\n python examples/track.py --tracking-method deepocsort --yolo-model yolov8n-seg.pt --source $IMG\n\n - name: Test tracking with pose models\n env:\n IMG: ./assets/MOT17-mini/train/MOT17-05-FRCNN/img1/000001.jpg\n shell: bash # for Windows compatibility\n run: |\n # tracking with POSE models\n python3 examples/track.py --yolo-model weights/yolov8n.pt --source $IMG --imgsz 320\n\n - name: Test validation on MOT17 subset\n shell: bash # for Windows compatibility\n run: |\n # validation on a few MOT17 imges\n python examples/val.py --tracking-method deepocsort --yolo-model yolov8n.pt --benchmark MOT17-mini --imgsz 320 --conf 0.2\n\n - name: Test evolution on MOT17 subset\n shell: bash # for Windows compatibility\n run: |\n # evolve a for a single set of parameters\n python examples/evolve.py --objectives HOTA,MOTA,IDF1 --benchmark MOT17-mini --n-trials 1 --imgsz 320 --conf 0.2\n", "logs": [{"step_name": "build (ubuntu-latest, 3.8)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n--2024-01-12 20:11:25-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 140.82.112.3\nConnecting to github.com (github.com)|140.82.112.3|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201125Z&X-Amz-Expires=300&X-Amz-Signature=3bfcceeafbe583d085cec55f86a8526c6e9f64b963ff80c4ac3797875e856d28&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:11:25-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201125Z&X-Amz-Expires=300&X-Amz-Signature=3bfcceeafbe583d085cec55f86a8526c6e9f64b963ff80c4ac3797875e856d28&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.110.133, 185.199.108.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.110.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 28.1M 0s\n 50K .......... .......... .......... .......... .......... 7% 36.5M 0s\n 100K .......... .......... .......... .......... .......... 11% 143M 0s\n 150K .......... .......... .......... .......... .......... 14% 42.2M 0s\n 200K .......... .......... .......... .......... .......... 18% 131M 0s\n 250K .......... .......... .......... .......... .......... 22% 167M 0s\n 300K .......... .......... .......... .......... .......... 25% 200M 0s\n 350K .......... .......... .......... .......... .......... 29% 74.0M 0s\n 400K .......... .......... .......... .......... .......... 33% 187M 0s\n 450K .......... .......... .......... .......... .......... 37% 172M 0s\n 500K .......... .......... .......... .......... .......... 40% 132M 0s\n 550K .......... .......... .......... .......... .......... 44% 233M 0s\n 600K .......... .......... .......... .......... .......... 48% 155M 0s\n 650K .......... .......... .......... .......... .......... 51% 208M 0s\n 700K .......... .......... .......... .......... .......... 55% 136M 0s\n 750K .......... .......... .......... .......... .......... 59% 223M 0s\n 800K .......... .......... .......... .......... .......... 62% 136M 0s\n 850K .......... .......... .......... .......... .......... 66% 140M 0s\n 900K .......... .......... .......... .......... .......... 70% 238M 0s\n 950K .......... .......... .......... .......... .......... 74% 241M 0s\n 1000K .......... .......... .......... .......... .......... 77% 242M 0s\n 1050K .......... .......... .......... .......... .......... 81% 215M 0s\n 1100K .......... .......... .......... .......... .......... 85% 238M 0s\n 1150K .......... .......... .......... .......... .......... 88% 240M 0s\n 1200K .......... .......... .......... .......... .......... 92% 236M 0s\n 1250K .......... .......... .......... .......... .......... 96% 198M 0s\n 1300K .......... .......... .......... .......... .......... 99% 251M 0s\n 1350K 100% 572G=0.01s\n\n2024-01-12 20:11:25 (120 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.8.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7f1dba152ca0>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 9%|\u2589 | 2.10M/22.4M [00:00<00:01, 19.5MB/s]\n 40%|\u2588\u2588\u2588\u2589 | 8.91M/22.4M [00:00<00:00, 43.6MB/s]\n 82%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f | 18.4M/22.4M [00:00<00:00, 65.1MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 64.2MB/s]\n\u001b[32m2024-01-12 20:11:55.585\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:11:55.738\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('mpl_toolkits')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.8.18-final-0 -----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 9 warnings in 31.85s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (ubuntu-latest, 3.9)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n--2024-01-12 20:11:58-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 192.30.255.113\nConnecting to github.com (github.com)|192.30.255.113|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201159Z&X-Amz-Expires=300&X-Amz-Signature=709497883dd2c2a9471410daf12c63e704acb6b779fb216aa16d512cc5577229&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:11:59-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201159Z&X-Amz-Expires=300&X-Amz-Signature=709497883dd2c2a9471410daf12c63e704acb6b779fb216aa16d512cc5577229&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.111.133, 185.199.108.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.111.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 10.7M 0s\n 50K .......... .......... .......... .......... .......... 7% 44.1M 0s\n 100K .......... .......... .......... .......... .......... 11% 89.2M 0s\n 150K .......... .......... .......... .......... .......... 14% 62.8M 0s\n 200K .......... .......... .......... .......... .......... 18% 108M 0s\n 250K .......... .......... .......... .......... .......... 22% 163M 0s\n 300K .......... .......... .......... .......... .......... 25% 186M 0s\n 350K .......... .......... .......... .......... .......... 29% 202M 0s\n 400K .......... .......... .......... .......... .......... 33% 108M 0s\n 450K .......... .......... .......... .......... .......... 37% 199M 0s\n 500K .......... .......... .......... .......... .......... 40% 190M 0s\n 550K .......... .......... .......... .......... .......... 44% 198M 0s\n 600K .......... .......... .......... .......... .......... 48% 171M 0s\n 650K .......... .......... .......... .......... .......... 51% 212M 0s\n 700K .......... .......... .......... .......... .......... 55% 170M 0s\n 750K .......... .......... .......... .......... .......... 59% 185M 0s\n 800K .......... .......... .......... .......... .......... 62% 199M 0s\n 850K .......... .......... .......... .......... .......... 66% 259M 0s\n 900K .......... .......... .......... .......... .......... 70% 191M 0s\n 950K .......... .......... .......... .......... .......... 74% 205M 0s\n 1000K .......... .......... .......... .......... .......... 77% 193M 0s\n 1050K .......... .......... .......... .......... .......... 81% 267M 0s\n 1100K .......... .......... .......... .......... .......... 85% 271M 0s\n 1150K .......... .......... .......... .......... .......... 88% 239M 0s\n 1200K .......... .......... .......... .......... .......... 92% 268M 0s\n 1250K .......... .......... .......... .......... .......... 96% 260M 0s\n 1300K .......... .......... .......... .......... .......... 99% 269M 0s\n 1350K 100% 572G=0.01s\n\n2024-01-12 20:11:59 (103 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.9.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.9.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7ff526351e50>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 5%|\u258d | 1.05M/22.4M [00:00<00:02, 9.56MB/s]\n 21%|\u2588\u2588 | 4.72M/22.4M [00:00<00:00, 21.4MB/s]\n 54%|\u2588\u2588\u2588\u2588\u2588\u258e | 12.1M/22.4M [00:00<00:00, 32.0MB/s]\n 77%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b | 17.3M/22.4M [00:00<00:00, 32.2MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 37.0MB/s]\n\u001b[32m2024-01-12 20:12:31.441\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:12:31.599\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.9.18-final-0 -----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 8 warnings in 33.86s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (ubuntu-latest, 3.10)/6_Pytest tests.txt", "log": "##[group]Run # needed in TFLite export\n\u001b[36;1m\u001b[0m\n\u001b[36;1m# needed in TFLite export\u001b[0m\n\u001b[36;1mwget https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\u001b[0m\n\u001b[36;1mtar -zxvf flatc.tar.gz\u001b[0m\n\u001b[36;1msudo chmod +x flatc\u001b[0m\n\u001b[36;1msudo mv flatc /usr/bin/\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpytest --cov=$PACKAGE_DIR --cov-report=html -v tests\u001b[0m\n\u001b[36;1mcoverage report --fail-under=$COVERAGE_FAIL_UNDER\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PACKAGE_DIR: boxmot\n COVERAGE_FAIL_UNDER: 29\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n--2024-01-12 20:12:02-- https://github.com/PINTO0309/onnx2tf/releases/download/1.16.31/flatc.tar.gz\nResolving github.com (github.com)... 140.82.113.3\nConnecting to github.com (github.com)|140.82.113.3|:443... connected.\nHTTP request sent, awaiting response... 302 Found\nLocation: https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201203Z&X-Amz-Expires=300&X-Amz-Signature=f339441879645b502b6bf0bbc8e04deb0dcdf54ed1dcd223e8a7e04adef107ad&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream [following]\n--2024-01-12 20:12:03-- https://objects.githubusercontent.com/github-production-release-asset-2e65be/541831874/29499355-44ab-4fb6-86c8-582f4bad68a3?X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Credential=AKIAVCODYLSA53PQK4ZA%2F20240112%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Date=20240112T201203Z&X-Amz-Expires=300&X-Amz-Signature=f339441879645b502b6bf0bbc8e04deb0dcdf54ed1dcd223e8a7e04adef107ad&X-Amz-SignedHeaders=host&actor_id=0&key_id=0&repo_id=541831874&response-content-disposition=attachment%3B%20filename%3Dflatc.tar.gz&response-content-type=application%2Foctet-stream\nResolving objects.githubusercontent.com (objects.githubusercontent.com)... 185.199.108.133, 185.199.110.133, 185.199.109.133, ...\nConnecting to objects.githubusercontent.com (objects.githubusercontent.com)|185.199.108.133|:443... connected.\nHTTP request sent, awaiting response... 200 OK\nLength: 1382707 (1.3M) [application/octet-stream]\nSaving to: \u2018flatc.tar.gz\u2019\n\n 0K .......... .......... .......... .......... .......... 3% 5.77M 0s\n 50K .......... .......... .......... .......... .......... 7% 6.18M 0s\n 100K .......... .......... .......... .......... .......... 11% 30.6M 0s\n 150K .......... .......... .......... .......... .......... 14% 39.6M 0s\n 200K .......... .......... .......... .......... .......... 18% 9.54M 0s\n 250K .......... .......... .......... .......... .......... 22% 49.5M 0s\n 300K .......... .......... .......... .......... .......... 25% 65.1M 0s\n 350K .......... .......... .......... .......... .......... 29% 45.1M 0s\n 400K .......... .......... .......... .......... .......... 33% 47.7M 0s\n 450K .......... .......... .......... .......... .......... 37% 86.6M 0s\n 500K .......... .......... .......... .......... .......... 40% 13.9M 0s\n 550K .......... .......... .......... .......... .......... 44% 44.2M 0s\n 600K .......... .......... .......... .......... .......... 48% 74.1M 0s\n 650K .......... .......... .......... .......... .......... 51% 127M 0s\n 700K .......... .......... .......... .......... .......... 55% 88.8M 0s\n 750K .......... .......... .......... .......... .......... 59% 126M 0s\n 800K .......... .......... .......... .......... .......... 62% 178M 0s\n 850K .......... .......... .......... .......... .......... 66% 77.3M 0s\n 900K .......... .......... .......... .......... .......... 70% 209M 0s\n 950K .......... .......... .......... .......... .......... 74% 180M 0s\n 1000K .......... .......... .......... .......... .......... 77% 159M 0s\n 1050K .......... .......... .......... .......... .......... 81% 14.6M 0s\n 1100K .......... .......... .......... .......... .......... 85% 52.9M 0s\n 1150K .......... .......... .......... .......... .......... 88% 186M 0s\n 1200K .......... .......... .......... .......... .......... 92% 127M 0s\n 1250K .......... .......... .......... .......... .......... 96% 210M 0s\n 1300K .......... .......... .......... .......... .......... 99% 216M 0s\n 1350K 100% 8.87M=0.04s\n\n2024-01-12 20:12:03 (31.1 MB/s) - \u2018flatc.tar.gz\u2019 saved [1382707/1382707]\n\nflatc\n============================= test session starts ==============================\nplatform linux -- Python 3.10.13, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.10.13/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yolo_tracking/yolo_tracking\nplugins: hydra-core-1.3.2, cov-4.1.0\ncollecting ... collected 14 items\n\ntests/test_exports.py::test_export_torchscript PASSED [ 7%]\ntests/test_exports.py::test_export_onnx PASSED [ 14%]\ntests/test_exports.py::test_export_openvino PASSED [ 21%]\ntests/test_python.py::test_strongsort_instantiation PASSED [ 28%]\ntests/test_python.py::test_botsort_instantiation PASSED [ 35%]\ntests/test_python.py::test_deepocsort_instantiation PASSED [ 42%]\ntests/test_python.py::test_ocsort_instantiation PASSED [ 50%]\ntests/test_python.py::test_bytetrack_instantiation PASSED [ 57%]\ntests/test_python.py::test_deepocsort_output FAILED [ 64%]\ntests/test_python.py::test_ocsort_output PASSED [ 71%]\ntests/test_python.py::test_botsort_output PASSED [ 78%]\ntests/test_python.py::test_bytetrack_output PASSED [ 85%]\ntests/test_python.py::test_strongsort_output PASSED [ 92%]\ntests/test_python.py::test_gsi PASSED [100%]\n\n=================================== FAILURES ===================================\n____________________________ test_deepocsort_output ____________________________\n\n def test_deepocsort_output():\n tracker_conf = get_tracker_config('deepocsort')\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n rgb = np.random.randint(255, size=(640, 640, 3), dtype=np.uint8)\n det = np.array([[144, 212, 578, 480, 0.82, 0],\n [425, 281, 576, 472, 0.56, 65]])\n output = tracker.update(det, rgb)\n # Works since frame count is less than min hits (1 <= 2)\n assert output.shape == (2, 8) # two inputs should give two outputs\n output = np.flip(np.delete(output, [4, 7], axis=1), axis=0)\n assert_allclose(det, output, atol=1, rtol=7e-3, verbose=True)\n \n # Instantiate new tracker and ensure minimum number of hits works\n tracker = create_tracker(\n tracker_type='deepocsort',\n tracker_config=tracker_conf,\n reid_weights=WEIGHTS / 'mobilenetv2_x1_4_dukemtmcreid.pt',\n device='cpu',\n half=False,\n per_class=False\n )\n tracker.min_hits = 2\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(np.empty((0, 6)), rgb)\n assert output.size == 0\n output = tracker.update(det, rgb)\n assert output.size == 0\n> output = tracker.update(det, rgb)\n\ntests/test_python.py:86: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nboxmot/trackers/deepocsort/deep_ocsort.py:434: in update\n matched, unmatched_dets, unmatched_trks = associate(\nboxmot/utils/association.py:146: in associate\n iou_matrix = run_asso_func(asso_func, detections, trackers, w, h)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nfunc = <function giou_batch at 0x7f20a69f0700>\nargs = (array([[1.44e+02, 2.12e+02, 5.78e+02, 4.80e+02, 8.20e-01],\n [4.25e+02, 2.81e+02, 5.76e+02, 4.72e+02, 5.60e-01]]), array([[144., 212., 578., 480., 0.],\n [425., 281., 576., 472., 0.]]), 640, 640)\n\n def run_asso_func(func, *args):\n \"\"\"\n Wrapper function that checks the inputs to the association functions\n and then call either one of the iou association functions or centroid.\n \n Parameters:\n func: The batch function to call (either *iou*_batch or centroid_batch).\n *args: Variable length argument list, containing either bounding boxes and optionally size parameters.\n \"\"\"\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n return func(*args)\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n return func(*args)\n else:\n> raise ValueError(\"No such association method\")\nE ValueError: No such association method\n\nboxmot/utils/iou.py:212: ValueError\n----------------------------- Captured stderr call -----------------------------\nDownloading...\nFrom: https://drive.google.com/uc?id=12uD5FeVqLg9-AFDju2L7SQxjmPb4zpBN\nTo: /home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\n\n 0%| | 0.00/22.4M [00:00<?, ?B/s]\n 2%|\u258f | 524k/22.4M [00:00<00:05, 3.80MB/s]\n 35%|\u2588\u2588\u2588\u258c | 7.86M/22.4M [00:00<00:00, 38.7MB/s]\n 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258d | 16.8M/22.4M [00:00<00:00, 58.3MB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 22.4M/22.4M [00:00<00:00, 54.7MB/s]\n\u001b[32m2024-01-12 20:12:35.371\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n\u001b[32m2024-01-12 20:12:35.534\u001b[0m | \u001b[32m\u001b[1mSUCCESS \u001b[0m | \u001b[36mboxmot.appearance.reid_model_factory\u001b[0m:\u001b[36mload_pretrained_weights\u001b[0m:\u001b[36m207\u001b[0m - \u001b[32m\u001b[1mSuccessfully loaded pretrained weights from \"/home/runner/work/yolo_tracking/yolo_tracking/examples/weights/mobilenetv2_x1_4_dukemtmcreid.pt\"\u001b[0m\n=============================== warnings summary ===============================\nboxmot/utils/checks.py:5\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/utils/checks.py:5: DeprecationWarning: pkg_resources is deprecated as an API. See https://setuptools.pypa.io/en/latest/pkg_resources.html\n import pkg_resources\n\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('google')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n../../../../../opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pkg_resources/__init__.py:2868: DeprecationWarning: Deprecated call to `pkg_resources.declare_namespace('sphinxcontrib')`.\n Implementing implicit namespace packages (as specified in PEP 420) is preferred to `pkg_resources.declare_namespace`. See https://setuptools.pypa.io/en/latest/references/keywords.html#keyword-namespace-packages\n declare_namespace(pkg)\n\ntests/test_exports.py::test_export_onnx\n /home/runner/work/yolo_tracking/yolo_tracking/boxmot/appearance/backbones/osnet.py:389: TracerWarning: Converting a tensor to a Python boolean might cause the trace to be incorrect. We can't record the data flow of Python values, so this value will be treated as a constant in the future. This means that the trace might not generalize to other inputs!\n if return_featuremaps:\n\ntests/test_exports.py::test_export_openvino\n /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/defusedxml/__init__.py:30: DeprecationWarning: defusedxml.cElementTree is deprecated, import from defusedxml.ElementTree instead.\n from . import cElementTree\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\n---------- coverage: platform linux, python 3.10.13-final-0 ----------\nCoverage HTML written to dir htmlcov\n\n=========================== short test summary info ============================\nFAILED tests/test_python.py::test_deepocsort_output - ValueError: No such association method\n================== 1 failed, 13 passed, 8 warnings in 33.78s ===================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/boxmot/utils/iou.py b/boxmot/utils/iou.py\nindex d8bfc7e..144f40f 100644\n--- a/boxmot/utils/iou.py\n+++ b/boxmot/utils/iou.py\n@@ -200,10 +200,10 @@ def run_asso_func(func, *args):\n if func not in [iou_batch, giou_batch, diou_batch, ciou_batch, centroid_batch]:\n raise ValueError(\"Invalid function specified. Must be either '(g,d,c, )iou_batch' or 'centroid_batch'.\")\n \n- if func is (iou_batch or giou_batch or diou_batch or ciou_batch):\n+ if func in (iou_batch, giou_batch, diou_batch, ciou_batch):\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[0:2]):\n raise ValueError(\"Invalid arguments for iou_batch. Expected two bounding boxes.\")\n- return func(*args)\n+ return func(*args[0:2])\n elif func is centroid_batch:\n if len(args) != 4 or not all(isinstance(arg, (list, np.ndarray)) for arg in args[:2]) or not all(isinstance(arg, (int)) for arg in args[2:]):\n raise ValueError(\"Invalid arguments for centroid_batch. Expected two bounding boxes and two size parameters.\")\n", "difficulty": 2, "changed_files": ["boxmot/utils/iou.py"], "commit_link": "https://github.com/mikel-brostrom/yolo_tracking/tree/03669a5d72130c57575bedd657b82c601f08a982"}
|
|
|
|
data/python/06bc6b6.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 93, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "composer_lora", "workflow_name": "Code Quality Checks", "workflow_filename": "code-quality.yaml", "workflow_path": ".github/workflows/code-quality.yaml", "contributor": "dakinggg", "sha_fail": "06bc6b627d6afb938950382ba50b6b71432d7cf6", "sha_success": "23961743885e704b3f4bd77ba8410b2508d5127c", "workflow": "name: Code Quality Checks\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n code-quality:\n runs-on: ubuntu-20.04\n timeout-minutes: 15\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n pip_deps:\n - \"[dev]\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .${{ matrix.pip_deps }}\n - name: Run checks\n run: |\n pre-commit run --all-files\n", "logs": [{"step_name": "code-quality (3.9, [dev])/5_Run checks.txt", "log": "##[group]Run pre-commit run --all-files\n\u001b[36;1mpre-commit run --all-files\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n[INFO] Initializing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Initializing environment for https://github.com/google/yapf.\n[INFO] Initializing environment for https://github.com/google/yapf:toml.\n[INFO] Initializing environment for https://github.com/pycqa/isort.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle:toml.\n[INFO] Initializing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/kynan/nbstripout.\n[INFO] Initializing environment for local:[email protected].\n[INFO] Initializing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Installing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/google/yapf.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pycqa/isort.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/kynan/nbstripout.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for local.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\nruff.....................................................................Passed\nyapf.....................................................................Passed\nisort....................................................................Failed\n- hook id: isort\n- files were modified by this hook\n\nFixing /home/runner/work/composer/composer/tests/models/test_hf_model.py\nSkipped 1 files\n\npydocstyle...............................................................Passed\ncheck for added large files..............................................Passed\ncheck python ast.........................................................Passed\ncheck builtin type constructor use.......................................Passed\ncheck for case conflicts.................................................Passed\ncheck docstring is first.................................................Passed\ncheck that executables have shebangs.....................................Passed\ncheck json...............................................................Passed\ncheck that scripts with shebangs are executable..........................Passed\npretty format json.......................................................Passed\ncheck for merge conflicts................................................Passed\ncheck for broken symlinks................................................Passed\ncheck toml...............................................................Passed\ncheck vcs permalinks.....................................................Passed\ncheck xml................................................................Passed\ncheck yaml...............................................................Passed\ndebug statements (python)................................................Passed\ndetect destroyed symlinks................................................Passed\nfix double quoted strings................................................Passed\nfix end of files.........................................................Passed\nfix utf-8 byte order marker..............................................Passed\nmixed line ending........................................................Passed\ntrim trailing whitespace.................................................Passed\nInsert license in comments...............................................Passed\nnbstripout...............................................................Passed\npyright..................................................................Passed\nsecret scan..............................................................Passed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/composer/models/huggingface.py b/composer/models/huggingface.py\nindex f3855ce6..c223a917 100644\n--- a/composer/models/huggingface.py\n+++ b/composer/models/huggingface.py\n@@ -111,6 +111,13 @@ class HuggingFaceModel(ComposerModel):\n conda_package='peft',\n conda_channel='conda-forge')\n \n+ if peft_config is not None:\n+ # Hugging Face requires the peft type and task type to be upper case, so we do that here\n+ # https://github.com/huggingface/peft/blob/ebbff4023ad276cbcb2466fd7e99be7d3ae0ae11/src/peft/utils/peft_types.py#L22-L51\n+ if isinstance(peft_config.peft_type, str):\n+ peft_config.peft_type = peft_config.peft_type.upper()\n+ if isinstance(peft_config.task_type, str):\n+ peft_config.task_type = peft_config.task_type.upper()\n if peft_config is not None and peft_config.peft_type != 'LORA':\n raise ValueError(\n f'PEFT type {peft_config.peft_type} is not supported by HuggingFaceModel. Only LORA is supported.')\ndiff --git a/tests/models/test_hf_model.py b/tests/models/test_hf_model.py\nindex c3c59e1c..a69b56b5 100644\n--- a/tests/models/test_hf_model.py\n+++ b/tests/models/test_hf_model.py\n@@ -28,8 +28,6 @@ from tests.common.markers import device, world_size\n from tests.common.models import (configure_tiny_bert_model, configure_tiny_bert_tokenizer, configure_tiny_gpt2_model,\n configure_tiny_gpt2_tokenizer, configure_tiny_mistral_model,\n configure_tiny_mistral_tokenizer, configure_tiny_t5_model, configure_tiny_t5_tokenizer)\n-from tests.common.models import (configure_tiny_bert_model, configure_tiny_bert_tokenizer, configure_tiny_gpt2_model,\n- configure_tiny_gpt2_tokenizer, configure_tiny_t5_model, configure_tiny_t5_tokenizer)\n from tests.loggers.test_remote_uploader_downloader import DummyObjectStore\n \n if TYPE_CHECKING:\n", "difficulty": 0, "changed_files": ["composer/models/huggingface.py", "tests/models/test_hf_model.py"], "commit_link": "https://github.com/mosaicml/composer/tree/06bc6b627d6afb938950382ba50b6b71432d7cf6"}
|
|
|
|
data/python/077f6aa.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 116, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "main", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "c-core-labs", "sha_fail": "077f6aaac3ebb96626ac747fb126a0b4d752489c", "sha_success": "4c315803bfdb46fa7818cb08fd7224fb0bb599f1", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 26.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 90.4 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 116.1 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 81.8 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 92.1 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|22e1ac8b31209d20dbc1af062b7bb70cb716a8ce5ddbf255b52ad3a34004f0dc\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nReceived 0 of 48057368 (0.0%), 0.0 MBs/sec\nCache Size: ~46 MB (48057368 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/516a3dc9-7b6f-410a-9989-4b9d88d8a69b/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48057368 of 48057368 (100.0%), 22.9 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|22e1ac8b31209d20dbc1af062b7bb70cb716a8ce5ddbf255b52ad3a34004f0dc\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nruff.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: ruff\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\nFound 1 error (1 fixed, 0 remaining).\n\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted wandb/integration/ultralytics/callback.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m870 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/wandb/integration/ultralytics/bbox_utils.py b/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[1mindex 798aa57..9cee915 100644\u001b[m\n\u001b[1m--- a/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[1m+++ b/wandb/integration/ultralytics/bbox_utils.py\u001b[m\n\u001b[36m@@ -3,6 +3,7 @@\u001b[m \u001b[mfrom typing import Any, Dict, List, Optional, Tuple, Union\u001b[m\n import torch\u001b[m\n from ultralytics.engine.results import Results\u001b[m\n from ultralytics.models.yolo.detect import DetectionPredictor\u001b[m\n\u001b[32m+\u001b[m\n try:\u001b[m\n from ultralytics.yolo.utils import ops\u001b[m\n except ModuleNotFoundError:\u001b[m\n\u001b[1mdiff --git a/wandb/integration/ultralytics/callback.py b/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[1mindex ed8a224..618fc5a 100644\u001b[m\n\u001b[1m--- a/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[1m+++ b/wandb/integration/ultralytics/callback.py\u001b[m\n\u001b[36m@@ -42,6 +42,7 @@\u001b[m \u001b[mtry:\u001b[m\n SegmentationValidator,\u001b[m\n )\u001b[m\n from ultralytics.utils.torch_utils import de_parallel\u001b[m\n\u001b[32m+\u001b[m\n try:\u001b[m\n from ultralytics.yolo.utils import RANK, __version__\u001b[m\n except ModuleNotFoundError:\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/wandb/cli/cli.py b/wandb/cli/cli.py\nindex 94a3ece90..1d0031488 100644\n--- a/wandb/cli/cli.py\n+++ b/wandb/cli/cli.py\n@@ -1496,6 +1496,8 @@ def launch(\n if cli_template_vars:\n if queue is None:\n raise LaunchError(\"'--set-var' flag requires queue to be set\")\n+ if entity is None:\n+ entity = launch_utils.get_default_entity(api, config)\n public_api = PublicApi()\n runqueue = RunQueue(client=public_api.client, name=queue, entity=entity)\n template_variables = launch_utils.fetch_and_validate_template_variables(\ndiff --git a/wandb/integration/ultralytics/bbox_utils.py b/wandb/integration/ultralytics/bbox_utils.py\nindex 798aa573a..8b924697d 100644\n--- a/wandb/integration/ultralytics/bbox_utils.py\n+++ b/wandb/integration/ultralytics/bbox_utils.py\n@@ -3,10 +3,7 @@ from typing import Any, Dict, List, Optional, Tuple, Union\n import torch\n from ultralytics.engine.results import Results\n from ultralytics.models.yolo.detect import DetectionPredictor\n-try:\n- from ultralytics.yolo.utils import ops\n-except ModuleNotFoundError:\n- from ultralytics.utils import ops\n+from ultralytics.yolo.utils import ops\n \n import wandb\n \ndiff --git a/wandb/integration/ultralytics/callback.py b/wandb/integration/ultralytics/callback.py\nindex ed8a22465..2e133a163 100644\n--- a/wandb/integration/ultralytics/callback.py\n+++ b/wandb/integration/ultralytics/callback.py\n@@ -42,10 +42,7 @@ try:\n SegmentationValidator,\n )\n from ultralytics.utils.torch_utils import de_parallel\n- try:\n- from ultralytics.yolo.utils import RANK, __version__\n- except ModuleNotFoundError:\n- from ultralytics.utils import RANK, __version__\n+ from ultralytics.yolo.utils import RANK, __version__\n \n from wandb.integration.ultralytics.bbox_utils import (\n plot_predictions,\ndiff --git a/wandb/integration/yolov8/yolov8.py b/wandb/integration/yolov8/yolov8.py\nindex 1c5dbd0fe..310a39114 100644\n--- a/wandb/integration/yolov8/yolov8.py\n+++ b/wandb/integration/yolov8/yolov8.py\n@@ -2,13 +2,8 @@ from typing import Any, Callable, Dict, List, Optional\n \n from ultralytics.yolo.engine.model import YOLO\n from ultralytics.yolo.engine.trainer import BaseTrainer\n-\n-try:\n- from ultralytics.yolo.utils import RANK\n- from ultralytics.yolo.utils.torch_utils import get_flops, get_num_params\n-except ModuleNotFoundError:\n- from ultralytics.utils import RANK\n- from ultralytics.utils.torch_utils import get_flops, get_num_params\n+from ultralytics.yolo.utils import RANK\n+from ultralytics.yolo.utils.torch_utils import get_flops, get_num_params\n from ultralytics.yolo.v8.classify.train import ClassificationTrainer\n \n import wandb\ndiff --git a/wandb/sdk/launch/utils.py b/wandb/sdk/launch/utils.py\nindex 05efe128e..69f4e57bb 100644\n--- a/wandb/sdk/launch/utils.py\n+++ b/wandb/sdk/launch/utils.py\n@@ -204,10 +204,7 @@ def set_project_entity_defaults(\n config_project = launch_config.get(\"project\")\n project = config_project or source_uri or \"\"\n if entity is None:\n- config_entity = None\n- if launch_config:\n- config_entity = launch_config.get(\"entity\")\n- entity = config_entity or api.default_entity\n+ entity = get_default_entity(api, launch_config)\n prefix = \"\"\n if platform.system() != \"Windows\" and sys.stdout.encoding == \"UTF-8\":\n prefix = \"\ud83d\ude80 \"\n@@ -217,6 +214,13 @@ def set_project_entity_defaults(\n return project, entity\n \n \n+def get_default_entity(api: Api, launch_config: Optional[Dict[str, Any]]):\n+ config_entity = None\n+ if launch_config:\n+ config_entity = launch_config.get(\"entity\")\n+ return config_entity or api.default_entity\n+\n+\n def construct_launch_spec(\n uri: Optional[str],\n job: Optional[str],\n", "difficulty": 0, "changed_files": ["wandb/cli/cli.py", "wandb/integration/ultralytics/bbox_utils.py", "wandb/integration/ultralytics/callback.py", "wandb/integration/yolov8/yolov8.py", "wandb/sdk/launch/utils.py"], "commit_link": "https://github.com/wandb/wandb/tree/077f6aaac3ebb96626ac747fb126a0b4d752489c"}
|
|
|
|
data/python/07bc10c.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 189, "repo_owner": "openvinotoolkit", "repo_name": "open_model_zoo", "head_branch": "ac_sub_evaluation", "workflow_name": "github-actions", "workflow_filename": "accuracy_checker.yml", "workflow_path": ".github/workflows/accuracy_checker.yml", "contributor": "pwolnows", "sha_fail": "07bc10c0e7858b22e9345812af8e6bb6c4ef18be", "sha_success": "76a4c47015ae5c1280dcf23d3d8592df40dbced8", "workflow": "name: github-actions\non: pull_request\njobs:\n accuracy_checker:\n strategy:\n matrix:\n runs-on: [ubuntu-20.04, ubuntu-22.04]\n python-version: [3.8, 3.9, '3.10']\n runs-on: ${{ matrix.runs-on }}\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n - name: Pip install dependencies\n run: |\n python -m pip install --upgrade pip\n python -m pip install tools/accuracy_checker openvino -r tools/accuracy_checker/requirements-test.in\n - name: python -m pytest tools/accuracy_checker\n run: |\n python -m pytest tools/accuracy_checker\n - name: Pip install and run pylint\n working-directory: ./tools/accuracy_checker\n run: |\n python -m pip install pylint==2.10.2\n PYTHONPATH=. python -m pylint --rcfile=.pylintrc `find -wholename '?*/**/*.py' -not -path \"./tests/*\" -not -path \"./build/*\"`\n", "logs": [{"step_name": "accuracy_checker (ubuntu-20.04, 3.8)/6_Pip install and run pylint.txt", "log": "##[group]Run python -m pip install pylint==2.10.2\n\u001b[36;1mpython -m pip install pylint==2.10.2\u001b[0m\n\u001b[36;1mPYTHONPATH=. python -m pylint --rcfile=.pylintrc `find -wholename '?*/**/*.py' -not -path \"./tests/*\" -not -path \"./build/*\"`\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nCollecting pylint==2.10.2\n Downloading pylint-2.10.2-py3-none-any.whl (392 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 392.6/392.6 kB 15.3 MB/s eta 0:00:00\nCollecting platformdirs>=2.2.0 (from pylint==2.10.2)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting astroid<2.8,>=2.7.2 (from pylint==2.10.2)\n Downloading astroid-2.7.3-py3-none-any.whl (240 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 240.6/240.6 kB 46.7 MB/s eta 0:00:00\nCollecting isort<6,>=4.2.5 (from pylint==2.10.2)\n Downloading isort-5.13.2-py3-none-any.whl.metadata (12 kB)\nCollecting mccabe<0.7,>=0.6 (from pylint==2.10.2)\n Downloading mccabe-0.6.1-py2.py3-none-any.whl (8.6 kB)\nCollecting toml>=0.7.1 (from pylint==2.10.2)\n Downloading toml-0.10.2-py2.py3-none-any.whl (16 kB)\nCollecting lazy-object-proxy>=1.4.0 (from astroid<2.8,>=2.7.2->pylint==2.10.2)\n Downloading lazy_object_proxy-1.10.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.8 kB)\nCollecting wrapt<1.13,>=1.11 (from astroid<2.8,>=2.7.2->pylint==2.10.2)\n Downloading wrapt-1.12.1.tar.gz (27 kB)\n Installing build dependencies: started\n Installing build dependencies: finished with status 'done'\n Getting requirements to build wheel: started\n Getting requirements to build wheel: finished with status 'done'\n Preparing metadata (pyproject.toml): started\n Preparing metadata (pyproject.toml): finished with status 'done'\nRequirement already satisfied: setuptools>=20.0 in /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages (from astroid<2.8,>=2.7.2->pylint==2.10.2) (56.0.0)\nDownloading isort-5.13.2-py3-none-any.whl (92 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 92.3/92.3 kB 21.2 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading lazy_object_proxy-1.10.0-cp38-cp38-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (66 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 66.4/66.4 kB 16.7 MB/s eta 0:00:00\nBuilding wheels for collected packages: wrapt\n Building wheel for wrapt (pyproject.toml): started\n Building wheel for wrapt (pyproject.toml): finished with status 'done'\n Created wheel for wrapt: filename=wrapt-1.12.1-cp38-cp38-linux_x86_64.whl size=81820 sha256=be1d80d6ab9cc886aa44b352ed5ead61a5f3ec49ae66639ac53a214d9cb2af26\n Stored in directory: /home/runner/.cache/pip/wheels/5f/fd/9e/b6cf5890494cb8ef0b5eaff72e5d55a70fb56316007d6dfe73\nSuccessfully built wrapt\nInstalling collected packages: wrapt, mccabe, toml, platformdirs, lazy-object-proxy, isort, astroid, pylint\nSuccessfully installed astroid-2.7.3 isort-5.13.2 lazy-object-proxy-1.10.0 mccabe-0.6.1 platformdirs-4.1.0 pylint-2.10.2 toml-0.10.2 wrapt-1.12.1\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/astroid/raw_building.py:416: FutureWarning: In the future `np.long` will be defined as the corresponding NumPy scalar.\n getattr(sys.modules[modname], name)\n************* Module openvino.tools.accuracy_checker.evaluators.model_evaluator\nopenvino/tools/accuracy_checker/evaluators/model_evaluator.py:805:0: C0303: Trailing whitespace (trailing-whitespace)\n\n------------------------------------\nYour code has been rated at 10.00/10\n\n##[error]Process completed with exit code 16.\n"}], "diff": "diff --git a/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py b/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\nindex adb41abf5..5a310a979 100644\n--- a/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\n+++ b/tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py\n@@ -802,5 +802,4 @@ def get_config_metrics(config):\n # first subset_metrics or matching subsample_size\n metrics = item.get('metrics')\n break\n- \n return config.get('metrics',[]) if (metrics is None) else metrics\n", "difficulty": 0, "changed_files": ["tools/accuracy_checker/openvino/tools/accuracy_checker/evaluators/model_evaluator.py"], "commit_link": "https://github.com/openvinotoolkit/open_model_zoo/tree/07bc10c0e7858b22e9345812af8e6bb6c4ef18be"}
|
|
|
|
data/python/0828c8d.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/08c47c4.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 118, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "WB-11095", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "wandb", "sha_fail": "08c47c420629666a284cb013a39c907ac7451af5", "sha_success": "143c1ae682a944d2bfc8ebe9fb6ef7368cc58797", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.3-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 6.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 27.8 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 33.6 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 97.3 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 74.3 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.3-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 96.0 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.3 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.3\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCache Size: ~47 MB (48931996 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/b3bd0166-47b9-4cba-b7da-654927a1ab90/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48931996 of 48931996 (100.0%), 46.6 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nclippy...................................................................\u001b[42mPassed\u001b[m\nruff.....................................................................\u001b[42mPassed\u001b[m\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted tests/pytest_tests/unit_tests/test_public_api.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m887 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/tests/pytest_tests/unit_tests/test_public_api.py b/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[1mindex 6fb0799..bb0190b 100644\u001b[m\n\u001b[1m--- a/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[1m+++ b/tests/pytest_tests/unit_tests/test_public_api.py\u001b[m\n\u001b[36m@@ -97,6 +97,7 @@\u001b[m \u001b[mdef test_parse_path_proj():\u001b[m\n assert project == \"proj\"\u001b[m\n assert run == \"proj\"\u001b[m\n \u001b[m\n\u001b[32m+\u001b[m\n @pytest.mark.usefixtures(\"patch_apikey\", \"patch_prompt\")\u001b[m\n def test_parse_path_id():\u001b[m\n with mock.patch.dict(\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/tests/pytest_tests/unit_tests/test_public_api.py b/tests/pytest_tests/unit_tests/test_public_api.py\nindex 6fb079974..bb0190be6 100644\n--- a/tests/pytest_tests/unit_tests/test_public_api.py\n+++ b/tests/pytest_tests/unit_tests/test_public_api.py\n@@ -97,6 +97,7 @@ def test_parse_path_proj():\n assert project == \"proj\"\n assert run == \"proj\"\n \n+\n @pytest.mark.usefixtures(\"patch_apikey\", \"patch_prompt\")\n def test_parse_path_id():\n with mock.patch.dict(\n", "difficulty": 0, "changed_files": ["tests/pytest_tests/unit_tests/test_public_api.py"], "commit_link": "https://github.com/wandb/wandb/tree/08c47c420629666a284cb013a39c907ac7451af5"}
|
|
|
|
data/python/0b08b8e.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/0b5c0d2.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/0b93d2d.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/0d26cc1.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 38, "repo_owner": "beetbox", "repo_name": "beets", "head_branch": "master", "workflow_name": "Check code formatting", "workflow_filename": "formatting_check.yml", "workflow_path": ".github/workflows/formatting_check.yml", "contributor": "niacdoial", "sha_fail": "0d26cc1482ff5080ec579b17b29f22657a20c562", "sha_success": "828c6e7dbd1a1fe77b4b7be61aa2e3d220ff86d1", "workflow": "name: formatting_check\nrun-name: Check code formatting\non: [push, pull_request]\njobs:\n formatting_check:\n runs-on: ubuntu-latest\n steps:\n - name: Install dependencies\n uses: actions/checkout@v3\n\n - name: Run formatting check\n uses: paolorechia/[email protected]\n with:\n tox_env: \"format_check\"\n", "logs": [{"step_name": "formatting_check/3_Run formatting check.txt", "log": "##[group]Run paolorechia/[email protected]\nwith:\n tox_env: format_check\n##[endgroup]\nstdout: Defaulting to user installation because normal site-packages is not writeable\n\nstdout: Collecting tox\n\nstdout: Downloading tox-4.11.4-py3-none-any.whl (153 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 KB 4.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Requirement already satisfied: platformdirs>=3.10 in /usr/local/lib/python3.10/dist-packages (from tox) (4.1.0)\n\nstdout: Collecting pluggy>=1.3\n\nstdout: Downloading pluggy-1.3.0-py3-none-any.whl (18 kB)\n\nstdout: Collecting virtualenv>=20.24.3\n\nstdout: Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 12.0 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting cachetools>=5.3.1\n\nstdout: Downloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\n\nstdout: Requirement already satisfied: packaging>=23.1 in /usr/local/lib/python3.10/dist-packages (from tox) (23.2)\n\nstdout: Requirement already satisfied: tomli>=2.0.1 in /usr/local/lib/python3.10/dist-packages (from tox) (2.0.1)\n\nstdout: Collecting filelock>=3.12.3\n\nstdout: Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\n\nstdout: Collecting colorama>=0.4.6\n\nstdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\n\nstdout: Collecting pyproject-api>=1.6.1\n\nstdout: Downloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\n\nstdout: Collecting chardet>=5.2\n\nstdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 KB 21.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting distlib<1,>=0.3.7\n\nstdout: Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 15.4 MB/s eta 0:00:00\nstdout: \n\nstdout: Installing collected packages: distlib, pyproject-api, pluggy, filelock, colorama, chardet, cachetools, virtualenv, tox\n\nstdout: Successfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\nReceived 6565789 of 6565789 (100.0%), 18.9 MBs/sec\nCache Size: ~6 MB (6565789 B)\n[command]/usr/bin/tar -z -xf /home/runner/work/_temp/d6197f94-ef69-4efe-9ad2-cb7265f9e34d/cache.tgz -P -C /home/runner/work/beets/beets\nCache restored successfully\nHit cache on tox\nstdout: format_check: recreate env because python changed virtualenv version='20.24.5'->'20.25.0'\n\nstdout: format_check: remove tox env folder /home/runner/work/beets/beets/.tox/format_check\n\nstdout: format_check: install_deps> python -I -m pip install black isort\n\nstdout: format_check: commands[0]> isort beets beetsplug test --check\n\nstderr: ERROR: /home/runner/work/beets/beets/beetsplug/edit.py Imports are incorrectly sorted and/or formatted.\n\nstdout: format_check: exit 1 (0.54 seconds) /home/runner/work/beets/beets> isort beets beetsplug test --check pid=1759\n\nstdout: format_check: FAIL code 1 (3.49=setup[2.95]+cmd[0.54] seconds)\n evaluation failed :( (3.64 seconds)\n\nchild process exited with code 1\n##[warning]The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/\n##[error]Tox failed\n"}], "diff": "diff --git a/beetsplug/advancedrewrite.py b/beetsplug/advancedrewrite.py\nindex 639aa5247..20f2b7e03 100644\n--- a/beetsplug/advancedrewrite.py\n+++ b/beetsplug/advancedrewrite.py\n@@ -161,7 +161,7 @@ class AdvancedRewritePlugin(BeetsPlugin):\n )\n elif isinstance(replacement, str):\n if Item._fields[fieldname] is MULTI_VALUE_DSV:\n- replacement = list(replacement)\n+ replacement = [replacement]\n else:\n raise UserError(\n f\"Invalid type of replacement {replacement} \"\ndiff --git a/beetsplug/edit.py b/beetsplug/edit.py\nindex 8a07fccaa..323dd9e41 100644\n--- a/beetsplug/edit.py\n+++ b/beetsplug/edit.py\n@@ -23,7 +23,7 @@ from tempfile import NamedTemporaryFile\n \n import yaml\n \n-from beets import plugins, ui, util, config\n+from beets import plugins, ui, util\n from beets.dbcore import types\n from beets.importer import action\n from beets.ui.commands import PromptChoice, _do_query\n@@ -155,8 +155,6 @@ class EditPlugin(plugins.BeetsPlugin):\n }\n )\n \n- self.has_shown_ui = False\n-\n self.register_listener(\n \"before_choose_candidate\", self.before_choose_candidate_listener\n )\n@@ -238,23 +236,11 @@ class EditPlugin(plugins.BeetsPlugin):\n # Get the content to edit as raw data structures.\n old_data = [flatten(o, fields) for o in objs]\n \n- # take set fields into account\n- set_fields = config[\"import\"][\"set_fields\"]\n- if set_fields and not self.has_shown_ui:\n- old_str = \"\\n\\n# note: the following fields will be reset to their current values:\\n\"\n- for key in set_fields:\n- old_str += f\"# - {key}\\n\"\n- for obj in old_data:\n- # those values will be enforced later anyway\n- obj.update({k:v.get() for k,v in set_fields.items()})\n- else:\n- old_str = \"\"\n-\n # Set up a temporary file with the initial data for editing.\n new = NamedTemporaryFile(\n mode=\"w\", suffix=\".yaml\", delete=False, encoding=\"utf-8\"\n )\n- old_str = dump(old_data) + old_str\n+ old_str = dump(old_data)\n new.write(old_str)\n new.close()\n \n@@ -263,7 +249,6 @@ class EditPlugin(plugins.BeetsPlugin):\n while True:\n # Ask the user to edit the data.\n edit(new.name, self._log)\n- self.has_shown_ui = True\n \n # Read the data back after editing and check whether anything\n # changed.\ndiff --git a/beetsplug/web/__init__.py b/beetsplug/web/__init__.py\nindex cebb0be0a..cd7e8a3fc 100644\n--- a/beetsplug/web/__init__.py\n+++ b/beetsplug/web/__init__.py\n@@ -336,7 +336,6 @@ def item_file(item_id):\n response = flask.send_file(\n item_path, as_attachment=True, download_name=safe_filename\n )\n- response.headers[\"Content-Length\"] = os.path.getsize(item_path)\n return response\n \n \ndiff --git a/docs/changelog.rst b/docs/changelog.rst\nindex 73cc10f3d..b3b63f1c2 100644\n--- a/docs/changelog.rst\n+++ b/docs/changelog.rst\n@@ -279,6 +279,8 @@ Bug fixes:\n :bug:`4973`\n * Fix bug regarding displaying tracks that have been changed not being\n displayed unless the detail configuration is enabled.\n+* :doc:`/plugins/web`: Fix range request support, allowing to play large audio/\n+ opus files using e.g. a browser/firefox or gstreamer/mopidy directly.\n \n For plugin developers:\n \n", "difficulty": 0, "changed_files": ["beetsplug/advancedrewrite.py", "beetsplug/edit.py", "beetsplug/web/__init__.py", "docs/changelog.rst"], "commit_link": "https://github.com/beetbox/beets/tree/0d26cc1482ff5080ec579b17b29f22657a20c562"}
|
|
|
|
data/python/0d367ce.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/0f71221.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 20, "repo_owner": "scrapy", "repo_name": "scrapy", "head_branch": "component-getters", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "Gallaecio", "sha_fail": "0f71221cf9875ed8ef3400e1008408e79b6691e6", "sha_success": "c1ba9ccdf916b89d875628ba143dc5c9f6977430", "workflow": "name: Checks\non: [push, pull_request]\n\nconcurrency:\n group: ${{github.workflow}}-${{ github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n include:\n - python-version: \"3.12\"\n env:\n TOXENV: pylint\n - python-version: 3.8\n env:\n TOXENV: typing\n - python-version: \"3.11\" # Keep in sync with .readthedocs.yml\n env:\n TOXENV: docs\n - python-version: \"3.12\"\n env:\n TOXENV: twinecheck\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Run check\n env: ${{ matrix.env }}\n run: |\n pip install -U tox\n tox\n\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "checks (3.12, pylint)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n TOXENV: pylint\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 16.6 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 48.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 16.1 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 104.2 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 81.3 MB/s eta 0:00:00\nInstalling collected packages: distlib, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.3.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\npylint: install_deps> python -I -m pip install -ctests/upper-constraints.txt boto3 google-cloud-storage 'markupsafe<2.1.0' 'mitmproxy<8,>=4.0.4; python_version < \"3.9\" and implementation_name != \"pypy\"' Pillow pylint==3.0.1 robotexclusionrulesparser 'Twisted[http2]' -r tests/requirements.txt\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\npylint: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\npylint: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\npylint: freeze> python -m pip freeze --all\npylint: astroid==3.0.2,asttokens==2.4.1,attrs==23.1.0,Automat==22.10.0,blessed==1.20.0,boto3==1.34.3,botocore==1.34.3,bpython==0.24,Brotli==1.1.0,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,coverage==7.3.3,cryptography==41.0.7,cssselect==1.2.0,curtsies==0.4.2,cwcwidth==0.1.9,decorator==5.1.1,dill==0.3.7,execnet==2.0.2,executing==2.0.1,filelock==3.13.1,google-api-core==2.15.0,google-auth==2.25.2,google-cloud-core==2.4.1,google-cloud-storage==2.14.0,google-crc32c==1.5.0,google-resumable-media==2.7.0,googleapis-common-protos==1.62.0,greenlet==3.0.2,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,iniconfig==2.0.0,ipython==8.18.1,isort==5.13.2,itemadapter==0.8.0,itemloaders==1.1.0,jedi==0.19.1,jmespath==1.0.1,lxml==4.9.3,MarkupSafe==2.0.1,matplotlib-inline==0.1.6,mccabe==0.7.0,packaging==23.2,parsel==1.8.1,parso==0.8.3,pexpect==4.9.0,Pillow==10.1.0,pip==23.3.2,platformdirs==4.1.0,pluggy==1.3.0,priority==1.3.0,prompt-toolkit==3.0.43,Protego==0.3.0,protobuf==4.25.1,ptyprocess==0.7.0,pure-eval==0.2.2,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,PyDispatcher==2.0.7,pyftpdlib==1.5.9,Pygments==2.17.2,pylint==3.0.1,pyOpenSSL==23.3.0,pytest==7.4.3,pytest-cov==4.0.0,pytest-xdist==3.5.0,python-dateutil==2.8.2,pyxdg==0.28,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,robotexclusionrulesparser==1.7.1,rsa==4.9,s3transfer==0.9.0,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=6f43aa5d37ec10736bd947dbba89df54ebc97175083ab360a721b3360f0d895b,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,stack-data==0.6.3,sybil==6.0.2,testfixtures==7.2.2,tldextract==5.1.1,tomlkit==0.12.3,traitlets==5.14.0,Twisted==23.10.0,typing_extensions==4.9.0,urllib3==2.0.7,uvloop==0.19.0,w3lib==2.1.2,wcwidth==0.2.12,zope.interface==6.1,zstandard==0.22.0\npylint: commands[0]> pylint conftest.py docs extras scrapy setup.py tests\n<unknown>:230: SyntaxWarning: invalid escape sequence '\\d'\n************* Module scrapy.crawler\nscrapy/crawler.py:182:4: W0211: Static method with 'cls' as first argument (bad-staticmethod-argument)\nscrapy/crawler.py:184:15: C0123: Use isinstance() rather than type() for a typecheck. (unidiomatic-typecheck)\n\n------------------------------------\nYour code has been rated at 10.00/10\n\npylint: exit 20 (56.60 seconds) /home/runner/work/scrapy/scrapy> pylint conftest.py docs extras scrapy setup.py tests pid=2086\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n pylint: FAIL code 20 (100.72=setup[44.12]+cmd[56.60] seconds)\n evaluation failed :( (100.86 seconds)\n##[error]Process completed with exit code 20.\n"}], "diff": "diff --git a/scrapy/crawler.py b/scrapy/crawler.py\nindex 76f4048e4..333b39f33 100644\n--- a/scrapy/crawler.py\n+++ b/scrapy/crawler.py\n@@ -179,9 +179,12 @@ class Crawler:\n yield maybeDeferred(self.engine.stop)\n \n @staticmethod\n- def _get_component(cls, components):\n+ def _get_component(component_class, components):\n for component in components:\n- if type(component) is cls:\n+ if (\n+ type(component) # pylint: disable=unidiomatic-typecheck\n+ is component_class\n+ ):\n return component\n return None\n \n", "difficulty": 0, "changed_files": ["scrapy/crawler.py"], "commit_link": "https://github.com/scrapy/scrapy/tree/0f71221cf9875ed8ef3400e1008408e79b6691e6"}
|
|
|
|
data/python/102f918.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 21, "repo_owner": "mindsdb", "repo_name": "mindsdb", "head_branch": "staging", "workflow_name": "MindsDB Code Checks", "workflow_filename": "code_checks.yml", "workflow_path": ".github/workflows/code_checks.yml", "contributor": "mindsdb", "sha_fail": "102f918deb2532bb7b825f00258f2c1414cf94da", "sha_success": "1cf75d7a1071fba82d4190725e984e8d55db48fb", "workflow": "name: MindsDB Code Checks\n\non:\n pull_request:\n branches: [stable, staging]\n \njobs:\n check_requirements:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/[email protected]\n with:\n python-version: 3.9\n - name: Check main requirements\n shell: bash\n run: |\n pip install -r requirements/requirements-dev.txt\n\n python tests/scripts/check_requirements.py\n\n - name: Check requirements files are installable\n shell: bash\n run: |\n pip install --dry-run --ignore-installed . # Install only the default handlers. We can expand this to all handlers later with: .[all_handlers_extras]\n", "logs": [{"step_name": "check_requirements/4_Check main requirements.txt", "log": "##[group]Run pip install -r requirements/requirements-dev.txt\n\u001b[36;1mpip install -r requirements/requirements-dev.txt\u001b[0m\n\u001b[36;1m\u001b[0m\n\u001b[36;1mpython tests/scripts/check_requirements.py\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nCollecting black==23.3.0 (from -r requirements/requirements-dev.txt (line 1))\n Downloading black-23.3.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.7 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.7/1.7 MB 18.4 MB/s eta 0:00:00\nCollecting flake8==5.0.4 (from -r requirements/requirements-dev.txt (line 2))\n Downloading flake8-5.0.4-py2.py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.9/61.9 kB 19.0 MB/s eta 0:00:00\nCollecting isort==5.10.1 (from -r requirements/requirements-dev.txt (line 3))\n Downloading isort-5.10.1-py3-none-any.whl (103 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 103.4/103.4 kB 26.5 MB/s eta 0:00:00\nCollecting pre-commit>=2.16.0 (from -r requirements/requirements-dev.txt (line 4))\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting watchfiles==0.19.0 (from -r requirements/requirements-dev.txt (line 5))\n Downloading watchfiles-0.19.0-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.3/1.3 MB 31.1 MB/s eta 0:00:00\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from -r requirements/requirements-dev.txt (line 6)) (58.1.0)\nCollecting wheel (from -r requirements/requirements-dev.txt (line 7))\n Downloading wheel-0.42.0-py3-none-any.whl.metadata (2.2 kB)\nCollecting deptry==0.12.0 (from -r requirements/requirements-dev.txt (line 8))\n Downloading deptry-0.12.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting twine (from -r requirements/requirements-dev.txt (line 9))\n Downloading twine-4.0.2-py3-none-any.whl (36 kB)\nCollecting click>=8.0.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading click-8.1.7-py3-none-any.whl.metadata (3.0 kB)\nCollecting mypy-extensions>=0.4.3 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading mypy_extensions-1.0.0-py3-none-any.whl (4.7 kB)\nCollecting packaging>=22.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting pathspec>=0.9.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading pathspec-0.12.1-py3-none-any.whl.metadata (21 kB)\nCollecting platformdirs>=2 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting tomli>=1.1.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting typing-extensions>=3.10.0.0 (from black==23.3.0->-r requirements/requirements-dev.txt (line 1))\n Downloading typing_extensions-4.9.0-py3-none-any.whl.metadata (3.0 kB)\nCollecting mccabe<0.8.0,>=0.7.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading mccabe-0.7.0-py2.py3-none-any.whl (7.3 kB)\nCollecting pycodestyle<2.10.0,>=2.9.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading pycodestyle-2.9.1-py2.py3-none-any.whl (41 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.5/41.5 kB 11.9 MB/s eta 0:00:00\nCollecting pyflakes<2.6.0,>=2.5.0 (from flake8==5.0.4->-r requirements/requirements-dev.txt (line 2))\n Downloading pyflakes-2.5.0-py2.py3-none-any.whl (66 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 66.1/66.1 kB 20.4 MB/s eta 0:00:00\nCollecting anyio>=3.0.0 (from watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading anyio-4.2.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting chardet>=4.0.0 (from deptry==0.12.0->-r requirements/requirements-dev.txt (line 8))\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting cfgv>=2.0.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting pkginfo>=1.8.1 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pkginfo-1.9.6-py3-none-any.whl (30 kB)\nCollecting readme-renderer>=35.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading readme_renderer-42.0-py3-none-any.whl.metadata (2.8 kB)\nCollecting requests>=2.20 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting requests-toolbelt!=0.9.0,>=0.8.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading requests_toolbelt-1.0.0-py2.py3-none-any.whl (54 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 17.0 MB/s eta 0:00:00\nCollecting urllib3>=1.26.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading urllib3-2.1.0-py3-none-any.whl.metadata (6.4 kB)\nCollecting importlib-metadata>=3.6 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading importlib_metadata-7.0.1-py3-none-any.whl.metadata (4.9 kB)\nCollecting keyring>=15.1 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading keyring-24.3.0-py3-none-any.whl.metadata (20 kB)\nCollecting rfc3986>=1.4.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading rfc3986-2.0.0-py2.py3-none-any.whl (31 kB)\nCollecting rich>=12.0.0 (from twine->-r requirements/requirements-dev.txt (line 9))\n Downloading rich-13.7.0-py3-none-any.whl.metadata (18 kB)\nCollecting idna>=2.8 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading idna-3.6-py3-none-any.whl.metadata (9.9 kB)\nCollecting sniffio>=1.1 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading sniffio-1.3.0-py3-none-any.whl (10 kB)\nCollecting exceptiongroup>=1.0.2 (from anyio>=3.0.0->watchfiles==0.19.0->-r requirements/requirements-dev.txt (line 5))\n Downloading exceptiongroup-1.2.0-py3-none-any.whl.metadata (6.6 kB)\nCollecting zipp>=0.5 (from importlib-metadata>=3.6->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading zipp-3.17.0-py3-none-any.whl.metadata (3.7 kB)\nCollecting jaraco.classes (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading jaraco.classes-3.3.0-py3-none-any.whl.metadata (2.9 kB)\nCollecting SecretStorage>=3.2 (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading SecretStorage-3.3.3-py3-none-any.whl (15 kB)\nCollecting jeepney>=0.4.2 (from keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading jeepney-0.8.0-py3-none-any.whl (48 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 48.4/48.4 kB 17.1 MB/s eta 0:00:00\nCollecting nh3>=0.2.14 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading nh3-0.2.15-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.7 kB)\nCollecting docutils>=0.13.1 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading docutils-0.20.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting Pygments>=2.5.1 (from readme-renderer>=35.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pygments-2.17.2-py3-none-any.whl.metadata (2.6 kB)\nCollecting charset-normalizer<4,>=2 (from requests>=2.20->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (33 kB)\nCollecting certifi>=2017.4.17 (from requests>=2.20->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading certifi-2023.11.17-py3-none-any.whl.metadata (2.2 kB)\nCollecting markdown-it-py>=2.2.0 (from rich>=12.0.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading markdown_it_py-3.0.0-py3-none-any.whl.metadata (6.9 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit>=2.16.0->-r requirements/requirements-dev.txt (line 4))\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=12.0.0->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading mdurl-0.1.2-py3-none-any.whl (10.0 kB)\nCollecting cryptography>=2.0 (from SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading cryptography-41.0.7-cp37-abi3-manylinux_2_28_x86_64.whl.metadata (5.2 kB)\nCollecting more-itertools (from jaraco.classes->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading more_itertools-10.1.0-py3-none-any.whl.metadata (33 kB)\nCollecting cffi>=1.12 (from cryptography>=2.0->SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading cffi-1.16.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.5 kB)\nCollecting pycparser (from cffi>=1.12->cryptography>=2.0->SecretStorage>=3.2->keyring>=15.1->twine->-r requirements/requirements-dev.txt (line 9))\n Downloading pycparser-2.21-py2.py3-none-any.whl (118 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 118.7/118.7 kB 35.3 MB/s eta 0:00:00\nDownloading deptry-0.12.0-py3-none-any.whl (41 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.7/41.7 kB 13.1 MB/s eta 0:00:00\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 42.9 MB/s eta 0:00:00\nDownloading wheel-0.42.0-py3-none-any.whl (65 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 65.4/65.4 kB 22.9 MB/s eta 0:00:00\nDownloading anyio-4.2.0-py3-none-any.whl (85 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.5/85.5 kB 29.1 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 43.8 MB/s eta 0:00:00\nDownloading click-8.1.7-py3-none-any.whl (97 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 97.9/97.9 kB 33.7 MB/s eta 0:00:00\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.2 MB/s eta 0:00:00\nDownloading importlib_metadata-7.0.1-py3-none-any.whl (23 kB)\nDownloading keyring-24.3.0-py3-none-any.whl (38 kB)\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 18.2 MB/s eta 0:00:00\nDownloading pathspec-0.12.1-py3-none-any.whl (31 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (738 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 738.9/738.9 kB 37.6 MB/s eta 0:00:00\nDownloading readme_renderer-42.0-py3-none-any.whl (13 kB)\nDownloading requests-2.31.0-py3-none-any.whl (62 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 62.6/62.6 kB 21.9 MB/s eta 0:00:00\nDownloading rich-13.7.0-py3-none-any.whl (240 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 240.6/240.6 kB 51.0 MB/s eta 0:00:00\nDownloading typing_extensions-4.9.0-py3-none-any.whl (32 kB)\nDownloading urllib3-2.1.0-py3-none-any.whl (104 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.6/104.6 kB 33.0 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 53.1 MB/s eta 0:00:00\nDownloading certifi-2023.11.17-py3-none-any.whl (162 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 162.5/162.5 kB 43.3 MB/s eta 0:00:00\nDownloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (142 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 142.3/142.3 kB 40.8 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 68.6 MB/s eta 0:00:00\nDownloading docutils-0.20.1-py3-none-any.whl (572 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 572.7/572.7 kB 67.0 MB/s eta 0:00:00\nDownloading exceptiongroup-1.2.0-py3-none-any.whl (16 kB)\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading idna-3.6-py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.6/61.6 kB 20.6 MB/s eta 0:00:00\nDownloading markdown_it_py-3.0.0-py3-none-any.whl (87 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 87.5/87.5 kB 28.4 MB/s eta 0:00:00\nDownloading nh3-0.2.15-cp37-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.7 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.7/1.7 MB 72.8 MB/s eta 0:00:00\nDownloading pygments-2.17.2-py3-none-any.whl (1.2 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.2/1.2 MB 81.6 MB/s eta 0:00:00\nDownloading zipp-3.17.0-py3-none-any.whl (7.4 kB)\nDownloading jaraco.classes-3.3.0-py3-none-any.whl (5.9 kB)\nDownloading cryptography-41.0.7-cp37-abi3-manylinux_2_28_x86_64.whl (4.4 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.4/4.4 MB 94.7 MB/s eta 0:00:00\nDownloading more_itertools-10.1.0-py3-none-any.whl (55 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 55.8/55.8 kB 17.6 MB/s eta 0:00:00\nDownloading cffi-1.16.0-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (443 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 443.4/443.4 kB 72.9 MB/s eta 0:00:00\nInstalling collected packages: nh3, distlib, zipp, wheel, urllib3, typing-extensions, tomli, sniffio, rfc3986, pyyaml, Pygments, pyflakes, pycparser, pycodestyle, platformdirs, pkginfo, pathspec, packaging, nodeenv, mypy-extensions, more-itertools, mdurl, mccabe, jeepney, isort, idna, identify, filelock, exceptiongroup, docutils, click, charset-normalizer, chardet, cfgv, certifi, virtualenv, requests, readme-renderer, markdown-it-py, jaraco.classes, importlib-metadata, flake8, deptry, cffi, black, anyio, watchfiles, rich, requests-toolbelt, pre-commit, cryptography, SecretStorage, keyring, twine\nSuccessfully installed Pygments-2.17.2 SecretStorage-3.3.3 anyio-4.2.0 black-23.3.0 certifi-2023.11.17 cffi-1.16.0 cfgv-3.4.0 chardet-5.2.0 charset-normalizer-3.3.2 click-8.1.7 cryptography-41.0.7 deptry-0.12.0 distlib-0.3.8 docutils-0.20.1 exceptiongroup-1.2.0 filelock-3.13.1 flake8-5.0.4 identify-2.5.33 idna-3.6 importlib-metadata-7.0.1 isort-5.10.1 jaraco.classes-3.3.0 jeepney-0.8.0 keyring-24.3.0 markdown-it-py-3.0.0 mccabe-0.7.0 mdurl-0.1.2 more-itertools-10.1.0 mypy-extensions-1.0.0 nh3-0.2.15 nodeenv-1.8.0 packaging-23.2 pathspec-0.12.1 pkginfo-1.9.6 platformdirs-4.1.0 pre-commit-3.6.0 pycodestyle-2.9.1 pycparser-2.21 pyflakes-2.5.0 pyyaml-6.0.1 readme-renderer-42.0 requests-2.31.0 requests-toolbelt-1.0.0 rfc3986-2.0.0 rich-13.7.0 sniffio-1.3.0 tomli-2.0.1 twine-4.0.2 typing-extensions-4.9.0 urllib3-2.1.0 virtualenv-20.25.0 watchfiles-0.19.0 wheel-0.42.0 zipp-3.17.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\n--- Checking requirements files for duplicates ---\n\n--- Checking that requirements match imports ---\n- requirements/requirements.txt\n None:None: DEP002 'type_infer' defined as a dependency but not used in the codebase\n\n\n--- Checking handlers that require other handlers ---\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/mindsdb/__main__.py b/mindsdb/__main__.py\nindex cac3482883..6cc92a7899 100644\n--- a/mindsdb/__main__.py\n+++ b/mindsdb/__main__.py\n@@ -35,6 +35,7 @@ from mindsdb.utilities.fs import create_dirs_recursive, clean_process_marks, cle\n from mindsdb.utilities.telemetry import telemetry_file_exists, disable_telemetry\n from mindsdb.utilities.context import context as ctx\n from mindsdb.utilities.auth import register_oauth_client, get_aws_meta_data\n+import type_infer # noqa\n \n try:\n import torch.multiprocessing as mp\n", "difficulty": 1, "changed_files": ["mindsdb/__main__.py"], "commit_link": "https://github.com/mindsdb/mindsdb/tree/102f918deb2532bb7b825f00258f2c1414cf94da"}
|
|
|
|
data/python/1168b92.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 131, "repo_owner": "scrapy", "repo_name": "scrapy", "head_branch": "simplify-attempt2", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "monicaq21", "sha_fail": "1168b9244d680437e2a2683294a55c0b52c118f6", "sha_success": "b5b7c99e24c722fa09c426ed0245df1ae792d486", "workflow": "name: Checks\non: [push, pull_request]\n\nconcurrency:\n group: ${{github.workflow}}-${{ github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n include:\n - python-version: \"3.12\"\n env:\n TOXENV: pylint\n - python-version: 3.8\n env:\n TOXENV: typing\n - python-version: \"3.11\" # Keep in sync with .readthedocs.yml\n env:\n TOXENV: docs\n - python-version: \"3.12\"\n env:\n TOXENV: twinecheck\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Run check\n env: ${{ matrix.env }}\n run: |\n pip install -U tox\n tox\n\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "checks (3.12, pylint)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n TOXENV: pylint\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 19.9 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 51.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 16.5 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 117.4 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 82.7 MB/s eta 0:00:00\nInstalling collected packages: distlib, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.7 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\npylint: install_deps> python -I -m pip install -ctests/upper-constraints.txt boto3 google-cloud-storage 'markupsafe<2.1.0' 'mitmproxy<8,>=4.0.4; python_version < \"3.9\" and implementation_name != \"pypy\"' Pillow pylint==3.0.1 robotexclusionrulesparser 'Twisted[http2]' -r tests/requirements.txt\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\npylint: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\npylint: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\npylint: freeze> python -m pip freeze --all\npylint: annotated-types==0.6.0,astroid==3.0.1,asttokens==2.4.1,attrs==23.1.0,Automat==22.10.0,blessed==1.20.0,boto3==1.33.11,botocore==1.33.11,bpython==0.24,Brotli==1.1.0,cachetools==5.3.2,Cerberus==1.3.5,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,coverage==7.3.2,cryptography==41.0.7,cssselect==1.2.0,curtsies==0.4.2,cwcwidth==0.1.9,decorator==5.1.1,dill==0.3.7,distlib==0.3.7,docopt==0.6.2,execnet==2.0.2,executing==2.0.1,filelock==3.13.1,google-api-core==2.15.0,google-auth==2.25.2,google-cloud-core==2.4.1,google-cloud-storage==2.13.0,google-crc32c==1.5.0,google-resumable-media==2.6.0,googleapis-common-protos==1.62.0,greenlet==3.0.2,h2==4.1.0,hpack==4.0.0,hyperframe==6.0.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,iniconfig==2.0.0,ipython==8.18.1,isort==5.13.0,itemadapter==0.8.0,itemloaders==1.1.0,jedi==0.19.1,jmespath==1.0.1,lxml==4.9.3,MarkupSafe==2.0.1,matplotlib-inline==0.1.6,mccabe==0.7.0,packaging==23.2,parsel==1.8.1,parso==0.8.3,pep517==0.13.1,pexpect==4.9.0,Pillow==10.1.0,pip==23.3.1,pip-api==0.0.30,pipreqs==0.4.13,platformdirs==4.1.0,plette==0.4.4,pluggy==1.3.0,priority==1.3.0,prompt-toolkit==3.0.41,Protego==0.3.0,protobuf==4.25.1,ptyprocess==0.7.0,pure-eval==0.2.2,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,pydantic==2.5.2,pydantic_core==2.14.5,PyDispatcher==2.0.7,pyftpdlib==1.5.9,Pygments==2.17.2,pylint==3.0.1,pyOpenSSL==23.3.0,pytest==7.4.3,pytest-cov==4.0.0,pytest-xdist==3.5.0,python-dateutil==2.8.2,pyxdg==0.28,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,requirementslib==3.0.0,robotexclusionrulesparser==1.7.1,rsa==4.9,s3transfer==0.8.2,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=badae26f961983adc76ef32b8b07c6ba5b979a31f891e5f2e57590381cc5b815,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,stack-data==0.6.3,sybil==6.0.2,testfixtures==7.2.2,tldextract==5.1.1,tomlkit==0.12.3,traitlets==5.14.0,Twisted==23.10.0,typing_extensions==4.9.0,urllib3==2.0.7,uvloop==0.19.0,w3lib==2.1.2,wcwidth==0.2.12,yarg==0.1.9,zope.interface==6.1,zstandard==0.22.0\npylint: commands[0]> pylint conftest.py docs extras scrapy setup.py tests\n<unknown>:230: SyntaxWarning: invalid escape sequence '\\d'\n************* Module scrapy.crawler\nscrapy/crawler.py:112:37: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.engine\nscrapy/core/engine.py:361:20: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.scheduler\nscrapy/core/scheduler.py:325:15: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/scheduler.py:336:12: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.contextfactory\nscrapy/core/downloader/contextfactory.py:168:26: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:168:26: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:174:26: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/contextfactory.py:174:26: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers\nscrapy/core/downloader/handlers/__init__.py:58:17: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/__init__.py:58:17: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers.http10\nscrapy/core/downloader/handlers/http10.py:33:37: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/http10.py:33:37: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module scrapy.core.downloader.handlers.s3\nscrapy/core/downloader/handlers/s3.py:53:24: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\nscrapy/core/downloader/handlers/s3.py:53:24: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module tests.test_downloader_handlers\ntests/test_downloader_handlers.py:831:22: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:831:22: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:861:17: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:861:17: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:888:12: W1117: 'objcls' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_downloader_handlers.py:888:12: W1117: 'crawler' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n************* Module tests.test_webclient\ntests/test_webclient.py:473:33: W1117: 'settings' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\ntests/test_webclient.py:485:33: W1117: 'settings' will be included in '**kwargs' since a positional-only parameter with this name already exists (kwarg-superseded-by-positional-arg)\n\n-----------------------------------\nYour code has been rated at 9.99/10\n\npylint: exit 4 (53.83 seconds) /home/runner/work/scrapy/scrapy> pylint conftest.py docs extras scrapy setup.py tests pid=2037\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.0/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n pylint: FAIL code 4 (100.80=setup[46.97]+cmd[53.83] seconds)\n evaluation failed :( (100.88 seconds)\n##[error]Process completed with exit code 4.\n"}, {"step_name": "checks (3.8, typing)/4_Run check.txt", "log": "##[group]Run pip install -U tox\n\u001b[36;1mpip install -U tox\u001b[0m\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n TOXENV: typing\n##[endgroup]\nCollecting tox\n Downloading tox-4.11.4-py3-none-any.whl.metadata (5.0 kB)\nCollecting cachetools>=5.3.1 (from tox)\n Downloading cachetools-5.3.2-py3-none-any.whl.metadata (5.2 kB)\nCollecting chardet>=5.2 (from tox)\n Downloading chardet-5.2.0-py3-none-any.whl.metadata (3.4 kB)\nCollecting colorama>=0.4.6 (from tox)\n Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\nCollecting filelock>=3.12.3 (from tox)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting packaging>=23.1 (from tox)\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nCollecting platformdirs>=3.10 (from tox)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting pluggy>=1.3 (from tox)\n Downloading pluggy-1.3.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting pyproject-api>=1.6.1 (from tox)\n Downloading pyproject_api-1.6.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting tomli>=2.0.1 (from tox)\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting virtualenv>=20.24.3 (from tox)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.24.3->tox)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nDownloading tox-4.11.4-py3-none-any.whl (153 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 kB 8.4 MB/s eta 0:00:00\nDownloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\nDownloading chardet-5.2.0-py3-none-any.whl (199 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 kB 32.2 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 17.1 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pluggy-1.3.0-py3-none-any.whl (18 kB)\nDownloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 82.9 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 79.6 MB/s eta 0:00:00\nInstalling collected packages: distlib, tomli, pluggy, platformdirs, packaging, filelock, colorama, chardet, cachetools, virtualenv, pyproject-api, tox\nSuccessfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.7 filelock-3.13.1 packaging-23.2 platformdirs-4.1.0 pluggy-1.3.0 pyproject-api-1.6.1 tomli-2.0.1 tox-4.11.4 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.1\n[notice] To update, run: pip install --upgrade pip\ntyping: install_deps> python -I -m pip install -ctests/upper-constraints.txt mypy==1.6.1 types-attrs==19.1.0 types-lxml==2023.10.21 types-Pillow==10.1.0.0 types-Pygments==2.16.0.0 types-pyOpenSSL==23.3.0.0 types-setuptools==68.2.0.0 typing-extensions==4.8.0 'w3lib>=2.1.2'\n.pkg: install_requires> python -I -m pip install 'setuptools>=40.8.0' wheel\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: get_requires_for_build_wheel> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: install_requires_for_build_wheel> python -I -m pip install wheel\n.pkg: freeze> python -m pip freeze --all\n.pkg: pip==23.3.1,setuptools==69.0.2,wheel==0.42.0\n.pkg: prepare_metadata_for_build_wheel> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\ntyping: install_package_deps> python -I -m pip install -ctests/upper-constraints.txt 'PyDispatcher>=2.0.5; platform_python_implementation == \"CPython\"' 'PyPyDispatcher>=2.1.0; platform_python_implementation == \"PyPy\"' 'Twisted>=18.9.0' 'cryptography>=36.0.0' 'cssselect>=0.9.1' 'itemadapter>=0.1.0' 'itemloaders>=1.0.1' 'lxml>=4.4.1' packaging 'parsel>=1.5.0' 'protego>=0.1.15' 'pyOpenSSL>=21.0.0' 'queuelib>=1.4.2' 'service-identity>=18.1.0' setuptools tldextract 'w3lib>=1.17.0' 'zope.interface>=5.1.0'\ntyping: install_package> python -I -m pip install -ctests/upper-constraints.txt --force-reinstall --no-deps /home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz\ntyping: freeze> python -m pip freeze --all\ntyping: attrs==23.1.0,Automat==22.10.0,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,constantly==23.10.4,cryptography==41.0.7,cssselect==1.2.0,filelock==3.13.1,hyperlink==21.0.0,idna==3.6,incremental==22.10.0,itemadapter==0.8.0,itemloaders==1.1.0,jmespath==1.0.1,lxml==4.9.3,mypy==1.6.1,mypy-extensions==1.0.0,packaging==23.2,parsel==1.8.1,pip==23.3.1,Protego==0.3.0,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycparser==2.21,PyDispatcher==2.0.7,pyOpenSSL==23.3.0,queuelib==1.6.2,requests==2.31.0,requests-file==1.5.1,Scrapy @ file:///home/runner/work/scrapy/scrapy/.tox/.tmp/package/1/Scrapy-2.11.0.tar.gz#sha256=ebe501ee4a51f734f99dff82a0f62429c4051d3976e2377b4c07e9f5fb55d533,service-identity==23.1.0,setuptools==69.0.2,six==1.16.0,tldextract==5.1.1,tomli==2.0.1,Twisted==23.10.0,types-attrs==19.1.0,types-beautifulsoup4==4.12.0.7,types-docutils==0.20.0.3,types-html5lib==1.1.11.15,types-lxml==2023.10.21,types-Pillow==10.1.0.0,types-Pygments==2.16.0.0,types-pyOpenSSL==23.3.0.0,types-setuptools==68.2.0.0,typing_extensions==4.8.0,urllib3==2.1.0,w3lib==2.1.2,wheel==0.42.0,zope.interface==6.1\ntyping: commands[0]> mypy scrapy tests\nscrapy/core/downloader/handlers/__init__.py:58: error: Too few arguments for \"build_from_crawler\" [call-arg]\nscrapy/core/engine.py:361: error: Too few arguments for \"build_from_crawler\" [call-arg]\nscrapy/crawler.py:112: error: Too few arguments for \"build_from_crawler\" [call-arg]\nFound 3 errors in 3 files (checked 333 source files)\ntyping: exit 1 (8.99 seconds) /home/runner/work/scrapy/scrapy> mypy scrapy tests pid=1984\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True setuptools.build_meta __legacy__\n typing: FAIL code 1 (25.59=setup[16.60]+cmd[8.99] seconds)\n evaluation failed :( (25.71 seconds)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/scrapy/core/downloader/contextfactory.py b/scrapy/core/downloader/contextfactory.py\nindex 9047a5b83..05d6e7a09 100644\n--- a/scrapy/core/downloader/contextfactory.py\n+++ b/scrapy/core/downloader/contextfactory.py\n@@ -166,14 +166,14 @@ def load_context_factory_from_settings(settings, crawler):\n # try method-aware context factory\n try:\n context_factory = build_from_crawler(\n- objcls=context_factory_cls,\n- crawler=crawler,\n+ context_factory_cls,\n+ crawler,\n method=ssl_method,\n )\n except TypeError:\n context_factory = build_from_crawler(\n- objcls=context_factory_cls,\n- crawler=crawler,\n+ context_factory_cls,\n+ crawler,\n )\n msg = (\n f\"{settings['DOWNLOADER_CLIENTCONTEXTFACTORY']} does not accept \"\ndiff --git a/scrapy/core/downloader/handlers/__init__.py b/scrapy/core/downloader/handlers/__init__.py\nindex 466393b80..416669b7f 100644\n--- a/scrapy/core/downloader/handlers/__init__.py\n+++ b/scrapy/core/downloader/handlers/__init__.py\n@@ -56,8 +56,8 @@ class DownloadHandlers:\n if skip_lazy and getattr(dhcls, \"lazy\", True):\n return None\n dh = build_from_crawler(\n- objcls=dhcls,\n- crawler=self._crawler,\n+ dhcls,\n+ self._crawler,\n )\n except NotConfigured as ex:\n self._notconfigured[scheme] = str(ex)\ndiff --git a/scrapy/core/downloader/handlers/http10.py b/scrapy/core/downloader/handlers/http10.py\nindex 817a420c0..b6ac7a251 100644\n--- a/scrapy/core/downloader/handlers/http10.py\n+++ b/scrapy/core/downloader/handlers/http10.py\n@@ -31,8 +31,8 @@ class HTTP10DownloadHandler:\n host, port = to_unicode(factory.host), factory.port\n if factory.scheme == b\"https\":\n client_context_factory = build_from_crawler(\n- objcls=self.ClientContextFactory,\n- crawler=self._crawler,\n+ self.ClientContextFactory,\n+ self._crawler,\n )\n return reactor.connectSSL(host, port, factory, client_context_factory)\n return reactor.connectTCP(host, port, factory)\ndiff --git a/scrapy/core/downloader/handlers/s3.py b/scrapy/core/downloader/handlers/s3.py\nindex a37f29aac..1f7533759 100644\n--- a/scrapy/core/downloader/handlers/s3.py\n+++ b/scrapy/core/downloader/handlers/s3.py\n@@ -51,8 +51,8 @@ class S3DownloadHandler:\n )\n \n _http_handler = build_from_crawler(\n- objcls=httpdownloadhandler,\n- crawler=crawler,\n+ httpdownloadhandler,\n+ crawler,\n )\n self._download_http = _http_handler.download_request\n \ndiff --git a/scrapy/core/engine.py b/scrapy/core/engine.py\nindex 281dc8a54..545cd401f 100644\n--- a/scrapy/core/engine.py\n+++ b/scrapy/core/engine.py\n@@ -358,7 +358,7 @@ class ExecutionEngine:\n raise RuntimeError(f\"No free spider slot when opening {spider.name!r}\")\n logger.info(\"Spider opened\", extra={\"spider\": spider})\n nextcall = CallLaterOnce(self._next_request)\n- scheduler = build_from_crawler(self.scheduler_cls, crawler=self.crawler)\n+ scheduler = build_from_crawler(self.scheduler_cls, self.crawler)\n start_requests = yield self.scraper.spidermw.process_start_requests(\n start_requests, spider\n )\ndiff --git a/scrapy/core/scheduler.py b/scrapy/core/scheduler.py\nindex bf027e1f9..f41b83a67 100644\n--- a/scrapy/core/scheduler.py\n+++ b/scrapy/core/scheduler.py\n@@ -324,7 +324,7 @@ class Scheduler(BaseScheduler):\n \"\"\"Create a new priority queue instance, with in-memory storage\"\"\"\n return build_from_crawler(\n self.pqclass,\n- crawler=self.crawler,\n+ self.crawler,\n downstream_queue_cls=self.mqclass,\n key=\"\",\n )\n@@ -335,7 +335,7 @@ class Scheduler(BaseScheduler):\n state = self._read_dqs_state(self.dqdir)\n q = build_from_crawler(\n self.pqclass,\n- crawler=self.crawler,\n+ self.crawler,\n downstream_queue_cls=self.dqclass,\n key=self.dqdir,\n startprios=state,\ndiff --git a/scrapy/crawler.py b/scrapy/crawler.py\nindex b7a6df2f3..844d5f759 100644\n--- a/scrapy/crawler.py\n+++ b/scrapy/crawler.py\n@@ -111,7 +111,7 @@ class Crawler:\n \n self.request_fingerprinter = build_from_crawler(\n load_object(self.settings[\"REQUEST_FINGERPRINTER_CLASS\"]),\n- crawler=self,\n+ self,\n )\n \n reactor_class: str = self.settings[\"TWISTED_REACTOR\"]\ndiff --git a/tests/test_downloader_handlers.py b/tests/test_downloader_handlers.py\nindex 37a53643b..dd07d33f1 100644\n--- a/tests/test_downloader_handlers.py\n+++ b/tests/test_downloader_handlers.py\n@@ -829,8 +829,8 @@ class S3AnonTestCase(unittest.TestCase):\n skip_if_no_boto()\n crawler = get_crawler()\n self.s3reqh = build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n httpdownloadhandler=HttpDownloadHandlerMock,\n # anon=True, # implicit\n )\n@@ -859,8 +859,8 @@ class S3TestCase(unittest.TestCase):\n skip_if_no_boto()\n crawler = get_crawler()\n s3reqh = build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n aws_access_key_id=self.AWS_ACCESS_KEY_ID,\n aws_secret_access_key=self.AWS_SECRET_ACCESS_KEY,\n httpdownloadhandler=HttpDownloadHandlerMock,\n@@ -886,8 +886,8 @@ class S3TestCase(unittest.TestCase):\n try:\n crawler = get_crawler()\n build_from_crawler(\n- objcls=S3DownloadHandler,\n- crawler=crawler,\n+ S3DownloadHandler,\n+ crawler,\n extra_kw=True,\n )\n except Exception as e:\ndiff --git a/tests/test_webclient.py b/tests/test_webclient.py\nindex a69d9c1b0..d4b6ba15b 100644\n--- a/tests/test_webclient.py\n+++ b/tests/test_webclient.py\n@@ -471,7 +471,7 @@ class WebClientCustomCiphersSSLTestCase(WebClientSSLTestCase):\n s = \"0123456789\" * 10\n settings = Settings({\"DOWNLOADER_CLIENT_TLS_CIPHERS\": self.custom_ciphers})\n client_context_factory = build_from_settings(\n- ScrapyClientContextFactory, settings=settings\n+ ScrapyClientContextFactory, settings\n )\n return getPage(\n self.getURL(\"payload\"), body=s, contextFactory=client_context_factory\n@@ -483,7 +483,7 @@ class WebClientCustomCiphersSSLTestCase(WebClientSSLTestCase):\n {\"DOWNLOADER_CLIENT_TLS_CIPHERS\": \"ECDHE-RSA-AES256-GCM-SHA384\"}\n )\n client_context_factory = build_from_settings(\n- ScrapyClientContextFactory, settings=settings\n+ ScrapyClientContextFactory, settings\n )\n d = getPage(\n self.getURL(\"payload\"), body=s, contextFactory=client_context_factory\n", "difficulty": 2, "changed_files": ["scrapy/core/downloader/contextfactory.py", "scrapy/core/downloader/handlers/__init__.py", "scrapy/core/downloader/handlers/http10.py", "scrapy/core/downloader/handlers/s3.py", "scrapy/core/engine.py", "scrapy/core/scheduler.py", "scrapy/crawler.py", "tests/test_downloader_handlers.py", "tests/test_webclient.py"], "commit_link": "https://github.com/scrapy/scrapy/tree/1168b9244d680437e2a2683294a55c0b52c118f6"}
|
|
|
|
data/python/124065b.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 121, "repo_owner": "autogluon", "repo_name": "autogluon", "head_branch": "modify-sem-seg", "workflow_name": "Codespell", "workflow_filename": "codespell.yml", "workflow_path": ".github/workflows/codespell.yml", "contributor": "Harry-zzh", "sha_fail": "124065b09d5b45bfbd2a38262aa8e9fc11ded7a6", "sha_success": "0ed9a432f338d4a05f3b729c99c17f2546c96573", "workflow": "---\nname: Codespell\n\non:\n push:\n branches: [master]\n pull_request:\n branches: [master]\n\npermissions:\n contents: read\n\njobs:\n codespell:\n name: Check for spelling errors\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout\n uses: actions/checkout@v3\n - name: Codespell\n uses: codespell-project/actions-codespell@v2\n", "logs": [{"step_name": "Check for spelling errors/4_Codespell.txt", "log": "##[group]Run codespell-project/actions-codespell@v2\nwith:\n skip: ./.git\n##[endgroup]\n##[command]/usr/bin/docker run --name da69a10d75df8948c097276461d46605f8_7bca0f --label 8689da --workdir /github/workspace --rm -e \"INPUT_CHECK_FILENAMES\" -e \"INPUT_CHECK_HIDDEN\" -e \"INPUT_EXCLUDE_FILE\" -e \"INPUT_SKIP\" -e \"INPUT_BUILTIN\" -e \"INPUT_IGNORE_WORDS_FILE\" -e \"INPUT_IGNORE_WORDS_LIST\" -e \"INPUT_URI_IGNORE_WORDS_LIST\" -e \"INPUT_PATH\" -e \"INPUT_ONLY_WARN\" -e \"HOME\" -e \"GITHUB_JOB\" -e \"GITHUB_REF\" -e \"GITHUB_SHA\" -e \"GITHUB_REPOSITORY\" -e \"GITHUB_REPOSITORY_OWNER\" -e \"GITHUB_REPOSITORY_OWNER_ID\" -e \"GITHUB_RUN_ID\" -e \"GITHUB_RUN_NUMBER\" -e \"GITHUB_RETENTION_DAYS\" -e \"GITHUB_RUN_ATTEMPT\" -e \"GITHUB_REPOSITORY_ID\" -e \"GITHUB_ACTOR_ID\" -e \"GITHUB_ACTOR\" -e \"GITHUB_TRIGGERING_ACTOR\" -e \"GITHUB_WORKFLOW\" -e \"GITHUB_HEAD_REF\" -e \"GITHUB_BASE_REF\" -e \"GITHUB_EVENT_NAME\" -e \"GITHUB_SERVER_URL\" -e \"GITHUB_API_URL\" -e \"GITHUB_GRAPHQL_URL\" -e \"GITHUB_REF_NAME\" -e \"GITHUB_REF_PROTECTED\" -e \"GITHUB_REF_TYPE\" -e \"GITHUB_WORKFLOW_REF\" -e \"GITHUB_WORKFLOW_SHA\" -e \"GITHUB_WORKSPACE\" -e \"GITHUB_ACTION\" -e \"GITHUB_EVENT_PATH\" -e \"GITHUB_ACTION_REPOSITORY\" -e \"GITHUB_ACTION_REF\" -e \"GITHUB_PATH\" -e \"GITHUB_ENV\" -e \"GITHUB_STEP_SUMMARY\" -e \"GITHUB_STATE\" -e \"GITHUB_OUTPUT\" -e \"RUNNER_OS\" -e \"RUNNER_ARCH\" -e \"RUNNER_NAME\" -e \"RUNNER_ENVIRONMENT\" -e \"RUNNER_TOOL_CACHE\" -e \"RUNNER_TEMP\" -e \"RUNNER_WORKSPACE\" -e \"ACTIONS_RUNTIME_URL\" -e \"ACTIONS_RUNTIME_TOKEN\" -e \"ACTIONS_CACHE_URL\" -e \"ACTIONS_RESULTS_URL\" -e GITHUB_ACTIONS=true -e CI=true -v \"/var/run/docker.sock\":\"/var/run/docker.sock\" -v \"/home/runner/work/_temp/_github_home\":\"/github/home\" -v \"/home/runner/work/_temp/_github_workflow\":\"/github/workflow\" -v \"/home/runner/work/_temp/_runner_file_commands\":\"/github/file_commands\" -v \"/home/runner/work/autogluon/autogluon\":\"/github/workspace\" 8689da:69a10d75df8948c097276461d46605f8\nRunning codespell on '' with the following options...\nCheck filenames? ''\nCheck hidden? ''\nExclude file ''\nSkipping './.git'\nBuiltin dictionaries ''\nIgnore words file ''\nIgnore words list ''\nIgnore URI words list ''\nResulting CLI options --skip ./.git\n1\n##[error]./multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py:752: evalution ==> evaluation, evolution\nCodespell found one or more problems\n"}], "diff": "diff --git a/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py b/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\nindex 27032888..e6c62802 100644\n--- a/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\n+++ b/multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py\n@@ -749,7 +749,7 @@ class MAE(COD):\n COD_METRICS_NAMES = {\"sm\": SM(), \"fm\": FM(), \"em\": EM(), \"mae\": MAE()}\n \n \n-# TODO: Modify multi-gpu evalution error. Maybe there will be a more elegant way.\n+# TODO: Modify multi-gpu evaluation error. Maybe there will be a more elegant way.\n class Multiclass_IoU_Pred:\n \"\"\"\n Compute the IoU for multi-class semantic segmentation based on https://github.com/xieenze/Trans2Seg/blob/master/segmentron/utils/score.py.\n", "difficulty": 0, "changed_files": ["multimodal/src/autogluon/multimodal/optimization/semantic_seg_metrics.py"], "commit_link": "https://github.com/autogluon/autogluon/tree/124065b09d5b45bfbd2a38262aa8e9fc11ded7a6"}
|
|
|
|
data/python/16a0c04.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 140, "repo_owner": "dask", "repo_name": "dask", "head_branch": "milesgranger/10801-more-deprecations", "workflow_name": "Linting", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "milesgranger", "sha_fail": "16a0c04d06205527ec5e379df2596b399ee5dadc", "sha_success": "3716f88d477a969d0910c75b4268949af11c0e6e", "workflow": "name: Linting\n\non:\n push:\n branches: main\n pull_request:\n branches: main\n\njobs:\n checks:\n name: pre-commit hooks\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - uses: actions/setup-python@v5\n with:\n python-version: '3.9'\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "pre-commit hooks/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from nodeenv>=0.11.1->pre-commit) (58.1.0)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 19.7 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 27.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (738 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 738.9/738.9 kB 80.2 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 114.1 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 78.9 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nInstalling collected packages: distlib, pyyaml, platformdirs, nodeenv, identify, filelock, cfgv, virtualenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 virtualenv-20.25.0\n\n[notice] A new release of pip is available: 23.0.1 -> 23.3.2\n[notice] To update, run: pip install --upgrade pip\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.9.18/x64|2f6023be5c7e4f75d9ed9a00c3fcb4c3fa7d8493a70885078264f9accc27f632\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nReceived 0 of 69094681 (0.0%), 0.0 MBs/sec\nCache Size: ~66 MB (69094681 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/2e80fd5c-4883-4bdb-be22-ed0a545b7db7/cache.tzst -P -C /home/runner/work/dask/dask --use-compress-program unzstd\nReceived 69094681 of 69094681 (100.0%), 32.9 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.9.18/x64|2f6023be5c7e4f75d9ed9a00c3fcb4c3fa7d8493a70885078264f9accc27f632\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nfix end of files.........................................................\u001b[42mPassed\u001b[m\ndebug statements (python)................................................\u001b[42mPassed\u001b[m\nabsolufy-imports.........................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\npyupgrade................................................................\u001b[42mPassed\u001b[m\nblack....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted dask/utils.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m257 files \u001b[0mleft unchanged.\n\nflake8...................................................................\u001b[42mPassed\u001b[m\ncodespell................................................................\u001b[42mPassed\u001b[m\nmypy.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: mypy\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\ndask/utils.py:231: error: Unsupported operand types for + (\"str\" and \"None\") [operator]\ndask/utils.py:231: note: Right operand is of type \"Optional[str]\"\ndask/utils.py:257: error: Unsupported operand types for + (\"str\" and \"None\") [operator]\ndask/utils.py:257: note: Right operand is of type \"Optional[str]\"\nFound 2 errors in 1 file (checked 258 source files)\n\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/dask/utils.py b/dask/utils.py\u001b[m\n\u001b[1mindex 80ea6e8..26f1176 100644\u001b[m\n\u001b[1m--- a/dask/utils.py\u001b[m\n\u001b[1m+++ b/dask/utils.py\u001b[m\n\u001b[36m@@ -148,7 +148,7 @@\u001b[m \u001b[mdef _deprecated_kwarg(\u001b[m\n new_arg_name: str | None = None,\u001b[m\n mapping: Mapping[Any, Any] | Callable[[Any], Any] | None = None,\u001b[m\n stacklevel: int = 2,\u001b[m\n\u001b[31m- comment: str | None = None\u001b[m\n\u001b[32m+\u001b[m\u001b[32m comment: str | None = None,\u001b[m\n ) -> Callable[[F], F]:\u001b[m\n \"\"\"\u001b[m\n Decorator to deprecate a keyword argument of a function.\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/dask/utils.py b/dask/utils.py\nindex 80ea6e8d6..0da2953c8 100644\n--- a/dask/utils.py\n+++ b/dask/utils.py\n@@ -148,7 +148,7 @@ def _deprecated_kwarg(\n new_arg_name: str | None = None,\n mapping: Mapping[Any, Any] | Callable[[Any], Any] | None = None,\n stacklevel: int = 2,\n- comment: str | None = None\n+ comment: str | None = None,\n ) -> Callable[[F], F]:\n \"\"\"\n Decorator to deprecate a keyword argument of a function.\n@@ -219,7 +219,7 @@ def _deprecated_kwarg(\n \"mapping from old to new argument values must be dict or callable!\"\n )\n \n- comment = f\"\\n{comment}\" or \"\"\n+ comment_ = f\"\\n{comment}\" or \"\"\n \n def _deprecated_kwarg(func: F) -> F:\n @wraps(func)\n@@ -232,7 +232,7 @@ def _deprecated_kwarg(\n f\"the {repr(old_arg_name)} keyword is deprecated and \"\n \"will be removed in a future version. Please take \"\n f\"steps to stop the use of {repr(old_arg_name)}\"\n- ) + comment\n+ ) + comment_\n warnings.warn(msg, FutureWarning, stacklevel=stacklevel)\n kwargs[old_arg_name] = old_arg_value\n return func(*args, **kwargs)\n@@ -254,7 +254,7 @@ def _deprecated_kwarg(\n f\"use {repr(new_arg_name)} instead.\"\n )\n \n- warnings.warn(msg + comment, FutureWarning, stacklevel=stacklevel)\n+ warnings.warn(msg + comment_, FutureWarning, stacklevel=stacklevel)\n if kwargs.get(new_arg_name) is not None:\n msg = (\n f\"Can only specify {repr(old_arg_name)} \"\n", "difficulty": 0, "changed_files": ["dask/utils.py"], "commit_link": "https://github.com/dask/dask/tree/16a0c04d06205527ec5e379df2596b399ee5dadc"}
|
|
|
|
data/python/1afe2c9.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 107, "repo_owner": "encode", "repo_name": "httpx", "head_branch": "add-ssl-context-argument", "workflow_name": "Test Suite", "workflow_filename": "test-suite.yml", "workflow_path": ".github/workflows/test-suite.yml", "contributor": "karpetrosyan", "sha_fail": "1afe2c9cb192d3760d59190cc7892e7ac37d5e27", "sha_success": "6be802e03bdb78d20f9c8df63a43f9167dcbfd49", "workflow": "---\nname: Test Suite\n\non:\n push:\n branches: [\"master\"]\n pull_request:\n branches: [\"master\"]\n\njobs:\n tests:\n name: \"Python ${{ matrix.python-version }}\"\n runs-on: \"ubuntu-latest\"\n\n strategy:\n matrix:\n python-version: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n\n steps:\n - uses: \"actions/checkout@v4\"\n - uses: \"actions/setup-python@v4\"\n with:\n python-version: \"${{ matrix.python-version }}\"\n allow-prereleases: true\n - name: \"Install dependencies\"\n run: \"scripts/install\"\n - name: \"Run linting checks\"\n run: \"scripts/check\"\n - name: \"Build package & docs\"\n run: \"scripts/build\"\n - name: \"Run tests\"\n run: \"scripts/test\"\n - name: \"Enforce coverage\"\n run: \"scripts/coverage\"\n", "logs": [{"step_name": "Python 3.10/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Python 3.11/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Python 3.12/5_Run linting checks.txt", "log": "##[group]Run scripts/check\n\u001b[36;1mscripts/check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n+ ./scripts/sync-version\nCHANGELOG_VERSION: 0.26.0\nVERSION: 0.26.0\n+ ruff format httpx tests --diff\n60 files left unchanged\n+ mypy httpx tests\nhttpx/_client.py:1456: error: Unexpected keyword argument \"verify\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_client.py:1458: error: Name \"verify\" is not defined [name-defined]\nhttpx/_client.py:1459: error: Name \"cert\" is not defined [name-defined]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nhttpx/_client.py:1456: error: Unexpected keyword argument \"cert\" for \"AsyncHTTPTransport\" [call-arg]\nhttpx/_transports/default.py:260: note: \"AsyncHTTPTransport\" defined here\nFound 4 errors in 1 file (checked 60 source files)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/httpx/_client.py b/httpx/_client.py\nindex e5cff5d..4dd2d07 100644\n--- a/httpx/_client.py\n+++ b/httpx/_client.py\n@@ -1455,8 +1455,6 @@ class AsyncClient(BaseClient):\n ) -> AsyncBaseTransport:\n return AsyncHTTPTransport(\n ssl_context=ssl_context,\n- verify=verify,\n- cert=cert,\n http1=http1,\n http2=http2,\n limits=limits,\n", "difficulty": 2, "changed_files": ["httpx/_client.py"], "commit_link": "https://github.com/encode/httpx/tree/1afe2c9cb192d3760d59190cc7892e7ac37d5e27"}
|
|
|
|
data/python/1b22eb7.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 88, "repo_owner": "hugsy", "repo_name": "gef", "head_branch": "main", "workflow_name": "CI Test for GEF", "workflow_filename": "run-tests.yml", "workflow_path": ".github/workflows/run-tests.yml", "contributor": "Angelo942", "sha_fail": "1b22eb71da460e7d3a301842f65c256b920f6195", "sha_success": "a5bb1bce7421428c7c5bd36344708c7a418baa87", "workflow": "name: CI Test for GEF\n\non:\n push:\n branches:\n - main\n\n pull_request:\n branches:\n - main\n\njobs:\n build:\n strategy:\n fail-fast: false\n matrix:\n os:\n - ubuntu-22.04\n - ubuntu-20.04\n name: \"Run Unit tests on ${{ matrix.os }}\"\n runs-on: ${{ matrix.os }}\n defaults:\n run:\n shell: bash\n\n steps:\n - uses: actions/checkout@v3\n\n - name: Install python and toolchain\n run: |\n sudo apt-get update\n sudo apt-get install -y gdb-multiarch python3-dev python3-pip python3-wheel python3-setuptools git cmake gcc g++ pkg-config libglib2.0-dev gdbserver qemu-user\n sudo python3 -m pip install --upgrade pip\n\n - name: Set runtime environment variables\n run: |\n echo PY_VER=`gdb -q -nx -ex \"pi print('.'.join(map(str, sys.version_info[:2])))\" -ex quit` >> $GITHUB_ENV\n echo GEF_CI_NB_CPU=`grep -c ^processor /proc/cpuinfo` >> $GITHUB_ENV\n echo GEF_CI_ARCH=`uname --processor` >> $GITHUB_ENV\n echo GEF_CI_CACHE_DIR=`python3 -m pip cache dir` >> $GITHUB_ENV\n\n - name: Cache dependencies\n uses: actions/cache@v3\n id: cache-deps\n env:\n cache-name: cache-deps\n with:\n key: ${{ runner.os }}-pip-${{ hashFiles('**/requirements.txt') }}\n path: ${{ env.GEF_CI_CACHE_DIR }}\n restore-keys:\n ${{ runner.os }}-pip-${{ env.cache-name }}-\n ${{ runner.os }}-pip-\n ${{ runner.os }}-${{ env.cache-name }}-\n ${{ runner.os }}-\n\n - name: Install Python Requirements\n run: |\n mkdir -p ${{ env.GEF_CI_CACHE_DIR }}\n python${{ env.PY_VER }} -m pip install --user --upgrade -r tests/requirements.txt\n\n - name: Setup GEF\n run: |\n echo \"source $(pwd)/gef.py\" > ~/.gdbinit\n gdb -q -ex 'gef missing' -ex 'gef help' -ex 'gef config' -ex start -ex continue -ex quit /bin/pwd\n\n - name: Run Tests\n run: |\n make -C tests/binaries -j ${{ env.GEF_CI_NB_CPU }}\n python${{ env.PY_VER }} -m pytest --forked -n ${{ env.GEF_CI_NB_CPU }} -v -k \"not benchmark\" tests/\n", "logs": [{"step_name": "Run Unit tests on ubuntu-22.04/8_Run Tests.txt", "log": "##[group]Run make -C tests/binaries -j 4\n\u001b[36;1mmake -C tests/binaries -j 4\u001b[0m\n\u001b[36;1mpython3.10 -m pytest --forked -n 4 -v -k \"not benchmark\" tests/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PY_VER: 3.10\n GEF_CI_NB_CPU: 4\n GEF_CI_ARCH: x86_64\n GEF_CI_CACHE_DIR: /home/runner/.cache/pip\n##[endgroup]\nmake: Entering directory '/home/runner/work/gef/gef/tests/binaries'\n[+] Building '/tmp/bss.out'\n[+] Building '/tmp/canary.out'\n[+] Building '/tmp/checksec-no-canary.out'\n[+] Building '/tmp/checksec-no-nx.out'\n[+] Building '/tmp/checksec-no-pie.out'\n[+] Building '/tmp/default.out'\n[+] Building '/tmp/format-string-helper.out'\n[+] Building '/tmp/heap-analysis.out'\n[+] Building '/tmp/heap-bins.out'\n[+] Building '/tmp/heap-fastbins.out'\n[+] Building '/tmp/heap-multiple-heaps.out'\n[+] Building '/tmp/heap-non-main.out'\n[+] Building '/tmp/heap-tcache.out'\n[+] Building '/tmp/heap.out'\n[+] Building '/tmp/memwatch.out'\n[+] Building '/tmp/mmap-known-address.out'\n[+] Building '/tmp/nested.out'\n[+] Building '/tmp/nested2.out'\n[+] Building '/tmp/pattern.out'\n[+] Building '/tmp/pcustom.out'\n[+] Building '/tmp/class.out'\nmake: Leaving directory '/home/runner/work/gef/gef/tests/binaries'\n/home/runner/.local/lib/python3.10/site-packages/pytest_benchmark/logger.py:46: PytestBenchmarkWarning: Benchmarks are automatically disabled because xdist plugin is active.Benchmarks cannot be performed reliably in a parallelized environment.\n warner(PytestBenchmarkWarning(text))\n============================= test session starts ==============================\nplatform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.3.0 -- /usr/bin/python3.10\ncachedir: .pytest_cache\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/gef/gef/tests\nconfigfile: pytest.ini\nplugins: benchmark-4.0.0, forked-1.6.0, xdist-3.5.0, cov-4.1.0\ncreated: 4/4 workers\n4 workers [155 items]\n\nscheduling tests via LoadScheduling\n\ntests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \n[gw1] [ 0%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/gef_session.py::GefSessionApi::test_func_get_filepath \n[gw3] [ 1%] PASSED tests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \n[gw1] [ 1%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_filepath \ntests/api/gef_session.py::GefSessionApi::test_func_get_pid \n[gw2] [ 2%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \n[gw1] [ 3%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_pid \ntests/api/gef_session.py::GefSessionApi::test_root_dir_local \n[gw3] [ 3%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \n[gw1] [ 4%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_local \ntests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \n[gw2] [ 5%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \ntests/api/misc.py::MiscFunctionTest::test_func_update_gef \n[gw2] [ 5%] SKIPPED tests/api/misc.py::MiscFunctionTest::test_func_update_gef \ntests/api/misc.py::MiscFunctionTest::test_func_which \n[gw1] [ 6%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \ntests/api/gef_session.py::GefSessionApi::test_root_dir_remote \n[gw3] [ 7%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \n[gw2] [ 7%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_which \ntests/commands/aliases.py::AliasesCommand::test_cmd_aliases \n[gw3] [ 8%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \n[gw1] [ 9%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_remote \ntests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \n[gw1] [ 9%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \ntests/api/misc.py::MiscFunctionTest::test_func_parse_address \n[gw2] [ 10%] PASSED tests/commands/aliases.py::AliasesCommand::test_cmd_aliases \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \n[gw3] [ 10%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \n[gw0] [ 11%] PASSED tests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \n[gw2] [ 12%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \n[gw1] [ 12%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_address \ntests/api/misc.py::MiscFunctionTest::test_func_parse_maps \n[gw0] [ 13%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \ntests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \n[gw3] [ 14%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \n[gw0] [ 14%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \n[gw2] [ 15%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \ntests/commands/canary.py::CanaryCommand::test_cmd_canary \n[gw0] [ 16%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \n[gw3] [ 16%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \ntests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \n[gw0] [ 17%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \ntests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \n[gw3] [ 18%] PASSED tests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \ntests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \n[gw2] [ 18%] PASSED tests/commands/canary.py::CanaryCommand::test_cmd_canary \ntests/commands/canary.py::CanaryCommand::test_overwrite_canary \n[gw3] [ 19%] PASSED tests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \n[gw2] [ 20%] PASSED tests/commands/canary.py::CanaryCommand::test_overwrite_canary \ntests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \n[gw0] [ 20%] PASSED tests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \n[gw2] [ 21%] PASSED tests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \ntests/commands/functions.py::FunctionsCommand::test_cmd_functions \n[gw2] [ 21%] PASSED tests/commands/functions.py::FunctionsCommand::test_cmd_functions \ntests/commands/gef.py::GefCommand::test_cmd_gef \n[gw1] [ 22%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_maps \ntests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \n[gw2] [ 23%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef \ntests/commands/gef.py::GefCommand::test_cmd_gef_config \n[gw3] [ 23%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \n[gw2] [ 24%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_get \n[gw1] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \ntests/commands/gef.py::GefCommand::test_cmd_gef_save \n[gw2] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_get \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_set \n[gw3] [ 26%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_set \ntests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw1] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_save \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \n[gw2] [ 28%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_help \ntests/commands/gef.py::GefCommand::test_cmd_gef_install \n[gw3] [ 29%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \n[gw2] [ 29%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_install \ntests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \n[gw0] [ 30%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \n[gw3] [ 30%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \n[gw1] [ 31%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \n[gw2] [ 32%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \ntests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \n[gw1] [ 32%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \n[gw2] [ 33%] PASSED tests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \ntests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \n[gw3] [ 34%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \n[gw1] [ 34%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \ntests/commands/got.py::GotCommand::test_cmd_got \n[gw0] [ 35%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \n[gw1] [ 36%] PASSED tests/commands/got.py::GotCommand::test_cmd_got \ntests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \n[gw3] [ 36%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \n[gw2] [ 37%] PASSED tests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \ntests/commands/highlight.py::HighlightCommand::test_cmd_highlight \n[gw1] [ 38%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \n[gw2] [ 38%] PASSED tests/commands/highlight.py::HighlightCommand::test_cmd_highlight \ntests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \n[gw0] [ 39%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \ntests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \n[gw2] [ 40%] PASSED tests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \ntests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \n[gw1] [ 40%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \n[gw3] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \n[gw2] [ 41%] PASSED tests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_list \n[gw1] [ 42%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \n[gw0] [ 43%] PASSED tests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \ntests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \n[gw3] [ 43%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \n[gw1] [ 44%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \n[gw2] [ 45%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_list \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \n[gw1] [ 45%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \n[gw0] [ 46%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes \n[gw3] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \n[gw2] [ 47%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \n[gw1] [ 48%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \ntests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \n[gw0] [ 49%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw1] [ 49%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \ntests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \n[gw0] [ 50%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw2] [ 50%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \n[gw3] [ 51%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \n[gw1] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \n[gw0] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \n[gw1] [ 53%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \n[gw0] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \n[gw1] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \ntests/commands/nop.py::NopCommand::test_cmd_nop_inactive \n[gw0] [ 55%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \n[gw3] [ 56%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \n[gw1] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_inactive \ntests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \n[gw0] [ 57%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \n[gw3] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch \n[gw1] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \n[gw3] [ 59%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte \n[gw1] [ 60%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \n[gw3] [ 60%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \n[gw0] [ 61%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \ntests/commands/patch.py::PatchCommand::test_cmd_patch_string \n[gw3] [ 61%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \ntests/commands/patch.py::PatchCommand::test_cmd_patch_dword \n[gw0] [ 62%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_string \ntests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw1] [ 63%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \n[gw3] [ 63%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_dword \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword \n[gw0] [ 64%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_word \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \n[gw2] [ 65%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \n[gw3] [ 65%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \n[gw0] [ 66%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \n[gw2] [ 67%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop \n[gw3] [ 67%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \n[gw1] [ 68%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \ntests/commands/pie.py::PieCommand::test_cmd_pie \n[gw2] [ 69%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \n[gw3] [ 69%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \n[gw2] [ 70%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \n[gw1] [ 70%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \n[gw1] [ 71%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \ntests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \n[gw2] [ 72%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \n[gw1] [ 72%] PASSED tests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \ntests/commands/scan.py::ScanCommand::test_cmd_scan \n[gw2] [ 73%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \n[gw0] [ 74%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \ntests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \n[gw2] [ 74%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \n[gw3] [ 75%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \n[gw2] [ 76%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \n[gw1] [ 76%] PASSED tests/commands/scan.py::ScanCommand::test_cmd_scan \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \n[gw0] [ 77%] PASSED tests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \ntests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \n[gw3] [ 78%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 78%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \n[gw1] [ 79%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \ntests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \n[gw0] [ 80%] PASSED tests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \ntests/commands/registers.py::RegistersCommand::test_cmd_registers \n[gw1] [ 80%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \n[gw3] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \n[gw2] [ 82%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \n[gw1] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \ntests/commands/stub.py::StubCommand::test_cmd_stub \n[gw0] [ 83%] PASSED tests/commands/registers.py::RegistersCommand::test_cmd_registers \ntests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \n[gw3] [ 84%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \ntests/commands/theme.py::ThemeCommand::test_cmd_theme \n[gw2] [ 85%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \ntests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \n[gw1] [ 85%] PASSED tests/commands/stub.py::StubCommand::test_cmd_stub \ntests/commands/version.py::VersionCommand::test_cmd_version \n[gw2] [ 86%] PASSED tests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \n[gw1] [ 87%] PASSED tests/commands/version.py::VersionCommand::test_cmd_version \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \n[gw0] [ 87%] PASSED tests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \ntests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \n[gw1] [ 88%] FAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n[gw2] [ 89%] PASSED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \n[gw1] [ 89%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \ntests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \n[gw0] [ 90%] PASSED tests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \ntests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \n[gw2] [ 90%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \ntests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \n[gw0] [ 91%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \n[gw2] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \n[gw1] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \ntests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \n[gw2] [ 93%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \n[gw0] [ 94%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \n[gw1] [ 94%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw0] [ 95%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw2] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \ntests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw1] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw0] [ 97%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw1] [ 98%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw2] [ 98%] PASSED tests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw3] [ 99%] PASSED tests/commands/theme.py::ThemeCommand::test_cmd_theme \ntests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n[gw3] [100%] PASSED tests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n\n=================================== FAILURES ===================================\n_____________________ XinfoCommand.test_cmd_xinfo_on_class _____________________\n[gw1] linux -- Python 3.10.12 /usr/bin/python3.10\nself = <tests.commands.xinfo.XinfoCommand testMethod=test_cmd_xinfo_on_class>\n\n def test_cmd_xinfo_on_class(self):\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n self.assertNoException(res)\n> self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\nE AssertionError: 'Symbol: B<TraitA, TraitB>::Run()+4' not found in 'Reading symbols from /tmp/class.out...\\nError while writing index for `/tmp/class.out\\': mkstemp: No such file or directory.\\nGEF for linux ready, type `gef\\' to start, `gef config\\' to configure\\n89 commands loaded and 5 functions added for GDB 12.1 in 0.00ms using Python engine 3.10\\nBreakpoint 1 at 0x1328: file class.cpp, line 21.\\nStarting program: /tmp/class.out \\n[Thread debugging using libthread_db enabled]\\nUsing host libthread_db library \"/lib/x86_64-linux-gnu/libthread_db.so.1\".\\n\\nBreakpoint 1, B<TraitA, TraitB>::Run (this=0x55555556b2b0) at class.cpp:21\\n21\\t virtual void Run() { printf(\"I am B\\\\n\"); }\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500 xinfo: 0x55555555532c \u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\nPage: 0x0000555555555000 \u2192 0x0000555555556000 (size=0x1000)\\nPermissions: r-x\\nPathname: /tmp/class.out\\nOffset (from page): 0x32c\\nInode: 37098\\nSegment: .text (0x00005555555550a0-0x000055555555533a)\\nOffset (from segment): 0x28c\\nSymbol: B<TraitA, TraitB>::Run()+20'\n\ntests/commands/xinfo.py:27: AssertionError\n=========================== short test summary info ============================\nFAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class\n================== 1 failed, 153 passed, 1 skipped in 52.95s ===================\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Run Unit tests on ubuntu-20.04/8_Run Tests.txt", "log": "##[group]Run make -C tests/binaries -j 4\n\u001b[36;1mmake -C tests/binaries -j 4\u001b[0m\n\u001b[36;1mpython3.8 -m pytest --forked -n 4 -v -k \"not benchmark\" tests/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n PY_VER: 3.8\n GEF_CI_NB_CPU: 4\n GEF_CI_ARCH: x86_64\n GEF_CI_CACHE_DIR: /home/runner/.cache/pip\n##[endgroup]\nmake: Entering directory '/home/runner/work/gef/gef/tests/binaries'\n[+] Building '/tmp/heap.out'\n[+] Building '/tmp/bss.out'\n[+] Building '/tmp/canary.out'\n[+] Building '/tmp/heap-multiple-heaps.out'\n[+] Building '/tmp/checksec-no-canary.out'\n[+] Building '/tmp/format-string-helper.out'\n[+] Building '/tmp/checksec-no-nx.out'\n[+] Building '/tmp/heap-non-main.out'\n[+] Building '/tmp/nested2.out'\n[+] Building '/tmp/mmap-known-address.out'\n[+] Building '/tmp/heap-analysis.out'\n[+] Building '/tmp/default.out'\n[+] Building '/tmp/heap-tcache.out'\n[+] Building '/tmp/nested.out'\n[+] Building '/tmp/checksec-no-pie.out'\n[+] Building '/tmp/heap-bins.out'\n[+] Building '/tmp/pattern.out'\n[+] Building '/tmp/memwatch.out'\n[+] Building '/tmp/pcustom.out'\n[+] Building '/tmp/heap-fastbins.out'\n[+] Building '/tmp/class.out'\nmake: Leaving directory '/home/runner/work/gef/gef/tests/binaries'\n============================= test session starts ==============================\nplatform linux -- Python 3.8.10, pytest-7.4.4, pluggy-1.3.0 -- /usr/bin/python3.8\ncachedir: .pytest_cache\nbenchmark: 4.0.0 (defaults: timer=time.perf_counter disable_gc=False min_rounds=5 min_time=0.000005 max_time=1.0 calibration_precision=10 warmup=False warmup_iterations=100000)\nrootdir: /home/runner/work/gef/gef/tests\nconfigfile: pytest.ini\nplugins: forked-1.6.0, xdist-3.5.0, benchmark-4.0.0, cov-4.1.0\ncreated: 4/4 workers\n4 workers [155 items]\n\nscheduling tests via LoadScheduling\n\ntests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \n[gw1] [ 0%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_auxiliary_vector \ntests/api/gef_session.py::GefSessionApi::test_func_get_filepath \n[gw3] [ 1%] PASSED tests/commands/checksec.py::ChecksecCommand::test_cmd_checksec \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \n[gw1] [ 1%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_filepath \ntests/api/gef_session.py::GefSessionApi::test_func_get_pid \n[gw2] [ 2%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_permissions \ntests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \n[gw1] [ 3%] PASSED tests/api/gef_session.py::GefSessionApi::test_func_get_pid \ntests/api/gef_session.py::GefSessionApi::test_root_dir_local \n[gw3] [ 3%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \n[gw1] [ 4%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_local \ntests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \n[gw2] [ 5%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_show_last_exception \ntests/api/misc.py::MiscFunctionTest::test_func_update_gef \n[gw2] [ 5%] SKIPPED tests/api/misc.py::MiscFunctionTest::test_func_update_gef \ntests/api/misc.py::MiscFunctionTest::test_func_which \n[gw1] [ 6%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_qemu \ntests/api/gef_session.py::GefSessionApi::test_root_dir_remote \n[gw3] [ 7%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_backwards \ntests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \n[gw2] [ 7%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_which \ntests/commands/aliases.py::AliasesCommand::test_cmd_aliases \n[gw3] [ 8%] PASSED tests/commands/dereference.py::DereferenceCommand::test_cmd_dereference_forwards \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \n[gw1] [ 9%] PASSED tests/api/gef_session.py::GefSessionApi::test_root_dir_remote \ntests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \n[gw1] [ 9%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_gef_convenience \ntests/api/misc.py::MiscFunctionTest::test_func_parse_address \n[gw3] [ 10%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_disable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \n[gw0] [ 10%] PASSED tests/api/deprecated.py::GefFuncDeprecatedApi::test_deprecated_elf_values \ntests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \n[gw2] [ 11%] PASSED tests/commands/aliases.py::AliasesCommand::test_cmd_aliases \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \n[gw0] [ 12%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_gef_arch_ptrsize \ntests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \n[gw1] [ 12%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_address \ntests/api/misc.py::MiscFunctionTest::test_func_parse_maps \n[gw2] [ 13%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_show \ntests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \n[gw3] [ 14%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_enable \ntests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \n[gw0] [ 14%] PASSED tests/api/gef_arch.py::GefArchApi::test_func_reset_architecture \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \n[gw0] [ 15%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble \ntests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \n[gw2] [ 16%] PASSED tests/commands/aslr.py::AslrCommand::test_cmd_aslr_toggle \ntests/commands/canary.py::CanaryCommand::test_cmd_canary \n[gw3] [ 16%] PASSED tests/commands/edit_flags.py::EditFlagsCommand::test_cmd_edit_flags_toggle \ntests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \n[gw0] [ 17%] PASSED tests/api/gef_disasemble.py::GefDisassembleApiFunction::test_func_gef_disassemble_page_border \ntests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \n[gw3] [ 18%] PASSED tests/commands/elf_info.py::ElfInfoCommand::test_cmd_elf_info \ntests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \n[gw2] [ 18%] PASSED tests/commands/canary.py::CanaryCommand::test_cmd_canary \ntests/commands/canary.py::CanaryCommand::test_overwrite_canary \n[gw3] [ 19%] PASSED tests/commands/entry_break.py::EntryBreakCommand::test_cmd_entry_break \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \n[gw0] [ 20%] PASSED tests/api/gef_heap.py::GefHeapApi::test_class_glibcarena_main_arena \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \n[gw2] [ 20%] PASSED tests/commands/canary.py::CanaryCommand::test_overwrite_canary \ntests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \n[gw2] [ 21%] PASSED tests/commands/format_string_helper.py::FormatStringHelperCommand::test_cmd_format_string_helper \ntests/commands/functions.py::FunctionsCommand::test_cmd_functions \n[gw2] [ 21%] PASSED tests/commands/functions.py::FunctionsCommand::test_cmd_functions \ntests/commands/gef.py::GefCommand::test_cmd_gef \n[gw1] [ 22%] PASSED tests/api/misc.py::MiscFunctionTest::test_func_parse_maps \ntests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \n[gw2] [ 23%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef \ntests/commands/gef.py::GefCommand::test_cmd_gef_config \n[gw3] [ 23%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache_all \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \n[gw2] [ 24%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_get \n[gw1] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_run_and_run \ntests/commands/gef.py::GefCommand::test_cmd_gef_save \n[gw2] [ 25%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_get \ntests/commands/gef.py::GefCommand::test_cmd_gef_config_set \n[gw3] [ 26%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_unsorted \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_config_set \ntests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw2] [ 27%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_help \n[gw1] [ 28%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_save \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef.py::GefCommand::test_cmd_gef_install \n[gw3] [ 29%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_no_arg \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \n[gw0] [ 29%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_csize2tidx \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \n[gw3] [ 30%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunk_with_number \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \n[gw2] [ 30%] PASSED tests/commands/gef.py::GefCommand::test_cmd_gef_install \ntests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \n[gw1] [ 31%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \n[gw2] [ 32%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_set_arena \ntests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \n[gw1] [ 32%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_gef_remote_qemu_user \ntests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \n[gw0] [ 33%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_malloc_align_address \ntests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \n[gw2] [ 34%] PASSED tests/commands/heap_analysis.py::HeapAnalysisCommand::test_cmd_heap_analysis \ntests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \n[gw3] [ 34%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \n[gw1] [ 35%] PASSED tests/commands/gef_remote.py::GefRemoteCommand::test_cmd_target_remote \ntests/commands/got.py::GotCommand::test_cmd_got \n[gw3] [ 36%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_max_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \n[gw1] [ 36%] PASSED tests/commands/got.py::GotCommand::test_cmd_got \ntests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \n[gw0] [ 37%] PASSED tests/api/gef_heap.py::GefHeapApi::test_func_gef_heap_tidx2size \ntests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \n[gw2] [ 38%] PASSED tests/commands/hexdump.py::HexdumpCommand::test_cmd_hexdump \ntests/commands/highlight.py::HighlightCommand::test_cmd_highlight \n[gw1] [ 38%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_arenas \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \n[gw2] [ 39%] PASSED tests/commands/highlight.py::HighlightCommand::test_cmd_highlight \ntests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \n[gw0] [ 40%] PASSED tests/commands/name_break.py::NameBreakCommand::test_cmd_name_break \ntests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \n[gw2] [ 40%] PASSED tests/commands/hijack_fd.py::HijackFdCommand::test_cmd_hijack_fd \ntests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \n[gw3] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_min_size_filter \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \n[gw1] [ 41%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_fast \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \n[gw2] [ 42%] PASSED tests/commands/ksymaddr.py::KsymaddrCommand::test_cmd_ksymaddr \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_list \n[gw1] [ 43%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_large \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \n[gw0] [ 43%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_as_bytes_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes \n[gw3] [ 44%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_mult_heaps \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \n[gw2] [ 45%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_list \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \n[gw1] [ 45%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_non_main \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \n[gw0] [ 46%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \n[gw3] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \n[gw1] [ 47%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_small \ntests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \n[gw0] [ 48%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \n[gw2] [ 49%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_reset \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \n[gw1] [ 49%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_bins_tcache \ntests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \n[gw3] [ 50%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_summary_with_type_resolved \ntests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \n[gw0] [ 50%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \n[gw2] [ 51%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_unwatch \ntests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \n[gw1] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_check_b_and_n_same_time \ntests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \n[gw0] [ 52%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_arg_nops_no_fit_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \n[gw1] [ 53%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_force_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \n[gw0] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \n[gw1] [ 54%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \n[gw3] [ 55%] PASSED tests/commands/heap.py::HeapCommand::test_cmd_heap_chunks_with_count \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \n[gw0] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_bytes_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \n[gw1] [ 56%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_i_arg_reaching_unmapped_area \ntests/commands/nop.py::NopCommand::test_cmd_nop_inactive \n[gw3] [ 57%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg_break_instruction \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop \n[gw1] [ 58%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_inactive \ntests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \n[gw0] [ 58%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword_symbol \ntests/commands/patch.py::PatchCommand::test_cmd_patch_string \n[gw3] [ 59%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \n[gw1] [ 60%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_invalid_end_address \ntests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \n[gw0] [ 60%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_string \ntests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw3] [ 61%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \n[gw1] [ 61%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_no_arg \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \n[gw0] [ 62%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_word \n[gw3] [ 63%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \n[gw1] [ 63%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom \ntests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \n[gw3] [ 64%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_arg_multibnop_breaks_force \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \n[gw0] [ 65%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_create \ntests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \n[gw3] [ 65%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \n[gw2] [ 66%] PASSED tests/commands/memory.py::MemoryCommand::test_cmd_memory_watch \ntests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \n[gw3] [ 67%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_delete \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \n[gw2] [ 67%] PASSED tests/commands/nop.py::NopCommand::test_cmd_nop_nop_break_instruction_force \ntests/commands/patch.py::PatchCommand::test_cmd_patch \n[gw3] [ 68%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_run \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \n[gw2] [ 69%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte \n[gw1] [ 69%] PASSED tests/commands/pcustom.py::PcustomCommand::test_cmd_pcustom_show \ntests/commands/pie.py::PieCommand::test_cmd_pie \n[gw2] [ 70%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte \ntests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \n[gw1] [ 70%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie \ntests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \n[gw2] [ 71%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_byte_bytearray \ntests/commands/patch.py::PatchCommand::test_cmd_patch_dword \n[gw0] [ 72%] PASSED tests/commands/pattern.py::PatternCommand::test_cmd_pattern_search \ntests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \n[gw1] [ 72%] PASSED tests/commands/pie.py::PieCommand::test_cmd_pie_breakpoint_check \ntests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \n[gw2] [ 73%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_dword \ntests/commands/patch.py::PatchCommand::test_cmd_patch_qword \n[gw1] [ 74%] PASSED tests/commands/reset_cache.py::ResetCacheCommand::test_cmd_reset_cache \ntests/commands/scan.py::ScanCommand::test_cmd_scan \n[gw2] [ 74%] PASSED tests/commands/patch.py::PatchCommand::test_cmd_patch_qword \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \n[gw3] [ 75%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format \ntests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \n[gw0] [ 76%] PASSED tests/commands/process_search.py::ProcessSearchCommand::test_cmd_process_search \ntests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \n[gw2] [ 76%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern_regex \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \n[gw3] [ 77%] PASSED tests/commands/print_format.py::PrintFormatCommand::test_cmd_print_format_bytearray \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw0] [ 78%] PASSED tests/commands/process_status.py::ProcessStatusCommand::test_cmd_process_status \ntests/commands/registers.py::RegistersCommand::test_cmd_registers \n[gw2] [ 78%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \n[gw1] [ 79%] PASSED tests/commands/scan.py::ScanCommand::test_cmd_scan \ntests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \n[gw0] [ 80%] PASSED tests/commands/registers.py::RegistersCommand::test_cmd_registers \ntests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \n[gw3] [ 80%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_ok \n[gw2] [ 81%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_get_nok \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \n[gw1] [ 81%] PASSED tests/commands/search_pattern.py::SearchPatternCommand::test_cmd_search_pattern \ntests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \n[gw0] [ 82%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_nop_inactive \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \n[gw2] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_skip_two_instructions \ntests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \n[gw0] [ 83%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_no_arg \ntests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \n[gw3] [ 84%] PASSED tests/commands/shellcode.py::ShellcodeCommand::test_cmd_shellcode_search \ntests/commands/stub.py::StubCommand::test_cmd_stub \n[gw2] [ 85%] PASSED tests/commands/skipi.py::SkipiCommand::test_cmd_skipi_two_instructions_from_location \ntests/commands/version.py::VersionCommand::test_cmd_version \n[gw2] [ 85%] PASSED tests/commands/version.py::VersionCommand::test_cmd_version \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \n[gw0] [ 86%] PASSED tests/commands/trace_run.py::TraceRunCommand::test_cmd_trace_run \ntests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \n[gw1] [ 87%] PASSED tests/commands/smart_eval.py::SmartEvalCommand::test_cmd_smart_eval \ntests/commands/theme.py::ThemeCommand::test_cmd_theme \n[gw3] [ 87%] PASSED tests/commands/stub.py::StubCommand::test_cmd_stub \ntests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \n[gw2] [ 88%] PASSED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo \ntests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \n[gw3] [ 89%] PASSED tests/commands/xfiles.py::XfilesCommand::test_cmd_xfiles \ntests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \n[gw0] [ 89%] PASSED tests/commands/vmmap.py::VmmapCommand::test_cmd_vmmap \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \n[gw2] [ 90%] FAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class \ntests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \n[gw3] [ 90%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_hook_validator \ntests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \n[gw0] [ 91%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_display \ntests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \n[gw3] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_show_opcodes_size \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \n[gw2] [ 92%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_libc_version \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \n[gw3] [ 93%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_bss \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \n[gw0] [ 94%] PASSED tests/config/__init__.py::TestGefConfigUnit::test_config_type_validator \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \n[gw0] [ 94%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_got \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \n[gw2] [ 95%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_base \ntests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw3] [ 96%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_heap \ntests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw0] [ 96%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_context_correct_registers_refresh_with_frames \ntests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw2] [ 97%] PASSED tests/functions/elf_sections.py::ElfSectionGdbFunction::test_func_stack \n[gw0] [ 98%] PASSED tests/regressions/registers_register_order.py::RegressionRegisterOrder::test_registers_show_registers_in_correct_order \n[gw3] [ 98%] PASSED tests/regressions/gdbserver_connection.py::RegressionGdbserverConnection::test_can_establish_connection_to_gdbserver_again_after_disconnect \n[gw1] [ 99%] PASSED tests/commands/theme.py::ThemeCommand::test_cmd_theme \ntests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n[gw1] [100%] PASSED tests/commands/xor_memory.py::XorMemoryCommand::test_cmd_xor_memory_patch \n\n=================================== FAILURES ===================================\n_____________________ XinfoCommand.test_cmd_xinfo_on_class _____________________\n[gw2] linux -- Python 3.8.10 /usr/bin/python3.8\nself = <tests.commands.xinfo.XinfoCommand testMethod=test_cmd_xinfo_on_class>\n\n def test_cmd_xinfo_on_class(self):\n/home/runner/.local/lib/python3.8/site-packages/pytest_benchmark/logger.py:46: PytestBenchmarkWarning: Benchmarks are automatically disabled because xdist plugin is active.Benchmarks cannot be performed reliably in a parallelized environment.\n warner(PytestBenchmarkWarning(text))\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n self.assertNoException(res)\n> self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\nE AssertionError: 'Symbol: B<TraitA, TraitB>::Run()+4' not found in 'Reading symbols from /tmp/class.out...\\nError while writing index for `/tmp/class.out\\': mkstemp: No such file or directory.\\nGEF for linux ready, type `gef\\' to start, `gef config\\' to configure\\n89 commands loaded and 5 functions added for GDB 10.2 in 0.00ms using Python engine 3.8\\nBreakpoint 1 at 0x1328: file class.cpp, line 21.\\nStarting program: /tmp/class.out \\n\\nBreakpoint 1, B<TraitA, TraitB>::Run (this=0x55555556aeb0) at class.cpp:21\\n21\\t virtual void Run() { printf(\"I am B\\\\n\"); }\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\n\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500 xinfo: 0x55555555532c \u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\u2500\\nPage: 0x0000555555555000 \u2192 0x0000555555556000 (size=0x1000)\\nPermissions: r-x\\nPathname: /tmp/class.out\\nOffset (from page): 0x32c\\nInode: 36000\\nSegment: .text (0x00005555555550a0-0x00005555555553b5)\\nOffset (from segment): 0x28c\\nSymbol: B<TraitA, TraitB>::Run()+20'\n\ntests/commands/xinfo.py:27: AssertionError\n=========================== short test summary info ============================\nFAILED tests/commands/xinfo.py::XinfoCommand::test_cmd_xinfo_on_class\n================== 1 failed, 153 passed, 1 skipped in 49.25s ===================\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/tests/commands/xinfo.py b/tests/commands/xinfo.py\nindex 625615b..ea50e23 100644\n--- a/tests/commands/xinfo.py\n+++ b/tests/commands/xinfo.py\n@@ -22,6 +22,6 @@ class XinfoCommand(GefUnitTestGeneric):\n def test_cmd_xinfo_on_class(self):\n cmd = \"xinfo $pc+4\"\n target = debug_target(\"class\")\n- res = gdb_run_silent_cmd(cmd, target=target, before=[\"b B<TraitA, TraitB>::Run()\"])\n+ res = gdb_run_silent_cmd(cmd, target=target, before=[\"b *'B<TraitA, TraitB>::Run'\"])\n self.assertNoException(res)\n self.assertIn(\"Symbol: B<TraitA, TraitB>::Run()+4\", res)\n", "difficulty": 2, "changed_files": ["tests/commands/xinfo.py"], "commit_link": "https://github.com/hugsy/gef/tree/1b22eb71da460e7d3a301842f65c256b920f6195"}
|
|
|
|
data/python/1e48195.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 120, "repo_owner": "wandb", "repo_name": "wandb", "head_branch": "update/diffusers", "workflow_name": "pre-commit", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "wandb", "sha_fail": "1e48195d38e0d8de001e0f55bfbc830abf2e3f41", "sha_success": "5c24819003ab2cf1fdaa07e78a120fbf53875392", "workflow": "name: pre-commit\n\non:\n pull_request:\n push:\n branches: [main]\n\njobs:\n pre-commit:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-go@v4\n with:\n go-version: '1.21'\n - uses: actions/setup-python@v3\n - name: setup env\n run: |\n ./core/scripts/code-checks.sh update\n - uses: pre-commit/[email protected]\n with:\n extra_args: --hook-stage pre-push --all-files\n env:\n SKIP: go-generate-graphql\n", "logs": [{"step_name": "pre-commit/6_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --hook-stage pre-push --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.3-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 31.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 31.0 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 87.6 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 116.8 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 49.6 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.3-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 92.1 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.3 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.3\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nReceived 0 of 48931996 (0.0%), 0.0 MBs/sec\nCache Size: ~47 MB (48931996 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/2e6b382b-db1b-4411-add2-ddbcf402d5ce/cache.tzst -P -C /home/runner/work/wandb/wandb --use-compress-program unzstd\nReceived 48931996 of 48931996 (100.0%), 23.3 MBs/sec\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|a333fd88224707918ef5e53726a9b083fdaa934e65d17db828a35f3ab158770e\n##[group]Run pre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --hook-stage pre-push --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n SKIP: go-generate-graphql\n##[endgroup]\nclippy...................................................................\u001b[42mPassed\u001b[m\nruff.....................................................................\u001b[42mPassed\u001b[m\nblack-jupyter............................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: black-jupyter\u001b[m\n\u001b[2m- files were modified by this hook\u001b[m\n\n\u001b[1mreformatted wandb/integration/diffusers/autologger.py\u001b[0m\n\n\u001b[1mAll done! \u2728 \ud83c\udf70 \u2728\u001b[0m\n\u001b[34m\u001b[1m1 file \u001b[0m\u001b[1mreformatted\u001b[0m, \u001b[34m887 files \u001b[0mleft unchanged.\n\nblacken-docs.............................................................\u001b[42mPassed\u001b[m\nclang-format.............................................................\u001b[42mPassed\u001b[m\ncheck for merge conflicts................................................\u001b[42mPassed\u001b[m\ntrailing-whitespace......................................................\u001b[42mPassed\u001b[m\nend-of-file-fixer........................................................\u001b[42mPassed\u001b[m\ncheck-yaml...............................................................\u001b[42mPassed\u001b[m\ncheck-added-large-files..................................................\u001b[42mPassed\u001b[m\ncheck-shebang-scripts-are-executable.....................................\u001b[42mPassed\u001b[m\ncheck-executables-have-shebangs..........................................\u001b[42mPassed\u001b[m\ncheck-symlinks.......................................(no files to check)\u001b[46;30mSkipped\u001b[m\ncheck-toml...............................................................\u001b[42mPassed\u001b[m\ndebug-statements.........................................................\u001b[42mPassed\u001b[m\nforbid-submodules....................................(no files to check)\u001b[46;30mSkipped\u001b[m\nblack....................................................................\u001b[42mPassed\u001b[m\ngo-generate-proto........................................................\u001b[42mPassed\u001b[m\ngo-generate-graphql.....................................................\u001b[43;30mSkipped\u001b[m\ngo-fmt...................................................................\u001b[42mPassed\u001b[m\ngo-imports...............................................................\u001b[42mPassed\u001b[m\ngo-unit-tests............................................................\u001b[42mPassed\u001b[m\ngo-vet...................................................................\u001b[42mPassed\u001b[m\ngolangci-lint............................................................\u001b[42mPassed\u001b[m\ngo-critic................................................................\u001b[42mPassed\u001b[m\ngo-cyclo.................................................................\u001b[42mPassed\u001b[m\ngo-mod-tidy..............................................................\u001b[42mPassed\u001b[m\ngo-build.................................................................\u001b[42mPassed\u001b[m\npre-commit hook(s) made changes.\nIf you are seeing this message in CI, reproduce locally with: `pre-commit run --all-files`.\nTo run `pre-commit` as part of git workflow, use `pre-commit install`.\nAll changes made by hooks:\n\u001b[1mdiff --git a/wandb/integration/diffusers/autologger.py b/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[1mindex ab88d9c..e10ee15 100644\u001b[m\n\u001b[1m--- a/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[1m+++ b/wandb/integration/diffusers/autologger.py\u001b[m\n\u001b[36m@@ -65,7 +65,7 @@\u001b[m \u001b[mautolog = AutologAPI(\u001b[m\n \"TextToVideoSDPipeline.__call__\",\u001b[m\n \"TextToVideoZeroPipeline.__call__\",\u001b[m\n \"StableVideoDiffusionPipeline.__call__\",\u001b[m\n\u001b[31m- \"AmusedPipeline.__call__\"\u001b[m\n\u001b[32m+\u001b[m\u001b[32m \"AmusedPipeline.__call__\",\u001b[m\n ),\u001b[m\n resolver=DiffusersPipelineResolver(),\u001b[m\n telemetry_feature=\"diffusers_autolog\",\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/wandb/integration/diffusers/autologger.py b/wandb/integration/diffusers/autologger.py\nindex ab88d9c7d..e10ee1549 100644\n--- a/wandb/integration/diffusers/autologger.py\n+++ b/wandb/integration/diffusers/autologger.py\n@@ -65,7 +65,7 @@ autolog = AutologAPI(\n \"TextToVideoSDPipeline.__call__\",\n \"TextToVideoZeroPipeline.__call__\",\n \"StableVideoDiffusionPipeline.__call__\",\n- \"AmusedPipeline.__call__\"\n+ \"AmusedPipeline.__call__\",\n ),\n resolver=DiffusersPipelineResolver(),\n telemetry_feature=\"diffusers_autolog\",\ndiff --git a/wandb/integration/diffusers/pipeline_resolver.py b/wandb/integration/diffusers/pipeline_resolver.py\nindex 37cf75fd7..9d3736d2d 100644\n--- a/wandb/integration/diffusers/pipeline_resolver.py\n+++ b/wandb/integration/diffusers/pipeline_resolver.py\n@@ -18,6 +18,7 @@ class DiffusersPipelineResolver:\n \n def __init__(self) -> None:\n self.wandb_table = None\n+ self.pipeline_call_count = 1\n \n def __call__(\n self,\n@@ -43,7 +44,10 @@ class DiffusersPipelineResolver:\n pipeline_name = args[0].__class__.__name__\n resolver = None\n if pipeline_name in SUPPORTED_MULTIMODAL_PIPELINES:\n- resolver = DiffusersMultiModalPipelineResolver(pipeline_name)\n+ resolver = DiffusersMultiModalPipelineResolver(\n+ pipeline_name, self.pipeline_call_count\n+ )\n+ self.pipeline_call_count += 1\n elif pipeline_name in SUPPORTED_SDXL_PIPELINES:\n resolver = SDXLResolver(pipeline_name)\n loggable_dict = resolver(args, kwargs, response, start_time, time_elapsed)\ndiff --git a/wandb/integration/diffusers/resolvers/multimodal.py b/wandb/integration/diffusers/resolvers/multimodal.py\nindex 2948778f9..021efcf2c 100644\n--- a/wandb/integration/diffusers/resolvers/multimodal.py\n+++ b/wandb/integration/diffusers/resolvers/multimodal.py\n@@ -502,8 +502,9 @@ class DiffusersMultiModalPipelineResolver:\n pipeline_name: (str) The name of the Diffusion Pipeline.\n \"\"\"\n \n- def __init__(self, pipeline_name: str) -> None:\n+ def __init__(self, pipeline_name: str, pipeline_call_count: int) -> None:\n self.pipeline_name = pipeline_name\n+ self.pipeline_call_count = pipeline_call_count\n columns = []\n if pipeline_name in SUPPORTED_MULTIMODAL_PIPELINES:\n columns += SUPPORTED_MULTIMODAL_PIPELINES[pipeline_name][\"table-schema\"]\n@@ -543,9 +544,30 @@ class DiffusersMultiModalPipelineResolver:\n pipeline_configs = dict(pipeline.config)\n pipeline_configs[\"pipeline-name\"] = self.pipeline_name\n \n- wandb.config.update(\n- {\"workflow\": {\"pipeline\": pipeline_configs, \"params\": kwargs}}\n- )\n+ if \"workflow\" not in wandb.config:\n+ wandb.config.update(\n+ {\n+ \"workflow\": [\n+ {\n+ \"pipeline\": pipeline_configs,\n+ \"params\": kwargs,\n+ \"stage\": f\"Pipeline-Call-{self.pipeline_call_count}\",\n+ }\n+ ]\n+ }\n+ )\n+ else:\n+ existing_workflow = wandb.config.workflow\n+ updated_workflow = existing_workflow + [\n+ {\n+ \"pipeline\": pipeline_configs,\n+ \"params\": kwargs,\n+ \"stage\": f\"Pipeline-Call-{self.pipeline_call_count}\",\n+ }\n+ ]\n+ wandb.config.update(\n+ {\"workflow\": updated_workflow}, allow_val_change=True\n+ )\n \n # Return the WandB loggable dict\n loggable_dict = self.prepare_loggable_dict(response, kwargs)\n@@ -596,7 +618,13 @@ class DiffusersMultiModalPipelineResolver:\n caption = loggable_kwarg_chunks[prompt_index][idx]\n except ValueError:\n caption = None\n- wandb.log({\"Generated-Image\": wandb.Image(image, caption=caption)})\n+ wandb.log(\n+ {\n+ f\"Generated-Image/Pipeline-Call-{self.pipeline_call_count}\": wandb.Image(\n+ image, caption=caption\n+ )\n+ }\n+ )\n else:\n if (\n SUPPORTED_MULTIMODAL_PIPELINES[self.pipeline_name][\"output-type\"]\n@@ -611,7 +639,7 @@ class DiffusersMultiModalPipelineResolver:\n caption = None\n wandb.log(\n {\n- \"Generated-Video\": wandb.Video(\n+ f\"Generated-Video/Pipeline-Call-{self.pipeline_call_count}\": wandb.Video(\n postprocess_pils_to_np(image), fps=4, caption=caption\n )\n }\n@@ -629,7 +657,7 @@ class DiffusersMultiModalPipelineResolver:\n caption = None\n wandb.log(\n {\n- \"Generated-Audio\": wandb.Audio(\n+ f\"Generated-Audio/Pipeline-Call-{self.pipeline_call_count}\": wandb.Audio(\n image, sample_rate=16000, caption=caption\n )\n }\n@@ -696,7 +724,11 @@ class DiffusersMultiModalPipelineResolver:\n images, normalize=self.pipeline_name == \"TextToVideoZeroPipeline\"\n )\n wandb.log(\n- {\"Generated-Video\": wandb.Video(video, fps=4, caption=kwargs[\"prompt\"])}\n+ {\n+ f\"Generated-Video/Pipeline-Call-{self.pipeline_call_count}\": wandb.Video(\n+ video, fps=4, caption=kwargs[\"prompt\"]\n+ )\n+ }\n )\n loggable_kwarg_ids = SUPPORTED_MULTIMODAL_PIPELINES[self.pipeline_name][\n \"kwarg-logging\"\n@@ -727,5 +759,6 @@ class DiffusersMultiModalPipelineResolver:\n self.log_media(image, loggable_kwarg_chunks, idx)\n # Populate the row of the wandb_table\n self.add_data_to_table(image, loggable_kwarg_chunks, idx)\n-\n- return {\"Result-Table\": self.wandb_table}\n+ return {\n+ f\"Result-Table/Pipeline-Call-{self.pipeline_call_count}\": self.wandb_table\n+ }\n", "difficulty": 0, "changed_files": ["wandb/integration/diffusers/autologger.py", "wandb/integration/diffusers/pipeline_resolver.py", "wandb/integration/diffusers/resolvers/multimodal.py"], "commit_link": "https://github.com/wandb/wandb/tree/1e48195d38e0d8de001e0f55bfbc830abf2e3f41"}
|
|
|
|
data/python/212ff27.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 52, "repo_owner": "yt-dlp", "repo_name": "yt-dlp", "head_branch": "mx3", "workflow_name": "Quick Test", "workflow_filename": "quick-test.yml", "workflow_path": ".github/workflows/quick-test.yml", "contributor": "martinxyz", "sha_fail": "212ff27cb59b69bad24a2e675d97e52ead5b12e8", "sha_success": "1cb1df5d401172bb7b6770605a72fdb9bd03b0a8", "workflow": "name: Quick Test\non: [push, pull_request]\npermissions:\n contents: read\n\njobs:\n tests:\n name: Core Test\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - name: Set up Python 3.11\n uses: actions/setup-python@v4\n with:\n python-version: '3.11'\n - name: Install test requirements\n run: pip install pytest -r requirements.txt\n - name: Run tests\n run: |\n python3 -m yt_dlp -v || true\n ./devscripts/run_tests.sh core\n flake8:\n name: Linter\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v4\n - name: Install flake8\n run: pip install flake8\n - name: Make lazy extractors\n run: python devscripts/make_lazy_extractors.py\n - name: Run flake8\n run: flake8 .\n", "logs": [{"step_name": "Core Test/5_Run tests.txt", "log": "##[group]Run python3 -m yt_dlp -v || true\n\u001b[36;1mpython3 -m yt_dlp -v || true\u001b[0m\n\u001b[36;1mpython3 ./devscripts/run_tests.py core\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n[debug] Command-line config: ['-v']\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/runpy.py\", line 194, in _run_module_as_main\n return _run_code(code, main_globals, None,\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__main__.py\", line 17, in <module>\n yt_dlp.main()\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__init__.py\", line 1009, in main\n _exit(*variadic(_real_main(argv)))\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/__init__.py\", line 963, in _real_main\n with YoutubeDL(ydl_opts) as ydl:\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/YoutubeDL.py\", line 689, in __init__\n self.print_debug_header()\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/YoutubeDL.py\", line 3919, in print_debug_header\n from .extractor.extractors import _LAZY_LOADER\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/extractors.py\", line 17, in <module>\n from ._extractors import * # noqa: F403\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/_extractors.py\", line 1140, in <module>\n from .mx3 import (\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 80, in <module>\n class Mx3IE(Mx3BaseIE):\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 82, in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\nNameError: name '_Mx3BaseIE' is not defined\nRunning ['pytest', '-Werror', '--tb=short', '--color=yes', '-m', 'not download']\n\u001b[1m============================= test session starts ==============================\u001b[0m\nplatform linux -- Python 3.8.18, pytest-7.4.4, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yt-dlp/yt-dlp\nconfigfile: setup.cfg\n\u001b[1mcollecting ... \u001b[0mcollected 653 items / 9 errors / 6 deselected / 647 selected\n\n==================================== ERRORS ====================================\n\u001b[31m\u001b[1m_________________ ERROR collecting test/test_InfoExtractor.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_InfoExtractor.py\u001b[0m:16: in <module>\n from yt_dlp.extractor import YoutubeIE, get_info_extractor\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m___________________ ERROR collecting test/test_YoutubeDL.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_YoutubeDL.py\u001b[0m:17: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m____________________ ERROR collecting test/test_all_urls.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_all_urls.py\u001b[0m:14: in <module>\n from yt_dlp.extractor import FacebookIE, YoutubeIE, gen_extractors\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m____________________ ERROR collecting test/test_download.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_download.py\u001b[0m:286: in <module>\n inject_tests(normal_test_cases)\n\u001b[1m\u001b[31mtest/test_download.py\u001b[0m:275: in inject_tests\n for test_case in test_cases:\n\u001b[1m\u001b[31mtest/helper.py\u001b[0m:91: in gettestcases\n for ie in yt_dlp.extractor.gen_extractors():\n\u001b[1m\u001b[31myt_dlp/extractor/__init__.py\u001b[0m:20: in gen_extractors\n return [klass() for klass in gen_extractor_classes()]\n\u001b[1m\u001b[31myt_dlp/extractor/__init__.py\u001b[0m:11: in gen_extractor_classes\n from .extractors import _ALL_CLASSES\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_____________ ERROR collecting test/test_iqiyi_sdk_interpreter.py ______________\u001b[0m\n\u001b[1m\u001b[31mtest/test_iqiyi_sdk_interpreter.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import IqiyiIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m___________________ ERROR collecting test/test_subtitles.py ____________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_subtitles.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import (\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_________________ ERROR collecting test/test_youtube_lists.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_lists.py\u001b[0m:12: in <module>\n from yt_dlp.extractor import YoutubeIE, YoutubeTabIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m__________________ ERROR collecting test/test_youtube_misc.py __________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_misc.py\u001b[0m:11: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[31m\u001b[1m_______________ ERROR collecting test/test_youtube_signature.py ________________\u001b[0m\n\u001b[1m\u001b[31mtest/test_youtube_signature.py\u001b[0m:17: in <module>\n from yt_dlp.extractor import YoutubeIE\n\u001b[1m\u001b[31m<frozen importlib._bootstrap>\u001b[0m:1039: in _handle_fromlist\n ???\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:36: in __getattribute__\n ret = super().__getattribute__(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:54: in __getattr__\n ret = from_child(attr)\n\u001b[1m\u001b[31myt_dlp/compat/compat_utils.py\u001b[0m:68: in from_child\n child = importlib.import_module(child, parent.__name__)\n\u001b[1m\u001b[31m/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py\u001b[0m:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n\u001b[1m\u001b[31myt_dlp/extractor/extractors.py\u001b[0m:17: in <module>\n from ._extractors import * # noqa: F403\n\u001b[1m\u001b[31myt_dlp/extractor/_extractors.py\u001b[0m:1140: in <module>\n from .mx3 import (\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:80: in <module>\n class Mx3IE(Mx3BaseIE):\n\u001b[1m\u001b[31myt_dlp/extractor/mx3.py\u001b[0m:82: in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n\u001b[1m\u001b[31mE NameError: name '_Mx3BaseIE' is not defined\u001b[0m\n\u001b[36m\u001b[1m=========================== short test summary info ============================\u001b[0m\n\u001b[31mERROR\u001b[0m test/test_InfoExtractor.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_YoutubeDL.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_all_urls.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_download.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_iqiyi_sdk_interpreter.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_subtitles.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_lists.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_misc.py - NameError: name '_Mx3BaseIE' is not defined\n\u001b[31mERROR\u001b[0m test/test_youtube_signature.py - NameError: name '_Mx3BaseIE' is not defined\n!!!!!!!!!!!!!!!!!!! Interrupted: 9 errors during collection !!!!!!!!!!!!!!!!!!!!\n\u001b[31m======================= \u001b[33m6 deselected\u001b[0m, \u001b[31m\u001b[1m9 errors\u001b[0m\u001b[31m in 1.32s\u001b[0m\u001b[31m ========================\u001b[0m\n##[error]Process completed with exit code 2.\n"}, {"step_name": "Linter/5_Make lazy extractors.txt", "log": "##[group]Run python devscripts/make_lazy_extractors.py\n\u001b[36;1mpython devscripts/make_lazy_extractors.py\u001b[0m\nshell: /usr/bin/bash -e {0}\n##[endgroup]\nTraceback (most recent call last):\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 132, in <module>\n main()\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 41, in main\n _ALL_CLASSES = get_all_ies() # Must be before import\n File \"/home/runner/work/yt-dlp/yt-dlp/devscripts/make_lazy_extractors.py\", line 68, in get_all_ies\n from yt_dlp.extractor.extractors import _ALL_CLASSES\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/extractors.py\", line 17, in <module>\n from ._extractors import * # noqa: F403\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/_extractors.py\", line 1140, in <module>\n from .mx3 import (\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 80, in <module>\n class Mx3IE(Mx3BaseIE):\n File \"/home/runner/work/yt-dlp/yt-dlp/yt_dlp/extractor/mx3.py\", line 82, in Mx3IE\n _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\nNameError: name '_Mx3BaseIE' is not defined. Did you mean: 'Mx3BaseIE'?\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/yt_dlp/extractor/mx3.py b/yt_dlp/extractor/mx3.py\nindex b4fd5ab72..e39334a6e 100644\n--- a/yt_dlp/extractor/mx3.py\n+++ b/yt_dlp/extractor/mx3.py\n@@ -79,7 +79,7 @@ def get_info_field(name):\n \n class Mx3IE(Mx3BaseIE):\n _DOMAIN = 'mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://mx3.ch/t/1Cru',\n 'md5': '7ba09e9826b4447d4e1ce9d69e0e295f',\n@@ -131,7 +131,7 @@ class Mx3IE(Mx3BaseIE):\n \n class Mx3NeoIE(Mx3BaseIE):\n _DOMAIN = 'neo.mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://neo.mx3.ch/t/1hpd',\n 'md5': '6d9986bbae5cac3296ec8813bf965eb2',\n@@ -152,7 +152,7 @@ class Mx3NeoIE(Mx3BaseIE):\n \n class Mx3VolksmusikIE(Mx3BaseIE):\n _DOMAIN = 'volksmusik.mx3.ch'\n- _VALID_URL = _Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n+ _VALID_URL = Mx3BaseIE._VALID_URL_TMPL % re.escape(_DOMAIN)\n _TESTS = [{\n 'url': 'https://volksmusik.mx3.ch/t/Zx',\n 'md5': 'dd967a7b0c1ef898f3e072cf9c2eae3c',\n", "difficulty": 1, "changed_files": ["yt_dlp/extractor/mx3.py"], "commit_link": "https://github.com/yt-dlp/yt-dlp/tree/212ff27cb59b69bad24a2e675d97e52ead5b12e8"}
|
|
|
|
data/python/2201be2.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 159, "repo_owner": "mwaskom", "repo_name": "seaborn", "head_branch": "plot/layout_rect", "workflow_name": "CI", "workflow_filename": "ci.yaml", "workflow_path": ".github/workflows/ci.yaml", "contributor": "mwaskom", "sha_fail": "2201be21886bb82201f3c3487f5f1468f6e6ac81", "sha_success": "596ad4f85f31f39bcedd2fbca935647b1a2fdf84", "workflow": "name: CI\n\non:\n push:\n branches: [master, v0.*]\n pull_request:\n branches: master\n schedule:\n - cron: '0 6 * * 1,4' # Each Monday and Thursday at 06:00 UTC\n workflow_dispatch:\n\npermissions:\n contents: read\n\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: ${{ github.workspace }}/seaborn-data\n\njobs:\n build-docs:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python 3.11\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n with:\n python-version: \"3.11\"\n\n - name: Install seaborn\n run: |\n pip install --upgrade pip\n pip install .[stats,docs]\n\n - name: Install pandoc\n run: |\n sudo apt-get install pandoc\n\n - name: Cache datasets\n run: |\n git clone https://github.com/mwaskom/seaborn-data.git\n ls $SEABORN_DATA\n\n - name: Build docs\n env:\n SPHINXOPTS: -j `nproc`\n run: |\n cd doc\n make -j `nproc` notebooks\n make html\n\n\n run-tests:\n runs-on: ubuntu-latest\n\n strategy:\n matrix:\n python: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n install: [full]\n deps: [latest]\n\n include:\n - python: \"3.8\"\n install: full\n deps: pinned\n - python: \"3.11\"\n install: light\n deps: latest\n\n steps:\n - uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python ${{ matrix.python }}\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n with:\n python-version: ${{ matrix.python }}\n allow-prereleases: true\n\n - name: Install seaborn\n run: |\n pip install --upgrade pip wheel\n if [[ ${{matrix.install}} == 'full' ]]; then EXTRAS=',stats'; fi\n if [[ ${{matrix.deps }} == 'pinned' ]]; then DEPS='-r ci/deps_pinned.txt'; fi\n pip install .[dev$EXTRAS] $DEPS\n\n - name: Run tests\n run: make test\n\n - name: Upload coverage\n uses: codecov/codecov-action@eaaf4bedf32dbdc6b720b63067d99c4d77d6047d # v3.1.4\n if: ${{ success() }}\n\n lint:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n steps:\n\n - name: Checkout\n uses: actions/checkout@b4ffde65f46336ab88eb53be808477a3936bae11 # v4.1.1\n\n - name: Setup Python\n uses: actions/setup-python@65d7f2d534ac1bc67fcd62888c5f4f3d2cb2b236 # v4.7.1\n\n - name: Install tools\n run: pip install mypy flake8\n\n - name: Flake8\n run: make lint\n\n - name: Type checking\n run: make typecheck\n", "logs": [{"step_name": "build-docs/7_Build docs.txt", "log": "##[group]Run cd doc\n\u001b[36;1mcd doc\u001b[0m\n\u001b[36;1mmake -j `nproc` notebooks\u001b[0m\n\u001b[36;1mmake html\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n SPHINXOPTS: -j `nproc`\n##[endgroup]\nmake[1]: Entering directory '/home/runner/work/seaborn/seaborn/doc/_tutorial'\n../tools/nb_to_doc.py aesthetics.ipynb ../tutorial\n../tools/nb_to_doc.py axis_grids.ipynb ../tutorial\nmake[1]: Entering directory '/home/runner/work/seaborn/seaborn/doc/_docstrings'\n../tools/nb_to_doc.py FacetGrid.ipynb ../docstrings\n../tools/nb_to_doc.py categorical.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py JointGrid.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py PairGrid.ipynb ../docstrings\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py color_palettes.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py axes_style.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py data_structure.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py barplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py blend_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py distributions.ipynb ../tutorial\n../tools/nb_to_doc.py boxenplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py error_bars.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py boxplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py function_overview.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py catplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py clustermap.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py color_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py introduction.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py countplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py cubehelix_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py dark_palette.ipynb ../docstrings\n../tools/nb_to_doc.py objects_interface.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py displot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py diverging_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py ecdfplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py heatmap.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py histplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py hls_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py properties.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py husl_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py jointplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py regression.ipynb ../tutorial\n../tools/nb_to_doc.py kdeplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py relational.ipynb ../tutorial\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py light_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py lineplot.ipynb ../docstrings\n../tools/nb_to_doc.py lmplot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py move_legend.ipynb ../docstrings\n../tools/nb_to_doc.py mpl_palette.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Agg.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Area.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Band.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Bar.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\nmake[1]: Leaving directory '/home/runner/work/seaborn/seaborn/doc/_tutorial'\n../tools/nb_to_doc.py objects.Bars.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Count.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Dash.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Dodge.ipynb ../docstrings\n../tools/nb_to_doc.py objects.Dot.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Dots.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Est.ipynb ../docstrings\n0.01s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Hist.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Jitter.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.KDE.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Line.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Lines.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Norm.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Path.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Paths.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Perc.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.add.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.config.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.01s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.facet.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.label.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.layout.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n../tools/nb_to_doc.py objects.Plot.limit.ipynb ../docstrings\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n../tools/nb_to_doc.py objects.Plot.on.ipynb ../docstrings\nTraceback (most recent call last):\n File \"/home/runner/work/seaborn/seaborn/doc/_docstrings/../tools/nb_to_doc.py\", line 126, in <module>\n ep.preprocess(nb, {\"metadata\": {\"path\": basedir}})\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/preprocessors/execute.py\", line 102, in preprocess\n self.preprocess_cell(cell, resources, index)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/preprocessors/execute.py\", line 123, in preprocess_cell\n cell = self.execute_cell(cell, index, store_history=True)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/jupyter_core/utils/__init__.py\", line 173, in wrapped\n return loop.run_until_complete(inner)\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/asyncio/base_events.py\", line 653, in run_until_complete\n return future.result()\n ^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbclient/client.py\", line 1062, in async_execute_cell\n await self._check_raise_for_error(cell, cell_index, exec_reply)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbclient/client.py\", line 918, in _check_raise_for_error\n raise CellExecutionError.from_cell_and_msg(cell, exec_reply_content)\nnbclient.exceptions.CellExecutionError: An error occurred while executing the following cell:\n------------------\np.layout(extent=[0, 0, .8, 1]).show()\n------------------\n\n\n\u001b[0;31m---------------------------------------------------------------------------\u001b[0m\n\u001b[0;31mNameError\u001b[0m Traceback (most recent call last)\nCell \u001b[0;32mIn[5], line 1\u001b[0m\n\u001b[0;32m----> 1\u001b[0m \u001b[43mp\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mlayout\u001b[49m\u001b[43m(\u001b[49m\u001b[43mextent\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m[\u001b[49m\u001b[38;5;241;43m0\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m0\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m.8\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m1\u001b[39;49m\u001b[43m]\u001b[49m\u001b[43m)\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mshow\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:930\u001b[0m, in \u001b[0;36mPlot.show\u001b[0;34m(self, **kwargs)\u001b[0m\n\u001b[1;32m 913\u001b[0m \u001b[38;5;250m\u001b[39m\u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 914\u001b[0m \u001b[38;5;124;03mCompile the plot and display it by hooking into pyplot.\u001b[39;00m\n\u001b[1;32m 915\u001b[0m \n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 922\u001b[0m \n\u001b[1;32m 923\u001b[0m \u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 924\u001b[0m \u001b[38;5;66;03m# TODO make pyplot configurable at the class level, and when not using,\u001b[39;00m\n\u001b[1;32m 925\u001b[0m \u001b[38;5;66;03m# import IPython.display and call on self to populate cell output?\u001b[39;00m\n\u001b[1;32m 926\u001b[0m \n\u001b[1;32m 927\u001b[0m \u001b[38;5;66;03m# Keep an eye on whether matplotlib implements \"attaching\" an existing\u001b[39;00m\n\u001b[1;32m 928\u001b[0m \u001b[38;5;66;03m# figure to pyplot: https://github.com/matplotlib/matplotlib/pull/14024\u001b[39;00m\n\u001b[0;32m--> 930\u001b[0m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mplot\u001b[49m\u001b[43m(\u001b[49m\u001b[43mpyplot\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mTrue\u001b[39;49;00m\u001b[43m)\u001b[49m\u001b[38;5;241m.\u001b[39mshow(\u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:937\u001b[0m, in \u001b[0;36mPlot.plot\u001b[0;34m(self, pyplot)\u001b[0m\n\u001b[1;32m 933\u001b[0m \u001b[38;5;250m\u001b[39m\u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 934\u001b[0m \u001b[38;5;124;03mCompile the plot spec and return the Plotter object.\u001b[39;00m\n\u001b[1;32m 935\u001b[0m \u001b[38;5;124;03m\"\"\"\u001b[39;00m\n\u001b[1;32m 936\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m theme_context(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_theme_with_defaults()):\n\u001b[0;32m--> 937\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_plot\u001b[49m\u001b[43m(\u001b[49m\u001b[43mpyplot\u001b[49m\u001b[43m)\u001b[49m\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:971\u001b[0m, in \u001b[0;36mPlot._plot\u001b[0;34m(self, pyplot)\u001b[0m\n\u001b[1;32m 969\u001b[0m \u001b[38;5;66;03m# Add various figure decorations\u001b[39;00m\n\u001b[1;32m 970\u001b[0m plotter\u001b[38;5;241m.\u001b[39m_make_legend(\u001b[38;5;28mself\u001b[39m)\n\u001b[0;32m--> 971\u001b[0m \u001b[43mplotter\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_finalize_figure\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m)\u001b[49m\n\u001b[1;32m 973\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m plotter\n\nFile \u001b[0;32m/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/seaborn/_core/plot.py:1815\u001b[0m, in \u001b[0;36mPlotter._finalize_figure\u001b[0;34m(self, p)\u001b[0m\n\u001b[1;32m 1812\u001b[0m set_layout_engine(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure, \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtight\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 1814\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m (extent \u001b[38;5;241m:=\u001b[39m p\u001b[38;5;241m.\u001b[39m_layout_spec\u001b[38;5;241m.\u001b[39mget(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mextent\u001b[39m\u001b[38;5;124m\"\u001b[39m)) \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m-> 1815\u001b[0m engine \u001b[38;5;241m=\u001b[39m \u001b[43mget_layout_engine\u001b[49m(\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure)\n\u001b[1;32m 1816\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m engine \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[1;32m 1817\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_figure\u001b[38;5;241m.\u001b[39msubplots_adjust(\u001b[38;5;241m*\u001b[39mextent)\n\n\u001b[0;31mNameError\u001b[0m: name 'get_layout_engine' is not defined\n\nmake[1]: *** [Makefile:7: ../docstrings/objects.Plot.layout.rst] Error 1\nmake[1]: *** Waiting for unfinished jobs....\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\n0.00s - Debugger warning: It seems that frozen modules are being used, which may\n0.00s - make the debugger miss breakpoints. Please pass -Xfrozen_modules=off\n0.00s - to python to disable frozen modules.\n0.00s - Note: Debugging will proceed. Set PYDEVD_DISABLE_FILE_VALIDATION=1 to disable this validation.\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/nbconvert/utils/pandoc.py:51: RuntimeWarning: You are using an unsupported version of pandoc (2.9.2.1).\nYour version must be at least (2.14.2) but less than (4.0.0).\nRefer to https://pandoc.org/installing.html.\nContinuing with doubts...\n check_pandoc_version()\nmake[1]: Leaving directory '/home/runner/work/seaborn/seaborn/doc/_docstrings'\nmake: *** [Makefile:60: docstrings] Error 2\n##[error]Process completed with exit code 2.\n"}, {"step_name": "lint/5_Flake8.txt", "log": "##[group]Run make lint\n\u001b[36;1mmake lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n##[endgroup]\nflake8 seaborn/ tests/\nseaborn/_core/plot.py:1815:22: F821 undefined name 'get_layout_engine'\nmake: *** [Makefile:7: lint] Error 1\n##[error]Process completed with exit code 2.\n"}, {"step_name": "run-tests (3.11, light, latest)/5_Run tests.txt", "log": "##[group]Run make test\n\u001b[36;1mmake test\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n NB_KERNEL: python\n MPLBACKEND: Agg\n SEABORN_DATA: /home/runner/work/seaborn/seaborn/seaborn-data\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\npytest -n auto --cov=seaborn --cov=tests --cov-config=setup.cfg tests\n============================= test session starts ==============================\nplatform linux -- Python 3.11.7, pytest-7.4.3, pluggy-1.3.0\nrootdir: /home/runner/work/seaborn/seaborn\nconfigfile: pyproject.toml\nplugins: cov-4.1.0, xdist-3.5.0\ncreated: 4/4 workers\n4 workers [2379 items]\n\n........................................................................ [ 3%]\n........................................................................ [ 6%]\n........................................................................ [ 9%]\n........................................................................ [ 12%]\n........................................................................ [ 15%]\n........................................................................ [ 18%]\n........................................................................ [ 21%]\n........................................................................ [ 24%]\n........................................................................ [ 27%]\n........................................................................ [ 30%]\n........................................................................ [ 33%]\n.s...................................................................... [ 36%]\n........................................................................ [ 39%]\n........................................................................ [ 42%]\n.................................................s.s................ss.. [ 45%]\n..ss.............ssss................................................sss [ 48%]\nsssssssssssssssss.sssssssssssssssssss.ssssssssssssss.................... [ 51%]\n...............................................s........................ [ 54%]\n........................................................................ [ 57%]\n........................................................................ [ 60%]\n..............................................................s......... [ 63%]\n..............................................................x......... [ 66%]\n......x.........x.........................................x..x.......... [ 69%]\n.............................................s.....s.................... [ 72%]\n..................s...................................................... [ 75%]\n........................................................................ [ 78%]\n........................................................................ [ 81%]\n........................................................................ [ 84%]\n..................F.F..x................................................ [ 87%]\n...........................F............................................ [ 90%]\n........................................................................ [ 93%]\n........................................................................ [ 96%]\n........................s..s............................................ [ 99%]\n.. [100%]\n=================================== FAILURES ===================================\n_______________________ TestPlotting.test_layout_extent ________________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f06656abb10>\n\n @pytest.mark.skipif(\n _version_predates(mpl, \"3.6\"),\n reason=\"mpl<3.6 does not have get_layout_engine\",\n )\n def test_layout_extent(self):\n \n> p = Plot().layout(extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1100: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0654edebd0>\np = <seaborn._core.plot.Plot object at 0x7f06567fe150>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n_________________ TestPlotting.test_constrained_layout_extent __________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f0665702410>\n\n @pytest.mark.skipif(\n _version_predates(mpl, \"3.6\"),\n reason=\"mpl<3.6 does not have get_layout_engine\",\n )\n def test_constrained_layout_extent(self):\n \n> p = Plot().layout(engine=\"constrained\", extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1109: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0656f58050>\np = <seaborn._core.plot.Plot object at 0x7f0656f597d0>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n_____________________ TestPlotting.test_base_layout_extent _____________________\n[gw1] linux -- Python 3.11.7 /opt/hostedtoolcache/Python/3.11.7/x64/bin/python\n\nself = <tests._core.test_plot.TestPlotting object at 0x7f0664b34450>\n\n def test_base_layout_extent(self):\n \n> p = Plot().layout(engine=None, extent=(.1, .2, .6, 1)).plot()\n\ntests/_core/test_plot.py:1114: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \nseaborn/_core/plot.py:937: in plot\n return self._plot(pyplot)\nseaborn/_core/plot.py:971: in _plot\n plotter._finalize_figure(self)\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\nself = <seaborn._core.plot.Plotter object at 0x7f0656d300d0>\np = <seaborn._core.plot.Plot object at 0x7f0656d301d0>\n\n def _finalize_figure(self, p: Plot) -> None:\n \n for sub in self._subplots:\n ax = sub[\"ax\"]\n for axis in \"xy\":\n axis_key = sub[axis]\n axis_obj = getattr(ax, f\"{axis}axis\")\n \n # Axis limits\n if axis_key in p._limits or axis in p._limits:\n convert_units = getattr(ax, f\"{axis}axis\").convert_units\n a, b = p._limits.get(axis_key) or p._limits[axis]\n lo = a if a is None else convert_units(a)\n hi = b if b is None else convert_units(b)\n if isinstance(a, str):\n lo = cast(float, lo) - 0.5\n if isinstance(b, str):\n hi = cast(float, hi) + 0.5\n ax.set(**{f\"{axis}lim\": (lo, hi)})\n \n if axis_key in self._scales: # TODO when would it not be?\n self._scales[axis_key]._finalize(p, axis_obj)\n \n if (engine_name := p._layout_spec.get(\"engine\", default)) is not default:\n # None is a valid arg for Figure.set_layout_engine, hence `default`\n set_layout_engine(self._figure, engine_name)\n elif p._target is None:\n # Don't modify the layout engine if the user supplied their own\n # matplotlib figure and didn't specify an engine through Plot\n # TODO switch default to \"constrained\"?\n # TODO either way, make configurable\n set_layout_engine(self._figure, \"tight\")\n \n if (extent := p._layout_spec.get(\"extent\")) is not None:\n> engine = get_layout_engine(self._figure)\nE NameError: name 'get_layout_engine' is not defined\n\nseaborn/_core/plot.py:1815: NameError\n\n---------- coverage: platform linux, python 3.11.7-final-0 -----------\nName Stmts Miss Cover\n------------------------------------------------------\nseaborn/__init__.py 16 0 100%\nseaborn/_base.py 766 23 97%\nseaborn/_compat.py 77 47 39%\nseaborn/_core/__init__.py 0 0 100%\nseaborn/_core/data.py 116 5 96%\nseaborn/_core/exceptions.py 10 1 90%\nseaborn/_core/groupby.py 54 0 100%\nseaborn/_core/moves.py 109 0 100%\nseaborn/_core/plot.py 828 13 98%\nseaborn/_core/properties.py 425 4 99%\nseaborn/_core/rules.py 63 1 98%\nseaborn/_core/scales.py 502 49 90%\nseaborn/_core/subplots.py 140 0 100%\nseaborn/_core/typing.py 25 1 96%\nseaborn/_docstrings.py 40 5 88%\nseaborn/_marks/__init__.py 0 0 100%\nseaborn/_marks/area.py 86 3 97%\nseaborn/_marks/bar.py 123 0 100%\nseaborn/_marks/base.py 132 2 98%\nseaborn/_marks/dot.py 92 0 100%\nseaborn/_marks/line.py 116 0 100%\nseaborn/_marks/text.py 33 0 100%\nseaborn/_statistics.py 323 10 97%\nseaborn/_stats/__init__.py 0 0 100%\nseaborn/_stats/aggregation.py 42 1 98%\nseaborn/_stats/base.py 26 1 96%\nseaborn/_stats/counting.py 96 0 100%\nseaborn/_stats/density.py 99 3 97%\nseaborn/_stats/order.py 30 3 90%\nseaborn/_stats/regression.py 23 1 96%\nseaborn/_testing.py 52 4 92%\nseaborn/algorithms.py 60 0 100%\nseaborn/axisgrid.py 1039 32 97%\nseaborn/categorical.py 1204 13 99%\nseaborn/distributions.py 951 36 96%\nseaborn/matrix.py 551 317 42%\nseaborn/miscplot.py 24 0 100%\nseaborn/objects.py 15 0 100%\nseaborn/palettes.py 249 1 99%\nseaborn/rcmod.py 104 0 100%\nseaborn/regression.py 336 46 86%\nseaborn/relational.py 285 1 99%\nseaborn/utils.py 395 23 94%\ntests/__init__.py 0 0 100%\ntests/_core/__init__.py 0 0 100%\ntests/_core/test_data.py 293 2 99%\ntests/_core/test_groupby.py 83 0 100%\ntests/_core/test_moves.py 231 0 100%\ntests/_core/test_plot.py 1550 24 98%\ntests/_core/test_properties.py 365 4 99%\ntests/_core/test_rules.py 72 0 100%\ntests/_core/test_scales.py 547 1 99%\ntests/_core/test_subplots.py 368 0 100%\ntests/_marks/__init__.py 0 0 100%\ntests/_marks/test_area.py 84 0 100%\ntests/_marks/test_bar.py 152 0 100%\ntests/_marks/test_base.py 102 0 100%\ntests/_marks/test_dot.py 136 0 100%\ntests/_marks/test_line.py 298 0 100%\ntests/_marks/test_text.py 98 1 99%\ntests/_stats/__init__.py 0 0 100%\ntests/_stats/test_aggregation.py 84 0 100%\ntests/_stats/test_counting.py 180 0 100%\ntests/_stats/test_density.py 141 8 94%\ntests/_stats/test_order.py 64 1 98%\ntests/_stats/test_regression.py 36 0 100%\ntests/conftest.py 107 0 100%\ntests/test_algorithms.py 110 0 100%\ntests/test_axisgrid.py 1314 6 99%\ntests/test_base.py 1008 3 99%\ntests/test_categorical.py 2161 17 99%\ntests/test_distributions.py 1523 14 99%\ntests/test_docstrings.py 19 0 100%\ntests/test_matrix.py 864 454 47%\ntests/test_miscplot.py 24 0 100%\ntests/test_objects.py 11 0 100%\ntests/test_palettes.py 304 0 100%\ntests/test_rcmod.py 189 21 89%\ntests/test_regression.py 441 46 90%\ntests/test_relational.py 1213 8 99%\ntests/test_statistics.py 489 19 96%\ntests/test_utils.py 382 4 99%\n------------------------------------------------------\nTOTAL 24700 1279 95%\n\n=========================== short test summary info ============================\nFAILED tests/_core/test_plot.py::TestPlotting::test_layout_extent - NameError: name 'get_layout_engine' is not defined\nFAILED tests/_core/test_plot.py::TestPlotting::test_constrained_layout_extent - NameError: name 'get_layout_engine' is not defined\nFAILED tests/_core/test_plot.py::TestPlotting::test_base_layout_extent - NameError: name 'get_layout_engine' is not defined\n====== 3 failed, 2299 passed, 71 skipped, 6 xfailed in 244.97s (0:04:04) =======\nmake: *** [Makefile:4: test] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/seaborn/_core/plot.py b/seaborn/_core/plot.py\nindex af4669c6..39ccd2e0 100644\n--- a/seaborn/_core/plot.py\n+++ b/seaborn/_core/plot.py\n@@ -40,7 +40,7 @@ from seaborn._core.typing import (\n )\n from seaborn._core.exceptions import PlotSpecError\n from seaborn._core.rules import categorical_order\n-from seaborn._compat import set_layout_engine\n+from seaborn._compat import get_layout_engine, set_layout_engine\n from seaborn.rcmod import axes_style, plotting_context\n from seaborn.palettes import color_palette\n \n", "difficulty": 2, "changed_files": ["seaborn/_core/plot.py"], "commit_link": "https://github.com/mwaskom/seaborn/tree/2201be21886bb82201f3c3487f5f1468f6e6ac81"}
|
|
|
|
data/python/2327ca6.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/2a104bf.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 0, "repo_owner": "pyg-team", "repo_name": "pytorch_geometric", "head_branch": "master", "workflow_name": "Testing PyTorch nightly", "workflow_filename": "latest_testing.yml", "workflow_path": ".github/workflows/latest_testing.yml", "contributor": "pmpalang", "sha_fail": "2a104bfbc6cad12e95d941244955382e96f21c51", "sha_success": "060fc6a2edb9775331645512c1e6eb5ed4f0afbd", "workflow": "name: Testing PyTorch nightly\n\non: # yamllint disable-line rule:truthy\n push:\n branches:\n - master\n pull_request:\n\njobs:\n\n latest_pytest:\n runs-on: ubuntu-latest\n\n steps:\n - name: Checkout repository\n uses: actions/checkout@v4\n with:\n fetch-depth: 40\n\n # Skip workflow if only certain files have been changed.\n - name: Get changed files\n id: changed-files-specific\n uses: tj-actions/changed-files@v34\n with:\n files: |\n benchmark/**\n conda/**\n docker/**\n docs/**\n examples/**\n graphgym/**\n README.md\n CHANGELOG.md\n\n - name: Setup packages\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n uses: ./.github/actions/setup\n with:\n torch-version: nightly\n\n - name: Install main package\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n run: |\n pip install -e .[test]\n\n - name: Run tests\n if: steps.changed-files-specific.outputs.only_changed != 'true'\n timeout-minutes: 10\n run: |\n pytest\n", "logs": [{"step_name": "latest_pytest/6_Run tests.txt", "log": "##[group]Run pytest\n\u001b[36;1mpytest\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n\u001b[1m============================= test session starts ==============================\u001b[0m\nplatform linux -- Python 3.8.18, pytest-7.4.3, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.8.18/x64/bin/python\ncachedir: .pytest_cache\nrootdir: /home/runner/work/pytorch_geometric/pytorch_geometric\nconfigfile: pyproject.toml\nplugins: cov-4.1.0\n\u001b[1mcollecting ... \u001b[0mcollected 5718 items / 1 error\n\n==================================== ERRORS ====================================\n\u001b[31m\u001b[1m_____________ ERROR collecting test/loader/test_neighbor_loader.py _____________\u001b[0m\n\u001b[31mImportError while importing test module '/home/runner/work/pytorch_geometric/pytorch_geometric/test/loader/test_neighbor_loader.py'.\nHint: make sure your test modules/packages have valid Python names.\nTraceback:\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/python.py:617: in _importtestmodule\n mod = import_path(self.path, mode=importmode, root=self.config.rootpath)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/pathlib.py:567: in import_path\n importlib.import_module(module_name)\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/importlib/__init__.py:127: in import_module\n return _bootstrap._gcd_import(name[level:], package, level)\n<frozen importlib._bootstrap>:1014: in _gcd_import\n ???\n<frozen importlib._bootstrap>:991: in _find_and_load\n ???\n<frozen importlib._bootstrap>:975: in _find_and_load_unlocked\n ???\n<frozen importlib._bootstrap>:671: in _load_unlocked\n ???\n/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/_pytest/assertion/rewrite.py:186: in exec_module\n exec(co, module.__dict__)\ntest/loader/test_neighbor_loader.py:24: in <module>\n from torch_geometric.typing import (\nE ImportError: cannot import name 'WITH_EDGE_TIME_NEIGHBOR_SAMPLE' from 'torch_geometric.typing' (/home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/typing.py)\u001b[0m\n\u001b[33m=============================== warnings summary ===============================\u001b[0m\ntorch_geometric/graphgym/config.py:19\n /home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/graphgym/config.py:19: UserWarning: Could not define global config object. Please install 'yacs' via 'pip install yacs' in order to use GraphGym\n warnings.warn(\"Could not define global config object. Please install \"\n\ntorch_geometric/graphgym/imports.py:14\n /home/runner/work/pytorch_geometric/pytorch_geometric/torch_geometric/graphgym/imports.py:14: UserWarning: Please install 'pytorch_lightning' via 'pip install pytorch_lightning' in order to use GraphGym\n warnings.warn(\"Please install 'pytorch_lightning' via \"\n\n-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html\n\u001b[36m\u001b[1m=========================== short test summary info ============================\u001b[0m\n\u001b[31mERROR\u001b[0m test/loader/test_neighbor_loader.py\n!!!!!!!!!!!!!!!!!!!! Interrupted: 1 error during collection !!!!!!!!!!!!!!!!!!!!\n\u001b[31m========================= \u001b[33m2 warnings\u001b[0m, \u001b[31m\u001b[1m1 error\u001b[0m\u001b[31m in 7.62s\u001b[0m\u001b[31m =========================\u001b[0m\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/torch_geometric/typing.py b/torch_geometric/typing.py\nindex d9e687601..085ecc613 100644\n--- a/torch_geometric/typing.py\n+++ b/torch_geometric/typing.py\n@@ -57,6 +57,7 @@ except Exception as e:\n WITH_SAMPLED_OP = False\n WITH_INDEX_SORT = False\n WITH_METIS = False\n+ WITH_EDGE_TIME_NEIGHBOR_SAMPLE = False\n WITH_WEIGHTED_NEIGHBOR_SAMPLE = False\n \n try:\n", "difficulty": 1, "changed_files": ["torch_geometric/typing.py"], "commit_link": "https://github.com/pyg-team/pytorch_geometric/tree/2a104bfbc6cad12e95d941244955382e96f21c51"}
|
|
|
|
data/python/2a59b55.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 84, "repo_owner": "django-import-export", "repo_name": "django-import-export", "head_branch": "remove_count_queries", "workflow_name": ".github/workflows/pre-commit.yml", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "PetrDlouhy", "sha_fail": "2a59b55e6124b33dca7f48c12845c78130b20fd5", "sha_success": "36786764b3ef1c4c3c8d0db451cea9e1b325616c", "workflow": "on:\n pull_request:\n push:\n branches:\n - main\n\njobs:\n main:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: 3.x\n - uses: pre-commit/[email protected]\n - uses: pre-commit-ci/[email protected]\n if: always()\n", "logs": [{"step_name": "main/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.33-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.25.0-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.8-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 kB 6.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 26.4 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 31.5 MB/s eta 0:00:00\nDownloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 92.4 MB/s eta 0:00:00\nDownloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 78.2 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 93.4 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 platformdirs-4.1.0 pre-commit-3.6.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\nplatformdirs==4.1.0\npre-commit==3.6.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nCache Size: ~15 MB (15847689 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/83ce4811-68cf-4484-bf4e-4cfbe36bb31a/cache.tzst -P -C /home/runner/work/django-import-export/django-import-export --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.1/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: flake8\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\n\u001b[1mimport_export/admin.py\u001b[m\u001b[36m:\u001b[m749\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (96 > 88 characters)\n\u001b[1mimport_export/admin.py\u001b[m\u001b[36m:\u001b[m756\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (98 > 88 characters)\n\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/import_export/admin.py b/import_export/admin.py\nindex a1b288e9..8bb2db67 100644\n--- a/import_export/admin.py\n+++ b/import_export/admin.py\n@@ -746,14 +746,16 @@ class ExportMixin(BaseExportMixin, ImportExportMixinBase):\n def get_results(self, request):\n \"\"\"\n We override this method because we only call ChangeList.get_queryset()\n- so we don't need anything from this method. The get_results() gets called during\n- ChangeList.__init__() and we do want to avoid unnecessary COUNT queries.\n+ so we don't need anything from this method.\n+ The get_results() gets called during ChangeList.__init__()\n+ and we do want to avoid unnecessary COUNT queries.\n \"\"\"\n pass\n \n cl = ExportChangeList(**changelist_kwargs)\n \n- # get_queryset() is already called during initialization, it is enough to get it's results\n+ # get_queryset() is already called during initialization,\n+ # it is enough to get it's results\n if hasattr(cl, \"queryset\"):\n return cl.queryset\n \n", "difficulty": 0, "changed_files": ["import_export/admin.py"], "commit_link": "https://github.com/django-import-export/django-import-export/tree/2a59b55e6124b33dca7f48c12845c78130b20fd5"}
|
|
|
|
data/python/2ab9e84.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 171, "repo_owner": "qtile", "repo_name": "qtile", "head_branch": "lists-match", "workflow_name": "Run pre-commit", "workflow_filename": "pre_commit.yml", "workflow_path": ".github/workflows/pre_commit.yml", "contributor": "jwijenbergh", "sha_fail": "2ab9e843db39063cc9bc33b924cef535eb289894", "sha_success": "26c042804d308f4c1c5b5061ee10da1515129489", "workflow": "name: Run pre-commit\n\non:\n push:\n pull_request:\n\njobs:\n check:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v4\n - name: Install dependencies\n run: |\n sudo apt update\n sudo apt install --no-install-recommends libxkbcommon-dev\n - uses: pre-commit/[email protected]\n", "logs": [{"step_name": "check/5_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nDefaulting to user installation because normal site-packages is not writeable\nCollecting pre-commit\n Downloading pre_commit-3.6.0-py2.py3-none-any.whl (204 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 204.0/204.0 KB 6.3 MB/s eta 0:00:00\nCollecting nodeenv>=0.11.1\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nRequirement already satisfied: pyyaml>=5.1 in /usr/lib/python3/dist-packages (from pre-commit) (5.4.1)\nCollecting cfgv>=2.0.0\n Downloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nCollecting virtualenv>=20.10.0\n Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 24.4 MB/s eta 0:00:00\nCollecting identify>=1.0.0\n Downloading identify-2.5.33-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 KB 37.4 MB/s eta 0:00:00\nRequirement already satisfied: setuptools in /usr/lib/python3/dist-packages (from nodeenv>=0.11.1->pre-commit) (59.6.0)\nCollecting filelock<4,>=3.12.2\n Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\nRequirement already satisfied: platformdirs<5,>=3.9.1 in /usr/local/lib/python3.10/dist-packages (from virtualenv>=20.10.0->pre-commit) (4.1.0)\nCollecting distlib<1,>=0.3.7\n Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 43.2 MB/s eta 0:00:00\nInstalling collected packages: distlib, nodeenv, identify, filelock, cfgv, virtualenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.8 filelock-3.13.1 identify-2.5.33 nodeenv-1.8.0 pre-commit-3.6.0 virtualenv-20.25.0\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nargcomplete==3.2.1\ncfgv==3.4.0\ndistlib==0.3.8\nfilelock==3.13.1\nidentify==2.5.33\nnodeenv==1.8.0\npackaging==23.2\npipx==1.3.3\nplatformdirs==4.1.0\npre-commit==3.6.0\ntomli==2.0.1\nuserpath==1.9.1\nvirtualenv==20.25.0\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3||05c1fbbb63b353467651b511e6ac241ffd2d8e71749cfabfe5ee6bc8366d2d02\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\n##[endgroup]\nCache Size: ~36 MB (37431926 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/8e85cf73-4d99-4c40-a6e8-25c81b58e73c/cache.tzst -P -C /home/runner/work/qtile/qtile --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3||05c1fbbb63b353467651b511e6ac241ffd2d8e71749cfabfe5ee6bc8366d2d02\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[42mPassed\u001b[m\nmypy.....................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: mypy\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\nlibqtile/config.py:886: error: Statement is unreachable [unreachable]\nlibqtile/config.py:890: error: Statement is unreachable [unreachable]\nlibqtile/config.py:894: error: Statement is unreachable [unreachable]\nlibqtile/config.py:911: error: Statement is unreachable [unreachable]\nlibqtile/config.py:915: error: Statement is unreachable [unreachable]\nFound 5 errors in 1 file (checked 179 source files)\n\nvulture..................................................................\u001b[42mPassed\u001b[m\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/libqtile/config.py b/libqtile/config.py\nindex 9801b3ba..6a6163e2 100644\n--- a/libqtile/config.py\n+++ b/libqtile/config.py\n@@ -883,15 +883,15 @@ class Match:\n \n if title is not None:\n if isinstance(title, list): # type: ignore\n- title = convert_deprecated_list(title, \"title\")\n+ title = convert_deprecated_list(title, \"title\") # type: ignore\n self._rules[\"title\"] = title\n if wm_class is not None:\n if isinstance(wm_class, list): # type: ignore\n- wm_class = convert_deprecated_list(wm_class, \"wm_class\")\n+ wm_class = convert_deprecated_list(wm_class, \"wm_class\") # type: ignore\n self._rules[\"wm_class\"] = wm_class\n if wm_instance_class is not None:\n if isinstance(wm_instance_class, list): # type: ignore\n- wm_instance_class = convert_deprecated_list(\n+ wm_instance_class = convert_deprecated_list( # type: ignore\n wm_instance_class, \"wm_instance_class\"\n )\n self._rules[\"wm_instance_class\"] = wm_instance_class\n@@ -908,11 +908,11 @@ class Match:\n \n if role is not None:\n if isinstance(role, list): # type: ignore\n- role = convert_deprecated_list(role, \"role\")\n+ role = convert_deprecated_list(role, \"role\") # type: ignore\n self._rules[\"role\"] = role\n if wm_type is not None:\n if isinstance(wm_type, list): # type: ignore\n- wm_type = convert_deprecated_list(wm_type, \"wm_type\")\n+ wm_type = convert_deprecated_list(wm_type, \"wm_type\") # type: ignore\n self._rules[\"wm_type\"] = wm_type\n \n @staticmethod\n", "difficulty": 1, "changed_files": ["libqtile/config.py"], "commit_link": "https://github.com/qtile/qtile/tree/2ab9e843db39063cc9bc33b924cef535eb289894"}
|
|
|
|
data/python/2c06ffa.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 4, "repo_owner": "huggingface", "repo_name": "diffusers", "head_branch": "ipadapterfaceid", "workflow_name": "Run code quality checks", "workflow_filename": "pr_quality.yml", "workflow_path": ".github/workflows/pr_quality.yml", "contributor": "fabiorigano", "sha_fail": "2c06ffa4c9d2c37846c60ad75899b4d72f214ff9", "sha_success": "217d9d073981605acab5200fc841f20c798c1449", "workflow": "name: Run code quality checks\n\non:\n pull_request:\n branches:\n - main\n push:\n branches:\n - main\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.head_ref || github.run_id }}\n cancel-in-progress: true\n\njobs:\n check_code_quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python\n uses: actions/setup-python@v4\n with:\n python-version: \"3.8\"\n - name: Install dependencies\n run: |\n python -m pip install --upgrade pip\n pip install .[quality]\n - name: Check quality\n run: |\n ruff check examples tests src utils scripts\n ruff format examples tests src utils scripts --check\n\n check_repository_consistency:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python\n uses: actions/setup-python@v4\n with:\n python-version: \"3.8\"\n - name: Install dependencies\n run: |\n python -m pip install --upgrade pip\n pip install .[quality]\n - name: Check quality\n run: |\n python utils/check_copies.py\n python utils/check_dummies.py\n make deps_table_check_updated\n", "logs": [{"step_name": "check_code_quality/5_Check quality.txt", "log": "##[group]Run ruff check examples tests src utils scripts\n\u001b[36;1mruff check examples tests src utils scripts\u001b[0m\n\u001b[36;1mruff format examples tests src utils scripts --check\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/actions-runner/_work/_tool/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/actions-runner/_work/_tool/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/actions-runner/_work/_tool/Python/3.8.18/x64/lib\n##[endgroup]\nexamples/community/ip_adapter_face_id.py:15:1: I001 [*] Import block is un-sorted or un-formatted\nFound 1 error.\n[*] 1 fixable with the `--fix` option.\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/examples/community/README.md b/examples/community/README.md\nindex f205f3b70..2fdbdb414 100755\n--- a/examples/community/README.md\n+++ b/examples/community/README.md\n@@ -3307,7 +3307,7 @@ pipeline = DiffusionPipeline.from_pretrained(\n torch_dtype=torch.float16,\n scheduler=noise_scheduler,\n vae=vae,\n- custom_pipeline=\"./forked/diffusers/examples/community/ip_adapter_face_id.py\"\n+ custom_pipeline=\"ip_adapter_face_id\"\n )\n pipeline.load_ip_adapter_face_id(\"h94/IP-Adapter-FaceID\", \"ip-adapter-faceid_sd15.bin\")\n pipeline.to(\"cuda\")\ndiff --git a/examples/community/ip_adapter_face_id.py b/examples/community/ip_adapter_face_id.py\nindex e3c5a2c84..d9325742c 100644\n--- a/examples/community/ip_adapter_face_id.py\n+++ b/examples/community/ip_adapter_face_id.py\n@@ -14,12 +14,12 @@\n \n import inspect\n from typing import Any, Callable, Dict, List, Optional, Union\n-from safetensors import safe_open\n \n import torch\n import torch.nn as nn\n import torch.nn.functional as F\n from packaging import version\n+from safetensors import safe_open\n from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, CLIPVisionModelWithProjection\n \n from diffusers.configuration_utils import FrozenDict\n@@ -27,20 +27,20 @@ from diffusers.image_processor import VaeImageProcessor\n from diffusers.loaders import FromSingleFileMixin, IPAdapterMixin, LoraLoaderMixin, TextualInversionLoaderMixin\n from diffusers.models import AutoencoderKL, UNet2DConditionModel\n from diffusers.models.attention_processor import FusedAttnProcessor2_0\n-from diffusers.models.lora import adjust_lora_scale_text_encoder, LoRALinearLayer\n+from diffusers.models.lora import LoRALinearLayer, adjust_lora_scale_text_encoder\n+from diffusers.pipelines.pipeline_utils import DiffusionPipeline\n+from diffusers.pipelines.stable_diffusion.pipeline_output import StableDiffusionPipelineOutput\n+from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker\n from diffusers.schedulers import KarrasDiffusionSchedulers\n from diffusers.utils import (\n- _get_model_file,\n USE_PEFT_BACKEND,\n+ _get_model_file,\n deprecate,\n logging,\n scale_lora_layers,\n unscale_lora_layers,\n )\n from diffusers.utils.torch_utils import randn_tensor\n-from diffusers.pipelines.pipeline_utils import DiffusionPipeline\n-from diffusers.pipelines.stable_diffusion.pipeline_output import StableDiffusionPipelineOutput\n-from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker\n \n \n logger = logging.get_logger(__name__) # pylint: disable=invalid-name\n@@ -555,7 +555,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(\n revision=revision,\n subfolder=subfolder,\n user_agent=user_agent,\n- )\n+ )\n if weight_name.endswith(\".safetensors\"):\n state_dict = {\"image_proj\": {}, \"ip_adapter\": {}}\n with safe_open(model_file, framework=\"pt\", device=\"cpu\") as f:\n@@ -1438,7 +1438,7 @@ class IPAdapterFaceIDStableDiffusionPipeline(\n extra_step_kwargs = self.prepare_extra_step_kwargs(generator, eta)\n \n # 6.1 Add image embeds for IP-Adapter\n- added_cond_kwargs ={\"image_embeds\": image_embeds} if image_embeds is not None else None\n+ added_cond_kwargs = {\"image_embeds\": image_embeds} if image_embeds is not None else None\n \n # 6.2 Optionally get Guidance Scale Embedding\n timestep_cond = None\n", "difficulty": 0, "changed_files": ["examples/community/README.md", "examples/community/ip_adapter_face_id.py"], "commit_link": "https://github.com/huggingface/diffusers/tree/2c06ffa4c9d2c37846c60ad75899b4d72f214ff9"}
|
|
|
|
data/python/2ccb7d3.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 114, "repo_owner": "mikf", "repo_name": "gallery-dl", "head_branch": "steamgriddb", "workflow_name": "tests", "workflow_filename": "tests.yml", "workflow_path": ".github/workflows/tests.yml", "contributor": "the-blank-x", "sha_fail": "2ccb7d3bd3f071c6923ca6eb9baedd196665d769", "sha_success": "100966b122cd90ca139593cf8ff21fd0f777243a", "workflow": "name: tests\n\non:\n push:\n branches:\n - master\n pull_request:\n branches:\n - master\n\njobs:\n build:\n\n runs-on: ubuntu-20.04\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.5\", \"3.6\", \"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\", \"pypy3.9\"]\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Check file permissions\n run: |\n if [[ \"$(find ./gallery_dl -type f -not -perm 644)\" ]]; then exit 1; fi\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Install dependencies\n run: |\n pip install -r requirements.txt\n pip install flake8 youtube-dl\n\n - name: Install yt-dlp\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5)\n # don't install yt-dlp\n ;;\n 3.6|3.7)\n # install from PyPI\n pip install yt-dlp\n ;;\n *)\n # install from master\n pip install https://github.com/yt-dlp/yt-dlp/archive/refs/heads/master.tar.gz\n ;;\n esac\n\n - name: Lint with flake8\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5|3.6|3.7)\n flake8 --extend-exclude scripts/export_tests.py .\n ;;\n *)\n flake8 .\n ;;\n esac\n\n - name: Run tests\n run: |\n make test\n\n - name: Test autogeneration of man pages, bash/zsh/fish completion, etc\n run: |\n make\n", "logs": [{"step_name": "build (3.5)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.5\" in\n\u001b[36;1mcase \"3.5\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.5.10/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.6)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.6\" in\n\u001b[36;1mcase \"3.6\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.6.15/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.7)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.7\" in\n\u001b[36;1mcase \"3.7\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.8)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.8\" in\n\u001b[36;1mcase \"3.8\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.9\" in\n\u001b[36;1mcase \"3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.10)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.10\" in\n\u001b[36;1mcase \"3.10\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.11)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.11\" in\n\u001b[36;1mcase \"3.11\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.12)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.12\" in\n\u001b[36;1mcase \"3.12\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (pypy3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"pypy3.9\" in\n\u001b[36;1mcase \"pypy3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/PyPy/3.9.18/x64/bin/lib/pkgconfig\n##[endgroup]\n./gallery_dl/extractor/steamgriddb.py:176:80: E501 line too long (80 > 79 characters)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gallery_dl/extractor/steamgriddb.py b/gallery_dl/extractor/steamgriddb.py\nindex 516c422b..1f803ffd 100644\n--- a/gallery_dl/extractor/steamgriddb.py\n+++ b/gallery_dl/extractor/steamgriddb.py\n@@ -173,7 +173,8 @@ class SteamgriddbGridsExtractor(SteamgriddbAssetsExtractor):\n example = \"https://www.steamgriddb.com/game/1234/grids\"\n valid_dimensions = (\"460x215\", \"920x430\", \"600x900\", \"342x482\", \"660x930\",\n \"512x512\", \"1024x1024\")\n- valid_styles = (\"alternate\", \"blurred\", \"no_logo\", \"material\", \"white_logo\")\n+ valid_styles = (\"alternate\", \"blurred\", \"no_logo\", \"material\",\n+ \"white_logo\")\n valid_file_types = (\"png\", \"jpeg\", \"jpg\", \"webp\")\n \n \n", "difficulty": 0, "changed_files": ["gallery_dl/extractor/steamgriddb.py"], "commit_link": "https://github.com/mikf/gallery-dl/tree/2ccb7d3bd3f071c6923ca6eb9baedd196665d769"}
|
|
|
|
data/python/2e41e78.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/2f0605c.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 85, "repo_owner": "django-import-export", "repo_name": "django-import-export", "head_branch": "fix-declared-model-fields-altering-export-order", "workflow_name": ".github/workflows/pre-commit.yml", "workflow_filename": "pre-commit.yml", "workflow_path": ".github/workflows/pre-commit.yml", "contributor": "cocorocho", "sha_fail": "2f0605c9ec79b7a675728cb525ad55b36ade2e93", "sha_success": "d1e35b26f1cfc1bd08b03ac606fa8bcd26dc91cb", "workflow": "on:\n pull_request:\n push:\n branches:\n - main\n\njobs:\n main:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: 3.x\n - uses: pre-commit/[email protected]\n - uses: pre-commit-ci/[email protected]\n if: always()\n", "logs": [{"step_name": "main/4_Run [email protected]", "log": "##[group]Run pre-commit/[email protected]\nwith:\n extra_args: --all-files\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\n##[group]Run python -m pip install pre-commit\n\u001b[36;1mpython -m pip install pre-commit\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nCollecting pre-commit\n Downloading pre_commit-3.5.0-py2.py3-none-any.whl.metadata (1.3 kB)\nCollecting cfgv>=2.0.0 (from pre-commit)\n Downloading cfgv-3.4.0-py2.py3-none-any.whl.metadata (8.5 kB)\nCollecting identify>=1.0.0 (from pre-commit)\n Downloading identify-2.5.32-py2.py3-none-any.whl.metadata (4.4 kB)\nCollecting nodeenv>=0.11.1 (from pre-commit)\n Downloading nodeenv-1.8.0-py2.py3-none-any.whl.metadata (21 kB)\nCollecting pyyaml>=5.1 (from pre-commit)\n Downloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (2.1 kB)\nCollecting virtualenv>=20.10.0 (from pre-commit)\n Downloading virtualenv-20.24.7-py3-none-any.whl.metadata (4.5 kB)\nCollecting setuptools (from nodeenv>=0.11.1->pre-commit)\n Downloading setuptools-69.0.2-py3-none-any.whl.metadata (6.3 kB)\nCollecting distlib<1,>=0.3.7 (from virtualenv>=20.10.0->pre-commit)\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting filelock<4,>=3.12.2 (from virtualenv>=20.10.0->pre-commit)\n Downloading filelock-3.13.1-py3-none-any.whl.metadata (2.8 kB)\nCollecting platformdirs<5,>=3.9.1 (from virtualenv>=20.10.0->pre-commit)\n Downloading platformdirs-4.0.0-py3-none-any.whl.metadata (11 kB)\nDownloading pre_commit-3.5.0-py2.py3-none-any.whl (203 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 203.7/203.7 kB 15.8 MB/s eta 0:00:00\nDownloading cfgv-3.4.0-py2.py3-none-any.whl (7.2 kB)\nDownloading identify-2.5.32-py2.py3-none-any.whl (98 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 98.9/98.9 kB 32.6 MB/s eta 0:00:00\nDownloading nodeenv-1.8.0-py2.py3-none-any.whl (22 kB)\nDownloading PyYAML-6.0.1-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (724 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 725.0/725.0 kB 49.7 MB/s eta 0:00:00\nDownloading virtualenv-20.24.7-py3-none-any.whl (3.8 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 124.4 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 84.7 MB/s eta 0:00:00\nDownloading filelock-3.13.1-py3-none-any.whl (11 kB)\nDownloading platformdirs-4.0.0-py3-none-any.whl (17 kB)\nDownloading setuptools-69.0.2-py3-none-any.whl (819 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 819.5/819.5 kB 99.2 MB/s eta 0:00:00\nInstalling collected packages: distlib, setuptools, pyyaml, platformdirs, identify, filelock, cfgv, virtualenv, nodeenv, pre-commit\nSuccessfully installed cfgv-3.4.0 distlib-0.3.7 filelock-3.13.1 identify-2.5.32 nodeenv-1.8.0 platformdirs-4.0.0 pre-commit-3.5.0 pyyaml-6.0.1 setuptools-69.0.2 virtualenv-20.24.7\n\n[notice] A new release of pip is available: 23.2.1 -> 23.3.1\n[notice] To update, run: pip install --upgrade pip\n##[group]Run python -m pip freeze --local\n\u001b[36;1mpython -m pip freeze --local\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\ncfgv==3.4.0\ndistlib==0.3.7\nfilelock==3.13.1\nidentify==2.5.32\nnodeenv==1.8.0\nplatformdirs==4.0.0\npre-commit==3.5.0\nPyYAML==6.0.1\nsetuptools==69.0.2\nvirtualenv==20.24.7\n##[group]Run actions/cache@v3\nwith:\n path: ~/.cache/pre-commit\n key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.0/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n enableCrossOsArchive: false\n fail-on-cache-miss: false\n lookup-only: false\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nCache Size: ~15 MB (15834028 B)\n[command]/usr/bin/tar -xf /home/runner/work/_temp/98029e33-8cfe-40ab-80a4-dbe66e8ec82f/cache.tzst -P -C /home/runner/work/django-import-export/django-import-export --use-compress-program unzstd\nCache restored successfully\nCache restored from key: pre-commit-3|/opt/hostedtoolcache/Python/3.12.0/x64|78083b3309c934e46e7d0a952b7a4c6dc6df8860edaa74474d06dd8c36e1b0a2\n##[group]Run pre-commit run --show-diff-on-failure --color=always --all-files\n\u001b[36;1mpre-commit run --show-diff-on-failure --color=always --all-files\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nblack....................................................................\u001b[42mPassed\u001b[m\nisort....................................................................\u001b[42mPassed\u001b[m\nflake8...................................................................\u001b[41mFailed\u001b[m\n\u001b[2m- hook id: flake8\u001b[m\n\u001b[2m- exit code: 1\u001b[m\n\n\u001b[1mimport_export/resources.py\u001b[m\u001b[36m:\u001b[m1360\u001b[36m:\u001b[m89\u001b[36m:\u001b[m \u001b[1m\u001b[31mE501\u001b[m line too long (101 > 88 characters)\n\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/import_export/resources.py b/import_export/resources.py\nindex 7d7ad475..99c3fe91 100644\n--- a/import_export/resources.py\n+++ b/import_export/resources.py\n@@ -1357,7 +1357,8 @@ class ModelDeclarativeMetaclass(DeclarativeMetaclass):\n continue\n \n if f.name in declared_fields:\n- # If model field is declared in `ModelResource`, remove it from `declared_fields`\n+ # If model field is declared in `ModelResource`,\n+ # remove it from `declared_fields`\n # to keep exact order of model fields\n field = declared_fields.pop(f.name)\n else:\n", "difficulty": 0, "changed_files": ["import_export/resources.py"], "commit_link": "https://github.com/django-import-export/django-import-export/tree/2f0605c9ec79b7a675728cb525ad55b36ade2e93"}
|
|
|
|
data/python/3054659.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/3137ef6.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 69, "repo_owner": "optuna", "repo_name": "optuna", "head_branch": "unifiy-implementation-of-fast-nondominated-sort", "workflow_name": "Checks", "workflow_filename": "checks.yml", "workflow_path": ".github/workflows/checks.yml", "contributor": "Alnusjaponica", "sha_fail": "3137ef65975fc93e9e82b130e545028223cef408", "sha_success": "a4cd1cd2d275d5163bcbbb4cb29b3cf378789e99", "workflow": "name: Checks\n\non:\n push:\n branches:\n - master\n pull_request: {}\n schedule:\n - cron: '0 23 * * SUN-THU'\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.ref == 'refs/heads/master' && github.run_number || github.ref }}\n cancel-in-progress: true\n\njobs:\n checks:\n if: (github.event_name == 'schedule' && github.repository == 'optuna/optuna') || (github.event_name != 'schedule')\n runs-on: ubuntu-latest\n\n steps:\n - uses: actions/checkout@v3\n\n - uses: actions/setup-python@v4\n with:\n python-version: 3.8\n\n - name: Install\n run: |\n python -m pip install -U pip\n pip install --progress-bar off -U .[checking]\n\n - name: Output installed packages\n run: |\n pip freeze --all\n - name: Output dependency tree\n run: |\n pip install pipdeptree\n pipdeptree\n\n - name: black\n run: black . --check --diff\n - name: flake8\n run: flake8 .\n - name: isort\n run: isort . --check --diff\n - name: mypy\n run: mypy .\n - name: blackdoc\n run: blackdoc . --check --diff\n", "logs": [{"step_name": "checks/9_isort.txt", "log": "##[group]Run isort . --check --diff\n\u001b[36;1misort . --check --diff\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nERROR: /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py Imports are incorrectly sorted and/or formatted.\n--- /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py:before\t2023-12-22 12:41:32.147471\n+++ /home/runner/work/optuna/optuna/optuna/study/_multi_objective.py:after\t2023-12-22 12:42:07.977714\n@@ -1,13 +1,16 @@\n from __future__ import annotations\n \n from collections import defaultdict\n-from typing import List, Optional, Sequence\n+from typing import List\n+from typing import Optional\n+from typing import Sequence\n \n import numpy as np\n \n import optuna\n from optuna.study._study_direction import StudyDirection\n-from optuna.trial import FrozenTrial, TrialState\n+from optuna.trial import FrozenTrial\n+from optuna.trial import TrialState\n \n \n def _get_pareto_front_trials_2d(\nSkipped 30 files\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/optuna/study/_multi_objective.py b/optuna/study/_multi_objective.py\nindex 4e4a36557..18cc8bccf 100644\n--- a/optuna/study/_multi_objective.py\n+++ b/optuna/study/_multi_objective.py\n@@ -1,13 +1,16 @@\n from __future__ import annotations\n \n from collections import defaultdict\n-from typing import List, Optional, Sequence\n+from typing import List\n+from typing import Optional\n+from typing import Sequence\n \n import numpy as np\n \n import optuna\n from optuna.study._study_direction import StudyDirection\n-from optuna.trial import FrozenTrial, TrialState\n+from optuna.trial import FrozenTrial\n+from optuna.trial import TrialState\n \n \n def _get_pareto_front_trials_2d(\n", "difficulty": 0, "changed_files": ["optuna/study/_multi_objective.py"], "commit_link": "https://github.com/optuna/optuna/tree/3137ef65975fc93e9e82b130e545028223cef408"}
|
|
|
|
data/python/31c3e1d.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 104, "repo_owner": "deepchecks", "repo_name": "deepchecks", "head_branch": "main", "workflow_name": "Code Style", "workflow_filename": "lint.yml", "workflow_path": ".github/workflows/lint.yml", "contributor": "manijhariya", "sha_fail": "31c3e1dfba76d6e90b5580da2f80fe4b12e6cb5c", "sha_success": "dabfee7048e94ef6a7c8a89159a7ddb887eeacbb", "workflow": "# ----------------------------------------------------------------------------\n# Copyright (C) 2021-2023 Deepchecks (https://www.deepchecks.com)\n#\n# This file is part of Deepchecks.\n# Deepchecks is distributed under the terms of the GNU Affero General\n# Public License (version 3 or later).\n# You should have received a copy of the GNU Affero General Public License\n# along with Deepchecks. If not, see <http://www.gnu.org/licenses/>.\n# ----------------------------------------------------------------------------\n#\nname: Code Style\n\non:\n push:\n branches: [ main ]\n pull_request: \n branches: [ main, nlp-base]\n\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: true\n\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: \"true\"\n\njobs:\n pylint:\n runs-on: ubuntu-latest\n\n\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n\n - name: Run Pylint\n run: make pylint\n - name: Check import order\n uses: jamescurtin/isort-action@master\n with:\n sortPaths: deepchecks/\n docstring:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n - name: Run Pydocstyle\n run: make docstring\n\n license-check:\n runs-on: ubuntu-latest\n \n needs: docstring\n \n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.9\n uses: actions/setup-python@v2\n with:\n python-version: 3.9\n - name: Run License eye\n run: make license-check\n", "logs": [{"step_name": "pylint/5_Check import order.txt", "log": "##[group]Run jamescurtin/isort-action@master\nwith:\n sortPaths: deepchecks/\n isortVersion: latest\n configuration: --check-only --diff\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/ensure_python\n\u001b[36;1m$GITHUB_ACTION_PATH/bin/ensure_python\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/install_packages latest \n\u001b[36;1m$GITHUB_ACTION_PATH/bin/install_packages latest \u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Install isort\nInstalling latest version of isort\nCollecting isort[pipfile_deprecated_finder,requirements_deprecated_finder]\n Downloading isort-5.12.0-py3-none-any.whl (91 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 91.2/91.2 kB 4.2 MB/s eta 0:00:00\nCollecting pip-api (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip_api-0.0.30-py3-none-any.whl (111 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 111.6/111.6 kB 8.3 MB/s eta 0:00:00\nCollecting pipreqs (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pipreqs-0.4.13-py2.py3-none-any.whl (33 kB)\nCollecting pip-shims>=0.5.2 (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip_shims-0.7.3-py2.py3-none-any.whl (35 kB)\nCollecting requirementslib (from isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading requirementslib-3.0.0-py2.py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: pip>=20 in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder]) (23.0.1)\nCollecting packaging (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading packaging-23.2-py3-none-any.whl.metadata (3.2 kB)\nRequirement already satisfied: setuptools in /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder]) (58.1.0)\nCollecting wheel (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading wheel-0.42.0-py3-none-any.whl.metadata (2.2 kB)\nCollecting docopt (from pipreqs->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading docopt-0.6.2.tar.gz (25 kB)\n Installing build dependencies: started\n Installing build dependencies: finished with status 'done'\n Getting requirements to build wheel: started\n Getting requirements to build wheel: finished with status 'done'\n Preparing metadata (pyproject.toml): started\n Preparing metadata (pyproject.toml): finished with status 'done'\nCollecting yarg (from pipreqs->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading yarg-0.1.9-py2.py3-none-any.whl (19 kB)\nCollecting distlib>=0.2.8 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading distlib-0.3.7-py2.py3-none-any.whl.metadata (5.1 kB)\nCollecting pep517>=0.5.0 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pep517-0.13.1-py3-none-any.whl.metadata (4.1 kB)\nCollecting pip>=20 (from pip-shims>=0.5.2->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pip-23.3.1-py3-none-any.whl.metadata (3.5 kB)\nCollecting platformdirs (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading platformdirs-4.1.0-py3-none-any.whl.metadata (11 kB)\nCollecting plette[validation] (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading plette-0.4.4-py2.py3-none-any.whl (12 kB)\nCollecting pydantic (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pydantic-2.5.2-py3-none-any.whl.metadata (65 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 65.2/65.2 kB 8.9 MB/s eta 0:00:00\nCollecting requests (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading requests-2.31.0-py3-none-any.whl.metadata (4.6 kB)\nCollecting tomlkit>=0.5.3 (from requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading tomlkit-0.12.3-py3-none-any.whl.metadata (2.7 kB)\nCollecting tomli>=1.1.0 (from pep517>=0.5.0->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading tomli-2.0.1-py3-none-any.whl (12 kB)\nCollecting cerberus (from plette[validation]->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading Cerberus-1.3.5-py3-none-any.whl.metadata (6.0 kB)\nCollecting annotated-types>=0.4.0 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading annotated_types-0.6.0-py3-none-any.whl.metadata (12 kB)\nCollecting pydantic-core==2.14.5 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading pydantic_core-2.14.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.5 kB)\nCollecting typing-extensions>=4.6.1 (from pydantic->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading typing_extensions-4.8.0-py3-none-any.whl.metadata (3.0 kB)\nCollecting charset-normalizer<4,>=2 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (33 kB)\nCollecting idna<4,>=2.5 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading idna-3.6-py3-none-any.whl.metadata (9.9 kB)\nCollecting urllib3<3,>=1.21.1 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading urllib3-2.1.0-py3-none-any.whl.metadata (6.4 kB)\nCollecting certifi>=2017.4.17 (from requests->requirementslib->isort[pipfile_deprecated_finder,requirements_deprecated_finder])\n Downloading certifi-2023.11.17-py3-none-any.whl.metadata (2.2 kB)\nDownloading requirementslib-3.0.0-py2.py3-none-any.whl (96 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 96.7/96.7 kB 8.0 MB/s eta 0:00:00\nDownloading distlib-0.3.7-py2.py3-none-any.whl (468 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 kB 10.1 MB/s eta 0:00:00\nDownloading pep517-0.13.1-py3-none-any.whl (19 kB)\nDownloading pip-23.3.1-py3-none-any.whl (2.1 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.1/2.1 MB 17.3 MB/s eta 0:00:00\nDownloading tomlkit-0.12.3-py3-none-any.whl (37 kB)\nDownloading packaging-23.2-py3-none-any.whl (53 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.0/53.0 kB 15.6 MB/s eta 0:00:00\nDownloading platformdirs-4.1.0-py3-none-any.whl (17 kB)\nDownloading pydantic-2.5.2-py3-none-any.whl (381 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 381.9/381.9 kB 21.0 MB/s eta 0:00:00\nDownloading pydantic_core-2.14.5-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (2.1 MB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.1/2.1 MB 29.9 MB/s eta 0:00:00\nDownloading requests-2.31.0-py3-none-any.whl (62 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 62.6/62.6 kB 18.9 MB/s eta 0:00:00\nUsing cached wheel-0.42.0-py3-none-any.whl (65 kB)\nDownloading annotated_types-0.6.0-py3-none-any.whl (12 kB)\nDownloading certifi-2023.11.17-py3-none-any.whl (162 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 162.5/162.5 kB 33.6 MB/s eta 0:00:00\nDownloading charset_normalizer-3.3.2-cp39-cp39-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (142 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 142.3/142.3 kB 36.6 MB/s eta 0:00:00\nDownloading idna-3.6-py3-none-any.whl (61 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 61.6/61.6 kB 19.5 MB/s eta 0:00:00\nDownloading typing_extensions-4.8.0-py3-none-any.whl (31 kB)\nDownloading urllib3-2.1.0-py3-none-any.whl (104 kB)\n \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.6/104.6 kB 29.6 MB/s eta 0:00:00\nDownloading Cerberus-1.3.5-py3-none-any.whl (30 kB)\nBuilding wheels for collected packages: docopt\n Building wheel for docopt (pyproject.toml): started\n Building wheel for docopt (pyproject.toml): finished with status 'done'\n Created wheel for docopt: filename=docopt-0.6.2-py2.py3-none-any.whl size=13706 sha256=29c03dca51f57ce29e3477c3d0f4958f94387b06a456f0fb38fa4f2f840b837c\n Stored in directory: /home/runner/.cache/pip/wheels/70/4a/46/1309fc853b8d395e60bafaf1b6df7845bdd82c95fd59dd8d2b\nSuccessfully built docopt\nInstalling collected packages: docopt, distlib, cerberus, wheel, urllib3, typing-extensions, tomlkit, tomli, platformdirs, pip, packaging, isort, idna, charset-normalizer, certifi, annotated-types, requests, pydantic-core, plette, pip-shims, pip-api, pep517, yarg, pydantic, requirementslib, pipreqs\n Attempting uninstall: pip\n Found existing installation: pip 23.0.1\n Uninstalling pip-23.0.1:\n Successfully uninstalled pip-23.0.1\nSuccessfully installed annotated-types-0.6.0 cerberus-1.3.5 certifi-2023.11.17 charset-normalizer-3.3.2 distlib-0.3.7 docopt-0.6.2 idna-3.6 isort-5.12.0 packaging-23.2 pep517-0.13.1 pip-23.3.1 pip-api-0.0.30 pip-shims-0.7.3 pipreqs-0.4.13 platformdirs-4.1.0 plette-0.4.4 pydantic-2.5.2 pydantic-core-2.14.5 requests-2.31.0 requirementslib-3.0.0 tomli-2.0.1 tomlkit-0.12.3 typing-extensions-4.8.0 urllib3-2.1.0 wheel-0.42.0 yarg-0.1.9\n##[endgroup]\n##[group]Run echo \"::add-matcher::$GITHUB_ACTION_PATH/isort-matcher.json\"\n\u001b[36;1mecho \"::add-matcher::$GITHUB_ACTION_PATH/isort-matcher.json\"\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n##[group]Run $GITHUB_ACTION_PATH/bin/run_isort --check-only --diff deepchecks/\n\u001b[36;1m$GITHUB_ACTION_PATH/bin/run_isort --check-only --diff deepchecks/\u001b[0m\nshell: /usr/bin/bash --noprofile --norc -e -o pipefail {0}\nenv:\n DISABLE_DEEPCHECKS_ANONYMOUS_TELEMETRY: true\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nRunning isort --check-only --diff deepchecks/\n##[error]ERROR: /home/runner/work/deepchecks/deepchecks/deepchecks/nlp/utils/text_embeddings.py Imports are incorrectly sorted and/or formatted.\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/deepchecks/nlp/utils/text_embeddings.py b/deepchecks/nlp/utils/text_embeddings.py\nindex 758d3554b..85fc5ad47 100644\n--- a/deepchecks/nlp/utils/text_embeddings.py\n+++ b/deepchecks/nlp/utils/text_embeddings.py\n@@ -9,8 +9,8 @@\n # ----------------------------------------------------------------------------\n #\n \"\"\"Utils module for calculating embeddings for text.\"\"\"\n-import sys\n import re\n+import sys\n import warnings\n from itertools import islice\n from typing import Optional\n", "difficulty": 0, "changed_files": ["deepchecks/nlp/utils/text_embeddings.py"], "commit_link": "https://github.com/deepchecks/deepchecks/tree/31c3e1dfba76d6e90b5580da2f80fe4b12e6cb5c"}
|
|
|
|
data/python/385c14d.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 26, "repo_owner": "canonical", "repo_name": "cloud-init", "head_branch": "holman/dhcpcd", "workflow_name": "Lint Tests", "workflow_filename": "check_format.yml", "workflow_path": ".github/workflows/check_format.yml", "contributor": "holmanb", "sha_fail": "385c14d0ae500918cff5565ea836884bfaa2bfa5", "sha_success": "0193e09ca15ed70a351203ab6f21dc52a60d6253", "workflow": "name: Lint Tests\non:\n pull_request:\n push:\n branches:\n - main\n\nconcurrency:\n group: 'ci-${{ github.workflow }}-${{ github.ref }}'\n cancel-in-progress: true\ndefaults:\n run:\n shell: sh -ex {0}\n\njobs:\n check_format:\n strategy:\n fail-fast: false\n matrix:\n env: [ruff, mypy, pylint, black, isort]\n lint-with:\n - {tip-versions: false, os: ubuntu-20.04}\n - {tip-versions: true, os: ubuntu-latest}\n name: Check ${{ matrix.lint-with.tip-versions && 'tip-' || '' }}${{ matrix.env }}\n runs-on: ${{ matrix.lint-with.os }}\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n\n - name: Dependencies\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n\n - name: Print version\n run: python3 --version\n\n - name: Test\n if: ${{ !matrix.lint-with.tip-versions }}\n env:\n # matrix env: not to be confused w/environment variables or testenv\n TOXENV: ${{ matrix.env }}\n run: tox\n - name: Test (tip versions)\n if: matrix.lint-with.tip-versions\n continue-on-error: true\n env:\n TOXENV: tip-${{ matrix.env }}\n run: tox\n schema-format:\n strategy:\n fail-fast: false\n name: Check json format\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Test format\"\n run: |\n tools/check_json_format.sh cloudinit/config/schemas/schema-cloud-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/schema-network-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/versions.schema.cloud-config.json\n\n doc:\n strategy:\n fail-fast: false\n name: Check docs\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n - name: \"Install Python 3.10\"\n uses: actions/setup-python@v4\n with:\n python-version: '3.10.8'\n - name: \"Install dependencies\"\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n - name: \"Spellcheck\"\n run: |\n tox\n env:\n TOXENV: doc-spelling\n - name: \"Build docs\"\n env:\n TOXENV: doc\n run: |\n tox\n\n shell-lint:\n name: Shell Lint\n runs-on: ubuntu-22.04\n steps:\n - name: Checkout\n uses: actions/[email protected]\n\n - name: Install ShellCheck\n run: |\n sudo apt-get update\n sudo apt install shellcheck\n\n - name: Run ShellCheck\n run: |\n shellcheck ./tools/ds-identify\n\n check-cla-signers:\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/[email protected]\n\n - name: Check CLA signers file\n run: tools/check-cla-signers\n", "logs": [{"step_name": "Check pylint/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: pylint\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\npylint create: /home/runner/work/cloud-init/cloud-init/.tox/pylint\npylint installdeps: pylint===2.13.9, -r/home/runner/work/cloud-init/cloud-init/test-requirements.txt, -r/home/runner/work/cloud-init/cloud-init/integration-requirements.txt\npylint inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+77.g80f699fe.zip\npylint installed: adal==1.2.7,anyio==4.2.0,appdirs==1.4.4,applicationinsights==0.11.10,argcomplete==3.1.6,astroid==2.11.7,attrs==23.2.0,autopage==0.5.2,azure-cli-core==2.55.0,azure-cli-telemetry==1.1.0,azure-common==1.1.28,azure-core==1.29.6,azure-identity==1.15.0,azure-mgmt-compute==30.4.0,azure-mgmt-core==1.4.0,azure-mgmt-network==25.2.0,azure-mgmt-resource==23.0.1,bcrypt==4.1.2,boto3==1.34.14,botocore==1.34.14,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,circuitbreaker==1.4.0,cliff==4.4.0,cloud-init==23.4+77.g80f699fe,cmd2==2.4.3,configobj==5.0.8,coverage==7.4.0,cryptography==41.0.7,debtcollector==2.5.0,decorator==5.1.1,dill==0.3.7,distro==1.9.0,dogpile.cache==1.3.0,exceptiongroup==1.2.0,google-api-core==2.15.0,google-api-python-client==2.112.0,google-auth==2.26.1,google-auth-httplib2==0.2.0,googleapis-common-protos==1.62.0,httplib2==0.22.0,humanfriendly==10.0,ibm-cloud-sdk-core==3.18.2,ibm-platform-services==0.48.1,ibm-vpc==0.20.0,idna==3.6,importlib-metadata==7.0.1,importlib-resources==6.1.1,iniconfig==2.0.0,iso8601==2.1.0,isodate==0.6.1,isort==5.13.2,Jinja2==3.1.2,jmespath==1.0.1,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,keystoneauth1==5.4.0,knack==0.11.0,lazy-object-proxy==1.10.0,MarkupSafe==2.1.3,mccabe==0.7.0,msal==1.24.0b2,msal-extensions==1.0.0,msgpack==1.0.7,msrest==0.7.1,msrestazure==0.6.4,netaddr==0.10.1,netifaces==0.11.0,oauthlib==3.2.2,oci==2.118.0,openstacksdk==1.4.0,os-service-types==1.7.0,osc-lib==2.9.0,oslo.config==9.2.0,oslo.i18n==6.2.0,oslo.serialization==5.2.0,oslo.utils==6.3.0,packaging==23.2,paramiko==3.4.0,passlib==1.7.4,pbr==6.0.0,pkginfo==1.9.6,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pluggy==1.3.0,portalocker==2.8.2,prettytable==3.9.0,protobuf==3.19.6,psutil==5.9.7,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycloudlib==1!5.12.1,pycparser==2.21,pygments==2.17.2,PyJWT==2.8.0,pylint==2.13.9,PyNaCl==1.5.0,pyOpenSSL==23.3.0,pyparsing==2.4.7,pyperclip==1.8.2,pyserial==3.5,pytest==7.4.4,pytest-cov==4.1.0,pytest-mock==3.12.0,python-cinderclient==9.4.0,python-dateutil==2.8.2,python-keystoneclient==5.2.0,python-novaclient==18.4.0,python-openstackclient==6.4.0,pytz==2023.3.post1,PyYAML==6.0.1,qemu.qmp==0.0.3,referencing==0.32.1,requests==2.31.0,requests-oauthlib==1.3.1,requestsexceptions==1.4.0,responses==0.24.1,rfc3986==2.0.0,rpds-py==0.16.2,rsa==4.9,s3transfer==0.10.0,simplejson==3.19.2,six==1.16.0,sniffio==1.3.0,stevedore==5.1.0,tabulate==0.9.0,toml==0.10.0,tomli==2.0.1,typing-extensions==4.9.0,tzdata==2023.4,uritemplate==4.1.1,urllib3==2.1.0,wcwidth==0.2.12,wrapt==1.16.0,zipp==3.17.0\npylint run-test-pre: PYTHONHASHSEED='2276146897'\npylint run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py\n************* Module cloudinit.net.dhcp\ncloudinit/net/dhcp.py:575: [W0612(unused-variable), Dhcpcd.dhcp_discovery] Unused variable 'out'\ncloudinit/net/dhcp.py:575: [W0612(unused-variable), Dhcpcd.dhcp_discovery] Unused variable 'err'\n\n------------------------------------\nYour code has been rated at 10.00/10\n\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py (exited with code 4)\n___________________________________ summary ____________________________________\nERROR: pylint: commands failed\n##[error]Process completed with exit code 1.\n"}, {"step_name": "Check black/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: black\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\nblack create: /home/runner/work/cloud-init/cloud-init/.tox/black\nblack installdeps: black===22.3.0\nblack inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+77.g80f699fe.zip\nblack installed: attrs==23.2.0,black==22.3.0,certifi==2023.11.17,charset-normalizer==3.3.2,click==8.1.7,cloud-init==23.4+77.g80f699fe,configobj==5.0.8,idna==3.6,importlib-resources==6.1.1,Jinja2==3.1.2,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,MarkupSafe==2.1.3,mypy-extensions==1.0.0,netifaces==0.11.0,oauthlib==3.2.2,pathspec==0.12.1,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pyserial==3.5,PyYAML==6.0.1,referencing==0.32.1,requests==2.31.0,rpds-py==0.16.2,six==1.16.0,tomli==2.0.1,typing-extensions==4.9.0,urllib3==2.1.0,zipp==3.17.0\nblack run-test-pre: PYTHONHASHSEED='3554362748'\nblack run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/black/bin/python -m black . --check\nwould reformat tests/unittests/net/test_dhcp.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 549 files would be left unchanged.\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/black/bin/python -m black . --check (exited with code 1)\n___________________________________ summary ____________________________________\nERROR: black: commands failed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/cloudinit/net/dhcp.py b/cloudinit/net/dhcp.py\nindex 1b8caee4f..6a4e5a3fa 100644\n--- a/cloudinit/net/dhcp.py\n+++ b/cloudinit/net/dhcp.py\n@@ -175,6 +175,11 @@ class DhcpClient(abc.ABC):\n client_name = \"\"\n max_wait = 5\n \n+ def __init__(self):\n+ self.dhcp_client_path = subp.which(self.client_name)\n+ if not self.dhcp_client_path:\n+ raise NoDHCPLeaseMissingDhclientError()\n+\n @classmethod\n def kill_dhcp_client(cls):\n subp.subp([\"pkill\", cls.client_name], rcs=[0, 1])\n@@ -200,14 +205,6 @@ class DhcpClient(abc.ABC):\n class IscDhclient(DhcpClient):\n client_name = \"dhclient\"\n \n- def __init__(self):\n- self.dhclient_path = subp.which(\"dhclient\")\n- if not self.dhclient_path:\n- LOG.debug(\n- \"Skip dhclient configuration: No dhclient command found.\"\n- )\n- raise NoDHCPLeaseMissingDhclientError()\n-\n @staticmethod\n def parse_dhcp_lease_file(lease_file: str) -> List[Dict[str, Any]]:\n \"\"\"Parse the given dhcp lease file returning all leases as dicts.\n@@ -298,7 +295,7 @@ class IscDhclient(DhcpClient):\n try:\n out, err = subp.subp(\n distro.build_dhclient_cmd(\n- self.dhclient_path,\n+ self.dhcp_client_path,\n lease_file,\n pid_file,\n interface,\n@@ -536,14 +533,6 @@ class IscDhclient(DhcpClient):\n class Dhcpcd:\n client_name = \"dhcpcd\"\n \n- def __init__(self):\n- self.dhclient_path = subp.which(\"dhcpcd\")\n- if not self.dhclient_path:\n- LOG.debug(\n- \"Skip dhclient configuration: No dhclient command found.\"\n- )\n- raise NoDHCPLeaseMissingDhclientError()\n-\n def dhcp_discovery(\n self,\n interface,\n@@ -572,9 +561,9 @@ class Dhcpcd:\n # /lib/dhcpcd/dhcpcd-hooks/ and pass each of those with the --nohook\n # argument to dhcpcd\n try:\n- out, err = subp.subp(\n+ subp.subp(\n [\n- \"dhcpcd\",\n+ self.client_name,\n \"--oneshot\", # get lease then exit\n \"--nobackground\", # don't fork\n \"--ipv4only\", # only attempt configuring ipv4\n@@ -584,6 +573,7 @@ class Dhcpcd:\n interface,\n ]\n )\n+ return self.parse_dhcp_lease_file(interface)\n except subp.ProcessExecutionError as error:\n LOG.debug(\n \"dhclient exited with code: %s stderr: %r stdout: %r\",\n@@ -592,7 +582,6 @@ class Dhcpcd:\n error.stdout,\n )\n raise NoDHCPLeaseError from error\n- return self.parse_dhcp_lease_file(interface)\n \n @staticmethod\n def parse_dhcpcd_lease(lease_dump: str, interface: str) -> List[dict]:\n@@ -656,7 +645,7 @@ class Dhcpcd:\n \n except subp.ProcessExecutionError as error:\n LOG.debug(\n- \"dhclient exited with code: %s stderr: %r stdout: %r\",\n+ \"dhcpcd exited with code: %s stderr: %r stdout: %r\",\n error.exit_code,\n error.stderr,\n error.stdout,\n@@ -667,12 +656,6 @@ class Dhcpcd:\n class Udhcpc(DhcpClient):\n client_name = \"udhcpc\"\n \n- def __init__(self):\n- self.udhcpc_path = subp.which(\"udhcpc\")\n- if not self.udhcpc_path:\n- LOG.debug(\"Skip udhcpc configuration: No udhcpc command found.\")\n- raise NoDHCPLeaseMissingUdhcpcError()\n-\n def dhcp_discovery(\n self,\n interface,\n@@ -702,7 +685,7 @@ class Udhcpc(DhcpClient):\n util.write_file(udhcpc_script, UDHCPC_SCRIPT, 0o755)\n \n cmd = [\n- self.udhcpc_path,\n+ self.dhcp_client_path,\n \"-O\",\n \"staticroutes\",\n \"-i\",\ndiff --git a/tests/unittests/net/test_dhcp.py b/tests/unittests/net/test_dhcp.py\nindex 2841a7244..fa8308068 100644\n--- a/tests/unittests/net/test_dhcp.py\n+++ b/tests/unittests/net/test_dhcp.py\n@@ -8,6 +8,7 @@ import pytest\n import responses\n \n from cloudinit.net.dhcp import (\n+ Dhcpcd,\n InvalidDHCPLeaseFileError,\n IscDhclient,\n NoDHCPLeaseError,\n@@ -438,7 +439,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n subp.ProcessExecutionError(exit_code=-5),\n ]\n \n- m_which.side_effect = [False, True]\n+ m_which.side_effect = [False, False, False]\n with pytest.raises(NoDHCPLeaseError):\n maybe_perform_dhcp_discovery(MockDistro())\n \n@@ -446,10 +447,6 @@ class TestDHCPDiscoveryClean(CiTestCase):\n \"DHCP client not found: dhclient\",\n self.logs.getvalue(),\n )\n- self.assertIn(\n- \"DHCP client not found: dhcpcd\",\n- self.logs.getvalue(),\n- )\n \n @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=None)\n def test_provided_nic_does_not_exist(self, m_fallback_nic):\n@@ -473,7 +470,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n maybe_perform_dhcp_discovery(MockDistro())\n \n self.assertIn(\n- \"Skip dhclient configuration: No dhclient command found.\",\n+ \"DHCP client not found: dhclient\",\n self.logs.getvalue(),\n )\n \n@@ -975,24 +972,6 @@ class TestUDHCPCDiscoveryClean(CiTestCase):\n with_logs = True\n maxDiff = None\n \n- @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n- def test_absent_udhcpc_command(self, m_fallback, m_which):\n- \"\"\"When dhclient doesn't exist in the OS, log the issue and no-op.\"\"\"\n- m_fallback.return_value = \"eth9\"\n- m_which.return_value = None # udhcpc isn't found\n-\n- distro = MockDistro()\n- distro.dhcp_client_priority = [Udhcpc]\n-\n- with pytest.raises(NoDHCPLeaseMissingDhclientError):\n- maybe_perform_dhcp_discovery(distro)\n-\n- self.assertIn(\n- \"Skip udhcpc configuration: No udhcpc command found.\",\n- self.logs.getvalue(),\n- )\n-\n @mock.patch(\"cloudinit.net.dhcp.is_ib_interface\", return_value=False)\n @mock.patch(\"cloudinit.net.dhcp.subp.which\", return_value=\"/sbin/udhcpc\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n@@ -1135,3 +1114,30 @@ class TestUDHCPCDiscoveryClean(CiTestCase):\n ),\n ]\n )\n+\n+\n+class TestDhcpcd:\n+ def test_parse_lease(self):\n+ lease = dedent(\n+ \"\"\"\n+ broadcast_address='192.168.15.255'\n+ dhcp_lease_time='3600'\n+ dhcp_message_type='5'\n+ dhcp_server_identifier='192.168.0.1'\n+ domain_name='us-east-2.compute.internal'\n+ domain_name_servers='192.168.0.2'\n+ host_name='ip-192-168-0-212'\n+ interface_mtu='9001'\n+ ip_address='192.168.0.212'\n+ network_number='192.168.0.0'\n+ routers='192.168.0.1'\n+ subnet_cidr='20'\n+ subnet_mask='255.255.240.0'\n+ \"\"\"\n+ )\n+ parsed_lease = Dhcpcd.parse_dhcpcd_lease(lease, \"eth0\")[0]\n+ assert \"eth0\" == parsed_lease[\"interface\"]\n+ assert \"192.168.15.255\" == parsed_lease[\"broadcast-address\"]\n+ assert \"192.168.0.212\" == parsed_lease[\"fixed-address\"]\n+ assert \"255.255.240.0\" == parsed_lease[\"subnet-mask\"]\n+ assert \"192.168.0.1\" == parsed_lease[\"routers\"]\n", "difficulty": 0, "changed_files": ["cloudinit/net/dhcp.py", "tests/unittests/net/test_dhcp.py"], "commit_link": "https://github.com/canonical/cloud-init/tree/385c14d0ae500918cff5565ea836884bfaa2bfa5"}
|
|
|
|
data/python/386bf6f.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 196, "repo_owner": "danielegrattarola", "repo_name": "spektral", "head_branch": "master", "workflow_name": "Run examples", "workflow_filename": "examples.yml", "workflow_path": ".github/workflows/examples.yml", "contributor": "danielegrattarola", "sha_fail": "386bf6f0815368b78261be43bf90e203dfe9c13f", "sha_success": "bf6071475028cf9711b8264e8952deb2aee2e32b", "workflow": "name: Run examples\n\non: [push, pull_request]\n\njobs:\n build:\n\n runs-on: ubuntu-latest\n\n steps:\n - uses: actions/checkout@v2\n - name: Set up Python 3.11\n uses: actions/setup-python@v2\n with:\n python-version: 3.11\n - name: Install dependencies\n run: |\n pip install ogb matplotlib\n - name: Install Spektral\n run: |\n pip install .\n - name: Just one epoch\n run: |\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/node_prediction/*.py\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/graph_prediction/*.py\n sed -i -e 's/epochs = /epochs = 1 #/g' examples/other/*.py\n - name: Run all examples\n run: |\n cd examples/node_prediction/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n cd graph_prediction/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n cd other/\n for f in *.py; do\n echo \"##### $f #####\"\n python $f\n done\n cd ..\n", "logs": [{"step_name": "build/7_Run all examples.txt", "log": "##[group]Run cd examples/node_prediction/\n\u001b[36;1mcd examples/node_prediction/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\n\u001b[36;1mcd graph_prediction/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\n\u001b[36;1mcd other/\u001b[0m\n\u001b[36;1mfor f in *.py; do\u001b[0m\n\u001b[36;1m echo \"##### $f #####\"\u001b[0m\n\u001b[36;1m python $f\u001b[0m\n\u001b[36;1mdone\u001b[0m\n\u001b[36;1mcd ..\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n##### citation_arma.py #####\n2024-01-21 16:14:04.697281: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:04.697333: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:04.698603: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:04.704823: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:05.946659: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nDownloading cora dataset.\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n arma_conv (ARMAConv) (None, 16) 91744 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n dropout (Dropout) (None, 16) 0 ['arma_conv[0][0]'] \n \n arma_conv_1 (ARMAConv) (None, 7) 231 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 91975 (359.28 KB)\nTrainable params: 91975 (359.28 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 0.1167 - acc: 0.1714\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 1s 938ms/step - loss: 0.1167 - acc: 0.1714 - val_loss: 0.3394 - val_acc: 0.3080\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 0.6695 - acc: 0.3360\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 12ms/step - loss: 0.6695 - acc: 0.3360\nDone.\nTest loss: 0.6695395708084106\nTest accuracy: 0.335999995470047\n##### citation_cheby.py #####\n2024-01-21 16:14:13.031828: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:13.031877: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:13.033233: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:13.039361: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:13.892745: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n dropout (Dropout) (None, 1433) 0 ['input_1[0][0]'] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n cheb_conv (ChebConv) (None, 16) 45856 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n dropout_1 (Dropout) (None, 16) 0 ['cheb_conv[0][0]'] \n \n cheb_conv_1 (ChebConv) (None, 7) 224 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 46080 (180.00 KB)\nTrainable params: 46080 (180.00 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 1.9901 - acc: 0.1357\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 492ms/step - loss: 1.9901 - acc: 0.1357 - val_loss: 1.9229 - val_acc: 0.2740\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9236 - acc: 0.2650\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 15ms/step - loss: 1.9236 - acc: 0.2650\nDone.\nTest loss: 1.9235953092575073\nTest accuracy: 0.26499998569488525\n##### citation_gat.py #####\n2024-01-21 16:14:17.870770: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:17.870818: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:17.872219: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:17.878280: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:18.689165: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nPre-processing node features\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n dropout (Dropout) (None, 1433) 0 ['input_1[0][0]'] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n gat_conv (GATConv) (None, 64) 91904 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n dropout_1 (Dropout) (None, 64) 0 ['gat_conv[0][0]'] \n \n gat_conv_1 (GATConv) (None, 7) 469 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 92373 (360.83 KB)\nTrainable params: 92373 (360.83 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 1.9498 - acc: 0.1429\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 2s 2s/step - loss: 1.9498 - acc: 0.1429 - val_loss: 1.9496 - val_acc: 0.2020\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9495 - acc: 0.2120\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 15ms/step - loss: 1.9495 - acc: 0.2120\nDone.\nTest loss: 1.9495328664779663\nTest accuracy: 0.21199998259544373\n##### citation_gat_custom.py #####\n2024-01-21 16:14:24.090507: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:24.090565: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:24.092222: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:24.098707: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:24.924070: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nPre-processing node features\nLoss tr: 1.9471, Acc tr: 0.3214, Loss va: 1.9493, Acc va: 0.1920, Loss te: 1.9494, Acc te: 0.1790\nImproved\nGAT (1 epochs)\nElapsed: 2.46s\n##### citation_gcn.py #####\n2024-01-21 16:14:30.891866: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:30.891915: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:30.893225: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:30.899400: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:31.735939: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nPre-processing node features\n\n1/1 [==============================] - ETA: 0s - loss: 1.9535 - acc: 0.1286\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 1s 1s/step - loss: 1.9535 - acc: 0.1286 - val_loss: 1.9469 - val_acc: 0.4000\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 1.9470 - acc: 0.3910\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 9ms/step - loss: 1.9470 - acc: 0.3910\nDone.\nTest loss: 1.946979284286499\nTest accuracy: 0.390999972820282\n##### citation_gcn_custom.py #####\n2024-01-21 16:14:35.783612: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:35.783669: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:35.785139: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:35.791466: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:36.608646: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nPre-processing node features\nSpektral - GCN (200 epochs)\nElapsed: 2.85s\nFinal loss = 0.617151141166687\n##### citation_simple_gc.py #####\n2024-01-21 16:14:43.435717: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:43.435759: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:43.437049: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:43.443159: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:44.284258: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/scipy/sparse/_index.py:145: SparseEfficiencyWarning: Changing the sparsity structure of a csr_matrix is expensive. lil_matrix is more efficient.\n self._set_arrayXarray(i, j, x)\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 1433)] 0 [] \n \n input_2 (InputLayer) [(None, 2708)] 0 [] \n \n gcn_conv (GCNConv) (None, 7) 10031 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 10031 (39.18 KB)\nTrainable params: 10031 (39.18 KB)\nNon-trainable params: 0 (0.00 Byte)\n__________________________________________________________________________________________________\n\n1/1 [==============================] - ETA: 0s - loss: 0.1008 - acc: 0.1357\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 355ms/step - loss: 0.1008 - acc: 0.1357 - val_loss: 0.2166 - val_acc: 0.6780\nEvaluating model.\n\n1/1 [==============================] - ETA: 0s - loss: 0.4094 - acc: 0.7070\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\b\n1/1 [==============================] - 0s 12ms/step - loss: 0.4094 - acc: 0.7070\nDone.\nTest loss: 0.40943244099617004\nTest accuracy: 0.7070000171661377\n##### ogbn-arxiv_gcn.py #####\n2024-01-21 16:14:47.998765: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:14:47.998814: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:14:48.000133: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:14:48.006402: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:14:48.837666: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\nDownloading http://snap.stanford.edu/ogb/data/nodeproppred/arxiv.zip\n\n 0%| | 0/81 [00:00<?, ?it/s]\nDownloaded 0.00 GB: 0%| | 0/81 [00:00<?, ?it/s]\nDownloaded 0.00 GB: 1%| | 1/81 [00:00<01:13, 1.09it/s]\nDownloaded 0.00 GB: 1%| | 1/81 [00:01<01:13, 1.09it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 2%|\u258f | 2/81 [00:01<00:40, 1.95it/s]\nDownloaded 0.00 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.00 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 5%|\u258d | 4/81 [00:01<00:20, 3.81it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 9%|\u258a | 7/81 [00:01<00:10, 7.06it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.01 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 15%|\u2588\u258d | 12/81 [00:01<00:05, 12.81it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 21%|\u2588\u2588 | 17/81 [00:01<00:03, 17.75it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:01<00:02, 23.06it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.02 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 28%|\u2588\u2588\u258a | 23/81 [00:02<00:02, 23.06it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 36%|\u2588\u2588\u2588\u258c | 29/81 [00:02<00:01, 27.07it/s]\nDownloaded 0.03 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 43%|\u2588\u2588\u2588\u2588\u258e | 35/81 [00:02<00:01, 30.04it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 49%|\u2588\u2588\u2588\u2588\u2589 | 40/81 [00:02<00:01, 33.77it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.04 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 54%|\u2588\u2588\u2588\u2588\u2588\u258d | 44/81 [00:02<00:01, 32.84it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 60%|\u2588\u2588\u2588\u2588\u2588\u2588 | 49/81 [00:02<00:00, 35.51it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.05 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.06 GB: 65%|\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 53/81 [00:02<00:00, 35.14it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 70%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 57/81 [00:02<00:00, 36.16it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:02<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 75%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 61/81 [00:03<00:00, 37.12it/s]\nDownloaded 0.06 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.06 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 80%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588 | 65/81 [00:03<00:00, 37.16it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 85%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 69/81 [00:03<00:00, 36.70it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.07 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 91%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258f| 74/81 [00:03<00:00, 38.00it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 96%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258b| 78/81 [00:03<00:00, 36.73it/s]\nDownloaded 0.08 GB: 100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 81/81 [00:03<00:00, 23.38it/s]\n\n 0%| | 0/1 [00:00<?, ?it/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 1/1 [00:00<00:00, 21076.90it/s]\nExtracting dataset/arxiv.zip\nLoading necessary files...\nThis might take a while.\nProcessing graphs...\nSaving...\nModel: \"model\"\n__________________________________________________________________________________________________\n Layer (type) Output Shape Param # Connected to \n==================================================================================================\n input_1 (InputLayer) [(None, 128)] 0 [] \n \n input_2 (InputLayer) [(None, 169343)] 0 [] \n \n gcn_conv (GCNConv) (None, 256) 33024 ['input_1[0][0]', \n 'input_2[0][0]'] \n \n batch_normalization (Batch (None, 256) 1024 ['gcn_conv[0][0]'] \n Normalization) \n \n dropout (Dropout) (None, 256) 0 ['batch_normalization[0][0]'] \n \n gcn_conv_1 (GCNConv) (None, 256) 65792 ['dropout[0][0]', \n 'input_2[0][0]'] \n \n batch_normalization_1 (Bat (None, 256) 1024 ['gcn_conv_1[0][0]'] \n chNormalization) \n \n dropout_1 (Dropout) (None, 256) 0 ['batch_normalization_1[0][0]'\n ] \n \n gcn_conv_2 (GCNConv) (None, 40) 10280 ['dropout_1[0][0]', \n 'input_2[0][0]'] \n \n==================================================================================================\nTotal params: 111144 (434.16 KB)\nTrainable params: 110120 (430.16 KB)\nNon-trainable params: 1024 (4.00 KB)\n__________________________________________________________________________________________________\nEp. 1 - Loss: 5.289 - Acc: 0.109 - Val acc: 0.177 - Test acc: 0.241\nEvaluating model.\nDone! - Test acc: 0.241\n##### custom_dataset.py #####\n2024-01-21 16:15:07.319425: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:15:07.319469: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:15:07.320948: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:15:07.327116: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:15:08.139878: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/data/utils.py:221: UserWarning: you are shuffling a 'MyDataset' object which is not a subclass of 'Sequence'; `shuffle` is not guaranteed to behave correctly. E.g., non-numpy array/tensor objects with view semantics may contain duplicates after shuffling.\n np.random.shuffle(a)\n/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/keras/src/initializers/initializers.py:120: UserWarning: The initializer GlorotUniform is unseeded and being called multiple times, which will return identical values each time (even if the initializer is unseeded). Please update your code to provide a seed to the initializer, or avoid using the same initializer instance more than once.\n warnings.warn(\nEp. 1 - Loss: 0.729 - Acc: 0.694 - Val loss: 0.334 - Val acc: 0.880\nNew best val_loss 0.334\nDone. Test loss: 0.4293. Test acc: 0.84\n##### general_gnn.py #####\n2024-01-21 16:15:13.211838: E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9261] Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered\n2024-01-21 16:15:13.211888: E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607] Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered\n2024-01-21 16:15:13.213185: E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1515] Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered\n2024-01-21 16:15:13.219400: I tensorflow/core/platform/cpu_feature_guard.cc:182] This TensorFlow binary is optimized to use available CPU instructions in performance-critical operations.\nTo enable the following instructions: AVX2 FMA, in other operations, rebuild TensorFlow with the appropriate compiler flags.\n2024-01-21 16:15:14.044203: W tensorflow/compiler/tf2tensorrt/utils/py_utils.cc:38] TF-TRT Warning: Could not find TensorRT\nDownloading PROTEINS dataset.\n\n 0%| | 0.00/447k [00:00<?, ?B/s]\n 7%|\u2588\u2588\u258a | 32.0k/447k [00:00<00:01, 249kB/s]\n 21%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 96.0k/447k [00:00<00:00, 419kB/s]\n 50%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u258c | 224k/447k [00:00<00:00, 777kB/s]\n100%|\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588\u2588| 447k/447k [00:00<00:00, 1.02MB/s]\nTraceback (most recent call last):\n File \"/home/runner/work/spektral/spektral/examples/graph_prediction/general_gnn.py\", line 39, in <module>\n data = TUDataset(\"PROTEINS\")\n ^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 66, in __init__\n super().__init__(**kwargs)\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/data/dataset.py\", line 118, in __init__\n self.graphs = self.read()\n ^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 128, in read\n [_normalize(xl_[:, None], \"ohe\") for xl_ in x_labs.T], -1\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 128, in <listcomp>\n [_normalize(xl_[:, None], \"ohe\") for xl_ in x_labs.T], -1\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n File \"/opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/spektral/datasets/tudataset.py\", line 219, in _normalize\n fnorm = OneHotEncoder(sparse=False, categories=\"auto\")\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\nTypeError: OneHotEncoder.__init__() got an unexpected keyword argument 'sparse'\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/spektral/datasets/tudataset.py b/spektral/datasets/tudataset.py\nindex 3877535..f26b0de 100644\n--- a/spektral/datasets/tudataset.py\n+++ b/spektral/datasets/tudataset.py\n@@ -216,7 +216,7 @@ def _normalize(x, norm=None):\n Apply one-hot encoding or z-score to a list of node features\n \"\"\"\n if norm == \"ohe\":\n- fnorm = OneHotEncoder(sparse=False, categories=\"auto\")\n+ fnorm = OneHotEncoder(sparse_output=False, categories=\"auto\")\n elif norm == \"zscore\":\n fnorm = StandardScaler()\n else:\n", "difficulty": 2, "changed_files": ["spektral/datasets/tudataset.py"], "commit_link": "https://github.com/danielegrattarola/spektral/tree/386bf6f0815368b78261be43bf90e203dfe9c13f"}
|
|
|
|
data/python/38a3024.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 165, "repo_owner": "prompt-toolkit", "repo_name": "python-prompt-toolkit", "head_branch": "master", "workflow_name": "test", "workflow_filename": "test.yaml", "workflow_path": ".github/workflows/test.yaml", "contributor": "auxsvr", "sha_fail": "38a30240c4e21de4c03322cfbf982119094bdf45", "sha_success": "55cde3eb0e1be72d9232e979a708f1a792339ad9", "workflow": "name: test\n\non:\n push: # any branch\n pull_request:\n branches: [master]\n\nenv:\n FORCE_COLOR: 1\n\njobs:\n test-ubuntu:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\"]\n\n steps:\n - uses: actions/checkout@v4\n - name: Setup Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n allow-prereleases: true\n - name: Install Dependencies\n run: |\n sudo apt remove python3-pip\n python -m pip install --upgrade pip\n python -m pip install . ruff typos coverage codecov mypy pytest readme_renderer types-contextvars asyncssh\n pip list\n - name: Ruff\n run: |\n ruff .\n ruff format --check .\n typos .\n - name: Tests\n run: |\n coverage run -m pytest\n - if: \"matrix.python-version != '3.7'\"\n name: Mypy\n # Check whether the imports were sorted correctly.\n # When this fails, please run ./tools/sort-imports.sh\n run: |\n mypy --strict src/prompt_toolkit --platform win32\n mypy --strict src/prompt_toolkit --platform linux\n mypy --strict src/prompt_toolkit --platform darwin\n - name: Validate README.md\n # Ensure that the README renders correctly (required for uploading to PyPI).\n run: |\n python -m readme_renderer README.rst > /dev/null\n - name: Run codecov\n run: |\n codecov\n", "logs": [{"step_name": "test-ubuntu (3.7)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.8)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.9)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.10)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.11)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.11.6/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}, {"step_name": "test-ubuntu (3.12)/5_Ruff.txt", "log": "##[group]Run ruff .\n\u001b[36;1mruff .\u001b[0m\n\u001b[36;1mruff format --check .\u001b[0m\n\u001b[36;1mtypos .\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n FORCE_COLOR: 1\n pythonLocation: /opt/hostedtoolcache/Python/3.12.0/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.0/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.0/x64/lib\n##[endgroup]\nWould reformat: src/prompt_toolkit/application/application.py\nWould reformat: src/prompt_toolkit/eventloop/inputhook.py\n2 files would be reformatted, 284 files left unchanged\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/src/prompt_toolkit/application/application.py b/src/prompt_toolkit/application/application.py\nindex 6a8f0290..2dcb3130 100644\n--- a/src/prompt_toolkit/application/application.py\n+++ b/src/prompt_toolkit/application/application.py\n@@ -985,7 +985,7 @@ class Application(Generic[_AppResult]):\n # See whether a loop was installed already. If so, use that.\n # That's required for the input hooks to work, they are\n # installed using `set_event_loop`.\n- loop = asyncio.get_running_loop()\n+ loop = asyncio.get_event_loop()\n except RuntimeError:\n # No loop installed. Run like usual.\n return asyncio.run(coro)\n@@ -1545,7 +1545,7 @@ async def _do_wait_for_enter(wait_text: AnyFormattedText) -> None:\n \n @contextmanager\n def attach_winch_signal_handler(\n- handler: Callable[[], None]\n+ handler: Callable[[], None],\n ) -> Generator[None, None, None]:\n \"\"\"\n Attach the given callback as a WINCH signal handler within the context\ndiff --git a/src/prompt_toolkit/eventloop/inputhook.py b/src/prompt_toolkit/eventloop/inputhook.py\nindex 5731573f..a4c0eee6 100644\n--- a/src/prompt_toolkit/eventloop/inputhook.py\n+++ b/src/prompt_toolkit/eventloop/inputhook.py\n@@ -66,7 +66,7 @@ InputHook: TypeAlias = Callable[[InputHookContext], None]\n \n \n def new_eventloop_with_inputhook(\n- inputhook: Callable[[InputHookContext], None]\n+ inputhook: Callable[[InputHookContext], None],\n ) -> AbstractEventLoop:\n \"\"\"\n Create a new event loop with the given inputhook.\n@@ -77,7 +77,7 @@ def new_eventloop_with_inputhook(\n \n \n def set_eventloop_with_inputhook(\n- inputhook: Callable[[InputHookContext], None]\n+ inputhook: Callable[[InputHookContext], None],\n ) -> AbstractEventLoop:\n \"\"\"\n Create a new event loop with the given inputhook, and activate it.\ndiff --git a/src/prompt_toolkit/patch_stdout.py b/src/prompt_toolkit/patch_stdout.py\nindex 46619cac..528bec7f 100644\n--- a/src/prompt_toolkit/patch_stdout.py\n+++ b/src/prompt_toolkit/patch_stdout.py\n@@ -203,6 +203,13 @@ class StdoutProxy:\n \"\"\"\n \n def write_and_flush() -> None:\n+ # Ensure that autowrap is enabled before calling `write`.\n+ # XXX: On Windows, the `Windows10_Output` enables/disables VT\n+ # terminal processing for every flush. It turns out that this\n+ # causes autowrap to be reset (disabled) after each flush. So,\n+ # we have to enable it again before writing text.\n+ self._output.enable_autowrap()\n+\n if self.raw:\n self._output.write_raw(text)\n else:\n", "difficulty": 0, "changed_files": ["src/prompt_toolkit/application/application.py", "src/prompt_toolkit/eventloop/inputhook.py", "src/prompt_toolkit/patch_stdout.py"], "commit_link": "https://github.com/prompt-toolkit/python-prompt-toolkit/tree/38a30240c4e21de4c03322cfbf982119094bdf45"}
|
|
|
|
data/python/3dd8e44.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 128, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "device-agnostic-testing", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "statelesshz", "sha_fail": "3dd8e4404a0ce2e29db4911dc2cd7e94755be631", "sha_success": "c9a13fe9a890cfb48ee6c7b1c6abd64c8fb432da", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\ntests/deepspeed/test_deepspeed.py:15:1: I001 [*] Import block is un-sorted or un-formatted\ntests/fsdp/test_fsdp.py:16:1: I001 [*] Import block is un-sorted or un-formatted\nFound 2 errors.\n[*] 2 fixable with the `--fix` option.\nmake: *** [Makefile:16: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/tests/deepspeed/test_deepspeed.py b/tests/deepspeed/test_deepspeed.py\nindex 035965c..612f234 100644\n--- a/tests/deepspeed/test_deepspeed.py\n+++ b/tests/deepspeed/test_deepspeed.py\n@@ -36,9 +36,9 @@ from accelerate.test_utils.testing import (\n AccelerateTestCase,\n TempDirTestCase,\n execute_subprocess_async,\n- require_non_cpu,\n require_deepspeed,\n require_multi_device,\n+ require_non_cpu,\n slow,\n )\n from accelerate.test_utils.training import RegressionDataset\ndiff --git a/tests/fsdp/test_fsdp.py b/tests/fsdp/test_fsdp.py\nindex d939991..c494f5e 100644\n--- a/tests/fsdp/test_fsdp.py\n+++ b/tests/fsdp/test_fsdp.py\n@@ -28,9 +28,9 @@ from accelerate.test_utils.testing import (\n AccelerateTestCase,\n TempDirTestCase,\n execute_subprocess_async,\n- require_non_cpu,\n require_fsdp,\n require_multi_device,\n+ require_non_cpu,\n slow,\n )\n from accelerate.utils.constants import (\n", "difficulty": 0, "changed_files": ["tests/deepspeed/test_deepspeed.py", "tests/fsdp/test_fsdp.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/3dd8e4404a0ce2e29db4911dc2cd7e94755be631"}
|
|
|
|
data/python/3ed7a88.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/3fcfe29.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 110, "repo_owner": "mikf", "repo_name": "gallery-dl", "head_branch": "master", "workflow_name": "tests", "workflow_filename": "tests.yml", "workflow_path": ".github/workflows/tests.yml", "contributor": "bun-dev", "sha_fail": "3fcfe2925441e86f02d700a3bd080ee7d4c85153", "sha_success": "c51a3c67884331c409a05779bcec57ecd402cd18", "workflow": "name: tests\n\non:\n push:\n branches:\n - master\n pull_request:\n branches:\n - master\n\njobs:\n build:\n\n runs-on: ubuntu-20.04\n strategy:\n fail-fast: false\n matrix:\n python-version: [\"3.5\", \"3.6\", \"3.7\", \"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.12\", \"pypy3.9\"]\n\n steps:\n - uses: actions/checkout@v4\n\n - name: Check file permissions\n run: |\n if [[ \"$(find ./gallery_dl -type f -not -perm 644)\" ]]; then exit 1; fi\n\n - name: Set up Python ${{ matrix.python-version }}\n uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python-version }}\n\n - name: Install dependencies\n run: |\n pip install -r requirements.txt\n pip install flake8 youtube-dl\n\n - name: Install yt-dlp\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5)\n # don't install yt-dlp\n ;;\n 3.6|3.7)\n # install from PyPI\n pip install yt-dlp\n ;;\n *)\n # install from master\n pip install https://github.com/yt-dlp/yt-dlp/archive/refs/heads/master.tar.gz\n ;;\n esac\n\n - name: Lint with flake8\n run: |\n case \"${{ matrix.python-version }}\" in\n 3.4|3.5|3.6|3.7)\n flake8 --extend-exclude scripts/export_tests.py .\n ;;\n *)\n flake8 .\n ;;\n esac\n\n - name: Run tests\n run: |\n make test\n\n - name: Test autogeneration of man pages, bash/zsh/fish completion, etc\n run: |\n make\n", "logs": [{"step_name": "build (3.5)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.5\" in\n\u001b[36;1mcase \"3.5\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.5.10/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.5.10/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.5.10/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.6)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.6\" in\n\u001b[36;1mcase \"3.6\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.6.15/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.6.15/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.6.15/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.7)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.7\" in\n\u001b[36;1mcase \"3.7\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.7.17/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.7.17/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.7.17/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.8)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.8\" in\n\u001b[36;1mcase \"3.8\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.9\" in\n\u001b[36;1mcase \"3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.10)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.10\" in\n\u001b[36;1mcase \"3.10\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.11)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.11\" in\n\u001b[36;1mcase \"3.11\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (3.12)/7_Lint with flake8.txt", "log": "##[group]Run case \"3.12\" in\n\u001b[36;1mcase \"3.12\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "build (pypy3.9)/7_Lint with flake8.txt", "log": "##[group]Run case \"pypy3.9\" in\n\u001b[36;1mcase \"pypy3.9\" in\u001b[0m\n\u001b[36;1m 3.4|3.5|3.6|3.7)\u001b[0m\n\u001b[36;1m flake8 --extend-exclude scripts/export_tests.py .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1m *)\u001b[0m\n\u001b[36;1m flake8 .\u001b[0m\n\u001b[36;1m ;;\u001b[0m\n\u001b[36;1mesac\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/PyPy/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/PyPy/3.9.18/x64/bin/lib/pkgconfig\n##[endgroup]\n./gallery_dl/extractor/shimmie2.py:226:52: E126 continuation line over-indented for hanging indent\n./gallery_dl/extractor/shimmie2.py:226:80: E501 line too long (82 > 79 characters)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gallery_dl/extractor/shimmie2.py b/gallery_dl/extractor/shimmie2.py\nindex a9ed615b..299b2394 100644\n--- a/gallery_dl/extractor/shimmie2.py\n+++ b/gallery_dl/extractor/shimmie2.py\n@@ -223,7 +223,7 @@ class Shimmie2TagExtractor(Shimmie2Extractor):\n break\n \n tags, dimensions, size, ext = extr(\n- \"title=\\\"\", \"\\\"\").split(\" // \")\n+ \"title=\\\"\", \"\\\"\").split(\" // \")\n width, _, height = dimensions.partition(\"x\")\n md5 = extr(\"/_thumbs/\", \"/\")\n \n", "difficulty": 0, "changed_files": ["gallery_dl/extractor/shimmie2.py"], "commit_link": "https://github.com/mikf/gallery-dl/tree/3fcfe2925441e86f02d700a3bd080ee7d4c85153"}
|
|
|
|
data/python/434321a.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/43dd59c.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 9, "repo_owner": "yt-dlp", "repo_name": "yt-dlp", "head_branch": "master", "workflow_name": "Quick Test", "workflow_filename": "quick-test.yml", "workflow_path": ".github/workflows/quick-test.yml", "contributor": "Kenshin9977", "sha_fail": "43dd59c3137df77f5dd22cef4cb7bedfe9f6b12e", "sha_success": "0bedc0ee590f45a6229a4b78f83297f8c07392bb", "workflow": "name: Quick Test\non: [push, pull_request]\npermissions:\n contents: read\n\njobs:\n tests:\n name: Core Test\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - name: Set up Python 3.11\n uses: actions/setup-python@v4\n with:\n python-version: '3.11'\n - name: Install test requirements\n run: pip install pytest pycryptodomex\n - name: Run tests\n run: |\n python3 -m yt_dlp -v || true\n ./devscripts/run_tests.sh core\n flake8:\n name: Linter\n if: \"!contains(github.event.head_commit.message, 'ci skip all')\"\n runs-on: ubuntu-latest\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n - name: Install flake8\n run: pip install flake8\n - name: Make lazy extractors\n run: python devscripts/make_lazy_extractors.py\n - name: Run flake8\n run: flake8 .\n", "logs": [{"step_name": "Core Test/5_Run tests.txt", "log": "##[group]Run python3 -m yt_dlp -v || true\n\u001b[36;1mpython3 -m yt_dlp -v || true\u001b[0m\n\u001b[36;1m./devscripts/run_tests.sh core\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.6/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.6/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.6/x64/lib\n##[endgroup]\n[debug] Command-line config: ['-v']\n[debug] Encodings: locale UTF-8, fs utf-8, pref UTF-8, out utf-8 (No ANSI), error utf-8 (No ANSI), screen utf-8 (No ANSI)\n[debug] yt-dlp version [email protected] [b634ba742] (source)\n[debug] Lazy loading extractors is disabled\n[debug] Git HEAD: f8abf45\n[debug] Python 3.11.6 (CPython x86_64 64bit) - Linux-6.2.0-1015-azure-x86_64-with-glibc2.35 (OpenSSL 3.0.2 15 Mar 2022, glibc 2.35)\n[debug] exe versions: none\n[debug] Optional libraries: Cryptodome-3.19.0, sqlite3-3.37.2\n[debug] Proxy map: {}\n[debug] Request Handlers: urllib\n[debug] Loaded 1890 extractors\n\nUsage: yt-dlp [OPTIONS] URL [URL...]\n\nyt-dlp: error: You must provide at least one URL.\nType yt-dlp --help to see a list of all options.\n============================= test session starts ==============================\nplatform linux -- Python 3.11.6, pytest-7.4.3, pluggy-1.3.0 -- /opt/hostedtoolcache/Python/3.11.6/x64/bin/python3\ncachedir: .pytest_cache\nrootdir: /home/runner/work/yt-dlp/yt-dlp\nconfigfile: setup.cfg\ncollecting ... collected 6320 items / 5705 deselected / 615 selected\n\ntest/test_InfoExtractor.py::TestInfoExtractor::test_download_json PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_extract_jwplayer_data_realworld PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_html_search_meta PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_html_search_regex PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_ie_key PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_opengraph PASSED [ 0%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_f4m_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_html5_media_entries PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_ism_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_m3u8_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_mpd_formats PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_parse_xspf PASSED [ 1%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_response_with_expected_status_returns_content PASSED [ 2%]\ntest/test_InfoExtractor.py::TestInfoExtractor::test_search_json_ld_realworld PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_audio_only_extractor_format_selection PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_default_format_spec PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_filtering PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_not_available PASSED [ 2%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_audio PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_audio_exts PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_issue_10083 PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_string_ops PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_format_selection_video PASSED [ 3%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_invalid_format_specs PASSED [ 4%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_prefer_free_formats PASSED [ 4%]\ntest/test_YoutubeDL.py::TestFormatSelection::test_youtube_format_selection PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_add_extra_info PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_add_headers_cookie PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_do_not_override_ie_key_in_url_transparent PASSED [ 4%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_format_note PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_header_cookies PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_ignoreerrors_for_playlist_with_url_transparent_iterable_entries PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_infojson_cookies PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_match_filter PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_playlist_items_selection PASSED [ 5%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_postprocessors PASSED [ 6%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_prepare_outtmpl_and_filename PASSED [ 6%]\ntest/test_YoutubeDL.py::TestYoutubeDL::test_subtitles PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_get_cookie_header PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_get_cookies_for_url PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_keep_session_cookies PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_malformed_cookies PASSED [ 6%]\ntest/test_YoutubeDLCookieJar.py::TestYoutubeDLCookieJar::test_strip_httponly_prefix PASSED [ 7%]\ntest/test_aes.py::TestAES::test_cbc_decrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_cbc_encrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ctr_decrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ctr_encrypt PASSED [ 7%]\ntest/test_aes.py::TestAES::test_decrypt_text PASSED [ 7%]\ntest/test_aes.py::TestAES::test_ecb_decrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_ecb_encrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_encrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_gcm_decrypt PASSED [ 8%]\ntest/test_aes.py::TestAES::test_key_expansion PASSED [ 8%]\ntest/test_aes.py::TestAES::test_pad_block PASSED [ 8%]\ntest/test_all_urls.py::TestAllURLsMatching::test_facebook_matching PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_keywords PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_no_duplicated_ie_names PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_no_duplicates PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_pbs PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_soundcloud_not_matching_sets PASSED [ 9%]\ntest/test_all_urls.py::TestAllURLsMatching::test_tumblr PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_vimeo_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_channel_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_feeds PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_playlist_matching PASSED [ 10%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_search_matching PASSED [ 11%]\ntest/test_all_urls.py::TestAllURLsMatching::test_youtube_user_matching PASSED [ 11%]\ntest/test_cache.py::TestCache::test_cache PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_etree_fromstring PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_etree_fromstring_doctype PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_expanduser PASSED [ 11%]\ntest/test_compat.py::TestCompat::test_compat_passthrough PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_unquote PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_unquote_plus PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_compat_urllib_parse_urlencode PASSED [ 12%]\ntest/test_compat.py::TestCompat::test_struct_unpack PASSED [ 12%]\ntest/test_config.py::TestConfig::test_config__ENVIRON_DEFAULTS_sanity PASSED [ 12%]\ntest/test_config.py::TestConfig::test_config_all_environ_values PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_default_expected_locations PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_default_grouping PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_early_exit_commandline PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_early_exit_files PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_override_commandline PASSED [ 13%]\ntest/test_config.py::TestConfig::test_config_override_files PASSED [ 13%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_derive_key PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_linux_v11 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_mac_derive_key PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_mac_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_chrome_cookie_decryptor_windows_v10 PASSED [ 14%]\ntest/test_cookies.py::TestCookies::test_get_desktop_environment PASSED [ 15%]\ntest/test_cookies.py::TestCookies::test_pbkdf2_sha1 PASSED [ 15%]\ntest/test_cookies.py::TestCookies::test_safari_cookie_parsing PASSED [ 15%]\ntest/test_cookies.py::TestLenientSimpleCookie::test_lenient_parsing PASSED [ 15%]\ntest/test_cookies.py::TestLenientSimpleCookie::test_parsing PASSED [ 15%]\ntest/test_downloader_external.py::TestHttpieFD::test_make_cmd PASSED [ 15%]\ntest/test_downloader_external.py::TestAxelFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestWgetFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestCurlFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestAria2cFD::test_make_cmd PASSED [ 16%]\ntest/test_downloader_external.py::TestFFmpegFD::test_make_cmd SKIPPED [ 16%]\ntest/test_downloader_http.py::TestHttpFD::test_chunked FAILED [ 16%]\ntest/test_downloader_http.py::TestHttpFD::test_regular FAILED [ 17%]\ntest/test_execution.py::TestExecution::test_cmdline_umlauts PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_import PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_lazy_extractors PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_main_exec PASSED [ 17%]\ntest/test_execution.py::TestExecution::test_module_exec PASSED [ 17%]\ntest/test_jsinterp.py::TestJSInterpreter::test_add PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_array_access PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_assignments PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_basic PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_bitwise_operators_overflow PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_bitwise_operators_typecast PASSED [ 18%]\ntest/test_jsinterp.py::TestJSInterpreter::test_builtins PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_calc PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_call PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_catch PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_char_code_at PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_comma PASSED [ 19%]\ntest/test_jsinterp.py::TestJSInterpreter::test_comments SKIPPED (Not...) [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_date PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_div PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_empty_return PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_exp PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_finally PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop PASSED [ 20%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_break PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_continue PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_for_loop_try PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_if PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_literal_list PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_mod PASSED [ 21%]\ntest/test_jsinterp.py::TestJSInterpreter::test_morespace PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_mul PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_negative PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_nested_try PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_null PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_object PASSED [ 22%]\ntest/test_jsinterp.py::TestJSInterpreter::test_operators PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_packed SKIPPED (Not i...) [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_parens PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_precedence PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_quotes PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_regex PASSED [ 23%]\ntest/test_jsinterp.py::TestJSInterpreter::test_replace SKIPPED (Not ...) [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_return_function PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_strange_chars PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_sub PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_switch PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_switch_default PASSED [ 24%]\ntest/test_jsinterp.py::TestJSInterpreter::test_try PASSED [ 25%]\ntest/test_jsinterp.py::TestJSInterpreter::test_undefined PASSED [ 25%]\ntest/test_jsinterp.py::TestJSInterpreter::test_void PASSED [ 25%]\ntest/test_netrc.py::TestNetRc::test_netrc_present PASSED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_verify_cert[Urllib] PASSED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_verify_cert[Requests] SKIPPED [ 25%]\ntest/test_networking.py::TestHTTPRequestHandler::test_ssl_error[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_ssl_error[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_percent_encode[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_percent_encode[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_remove_dot_segments[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_remove_dot_segments[Requests] SKIPPED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unicode_path_redirection[Urllib] PASSED [ 26%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unicode_path_redirection[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_raise_http_error[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_raise_http_error[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_response_url[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_response_url[Requests] SKIPPED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect[Urllib] PASSED [ 27%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_request_cookie_header[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_request_cookie_header[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect_loop[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_redirect_loop[Requests] SKIPPED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_incompleteread[Urllib] PASSED [ 28%]\ntest/test_networking.py::TestHTTPRequestHandler::test_incompleteread[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_cookies[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_cookies[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_headers[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_headers[Requests] SKIPPED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_timeout[Urllib] PASSED [ 29%]\ntest/test_networking.py::TestHTTPRequestHandler::test_timeout[Requests] SKIPPED [ 30%]\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 691, in process_request_thread\n self.finish_request(request, client_address)\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 361, in finish_request\n self.RequestHandlerClass(request, client_address, self)\ntest/test_networking.py::TestHTTPRequestHandler::test_source_address[Urllib] PASSED [ 30%]\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/socketserver.py\", line 755, in __init__\n self.handle()\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/http/server.py\", line 436, in handle\n self.handle_one_request()\n File \"/opt/hostedtoolcache/Python/3.11.6/x64/lib/python3.11/http/server.py\", line 425, in handle_one_request\n self.wfile.flush() #actually send the response if not already done.\n ^^^^^^^^^^^^^^^^^^\nValueError: I/O operation on closed file.\n----------------------------------------\ntest/test_networking.py::TestHTTPRequestHandler::test_source_address[Requests] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip_trailing_garbage[Urllib] PASSED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip_trailing_garbage[Requests] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_brotli[Urllib] SKIPPED [ 30%]\ntest/test_networking.py::TestHTTPRequestHandler::test_brotli[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_deflate[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_deflate[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_gzip[Requests] SKIPPED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_multiple_encodings[Urllib] PASSED [ 31%]\ntest/test_networking.py::TestHTTPRequestHandler::test_multiple_encodings[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unsupported_encoding[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_unsupported_encoding[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_read[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPRequestHandler::test_read[Requests] SKIPPED [ 32%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy[Urllib] PASSED [ 32%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_noproxy[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_noproxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_allproxy[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_allproxy[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy_with_idn[Urllib] PASSED [ 33%]\ntest/test_networking.py::TestHTTPProxy::test_http_proxy_with_idn[Requests] SKIPPED [ 33%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_nopass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_nopass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_nopass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_nopass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_pass[Urllib] PASSED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_combined_pass[Requests] SKIPPED [ 34%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_pass[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestClientCertificate::test_certificate_nocombined_pass[Requests] SKIPPED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_file_urls[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_http_error_returns_content[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_verify_cert_error_text[Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req0-method can't contain control characters-<lambda>-Urllib] PASSED [ 35%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req1-URL can't contain control characters-<lambda>-Urllib] PASSED [ 36%]\ntest/test_networking.py::TestUrllibRequestHandler::test_httplib_validation_errors[req2-Invalid header name-None-Urllib] PASSED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError0] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError1] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError2] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError3] SKIPPED [ 36%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-ProxyError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-CertificateVerifyError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-SSLError] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError0] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError1] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-TransportError4] SKIPPED [ 37%]\ntest/test_networking.py::TestRequestsRequestHandler::test_request_error_mapping[Requests-<lambda>-RequestError2] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-SSLError-None] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None0] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None1] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None2] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None3] SKIPPED [ 38%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-TransportError-None4] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-IncompleteRead-3 bytes read, 4 more expected] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestsRequestHandler::test_response_error_mapping[Requests-<lambda>-IncompleteRead-3 bytes read, 5 more expected] SKIPPED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-http-False-handler_kwargs0] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-https-False-handler_kwargs1] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-data-False-handler_kwargs2] PASSED [ 39%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-ftp-False-handler_kwargs3] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-file-UnsupportedRequest-handler_kwargs4] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Urllib-file-False-handler_kwargs5] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Requests-http-False-handler_kwargs6] SKIPPED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[Requests-https-False-handler_kwargs7] SKIPPED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[NoCheckRH-http-False-handler_kwargs8] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_url_scheme[ValidationRH-http-UnsupportedRequest-handler_kwargs9] PASSED [ 40%]\ntest/test_networking.py::TestRequestHandlerValidation::test_no_proxy[Urllib-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_no_proxy[Requests-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Urllib-all-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Urllib-unrelated-False] PASSED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Requests-all-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[Requests-unrelated-False] SKIPPED [ 41%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[NoCheckRH-all-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[HTTPSupportedRH-all-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_key[HTTPSupportedRH-no-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-http-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-https-UnsupportedRequest] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks4-False] PASSED [ 42%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks4a-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks5-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks5h-False] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Urllib-socks-UnsupportedRequest] PASSED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-http-False] SKIPPED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-https-False] SKIPPED [ 43%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks4-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks4a-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks5-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[Requests-socks5h-False] SKIPPED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[NoCheckRH-http-False] PASSED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_proxy_scheme[HTTPSupportedRH-http-UnsupportedRequest] PASSED [ 44%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[Urllib] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[HTTPSupportedRH] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_empty_proxy[Requests] SKIPPED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-//example.com] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-example.com] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-127.0.0.1] PASSED [ 45%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Urllib-/a/b/c] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-//example.com] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-example.com] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-127.0.0.1] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_proxy_url[Requests-/a/b/c] SKIPPED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions0-AssertionError] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions1-False] PASSED [ 46%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions2-AssertionError] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions3-False] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions4-AssertionError] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Urllib-extensions5-UnsupportedRequest] PASSED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions6-AssertionError] SKIPPED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions7-False] SKIPPED [ 47%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions8-False] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions9-AssertionError] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[Requests-extensions10-UnsupportedRequest] SKIPPED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[NoCheckRH-extensions11-False] PASSED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_extension[NoCheckRH-extensions12-False] PASSED [ 48%]\ntest/test_networking.py::TestRequestHandlerValidation::test_invalid_request_type PASSED [ 48%]\ntest/test_networking.py::TestRequestDirector::test_handler_operations PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_send PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_unsupported_handlers PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_unexpected_error PASSED [ 49%]\ntest/test_networking.py::TestRequestDirector::test_preference PASSED [ 49%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_opener PASSED [ 49%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[http://127.0.0.1:8080-expected0] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[-expected1] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_proxy[None-expected2] PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_request PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_extract_basic_auth PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_sanitize_url PASSED [ 50%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_file_urls_error PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_legacy_server_connect_error PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-__noproxy__-None] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[no-127.0.0.1,foo.bar-127.0.0.1,foo.bar] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-example.com-http://example.com] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-//example.com-http://example.com] PASSED [ 51%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[https-socks5://example.com-socks5h://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-socks://example.com-socks4://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[http-socks4://example.com-socks4://example.com] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy[unrelated-/bad/proxy-/bad/proxy] PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_proxy_header PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_clean_header PASSED [ 52%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_build_handler_params PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params0] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params1] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params2] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_client_certificate[ydl_params3] PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_urllib_file_urls PASSED [ 53%]\ntest/test_networking.py::TestYoutubeDLNetworking::test_compat_opt_prefer_urllib PASSED [ 53%]\ntest/test_networking.py::TestRequest::test_query PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_method PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_request_helpers PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_headers PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_data_type PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_content_length_header PASSED [ 54%]\ntest/test_networking.py::TestRequest::test_content_type_header PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_update_req PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_proxies PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_extensions PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_copy PASSED [ 55%]\ntest/test_networking.py::TestRequest::test_url PASSED [ 55%]\ntest/test_networking.py::TestResponse::test_reason[custom-200-custom] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[None-404-Not Found] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[-403-Forbidden] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_reason[None-999-None] PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_headers PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_get_header PASSED [ 56%]\ntest/test_networking.py::TestResponse::test_compat PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_select_proxy PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[socks5h://example.com-expected0] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[***example.com:5555-expected1] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[***127.0.0.1:1080-expected2] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_opts[socks4a://:pa%[email protected]] PASSED [ 57%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_make_socks_proxy_unknown PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_load_certifi SKIPPED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[GET-303-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-303-HEAD] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[PUT-303-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-301-GET] PASSED [ 58%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-301-HEAD] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-302-GET] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-302-HEAD] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[PUT-302-PUT] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-308-POST] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[POST-307-POST] PASSED [ 59%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-308-HEAD] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_get_redirect_method[HEAD-307-HEAD] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers0-supported_encodings0-expected0] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers1-supported_encodings1-expected1] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingUtils::test_add_accept_encoding_header[headers2-supported_encodings2-expected2] PASSED [ 60%]\ntest/test_networking_utils.py::TestInstanceStoreMixin::test_mixin PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_http_error[HTTPError] PASSED [ 60%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_http_error[<lambda>] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_redirect_http_error[HTTPError] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_redirect_http_error[<lambda>] PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_compat_http_error PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_compat_http_error_autoclose PASSED [ 61%]\ntest/test_networking_utils.py::TestNetworkingExceptions::test_incomplete_read_error PASSED [ 61%]\ntest/test_plugins.py::TestPlugins::test_directories_containing_plugins PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_extractor_classes PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_importing_zipped_module PASSED [ 62%]\ntest/test_plugins.py::TestPlugins::test_postprocessor_classes PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_field_to_template PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_format_to_regex PASSED [ 62%]\ntest/test_postprocessors.py::TestMetadataFromField::test_metadatafromfield PASSED [ 63%]\ntest/test_postprocessors.py::TestConvertThumbnail::test_escaping PASSED [ 63%]\ntest/test_postprocessors.py::TestExec::test_parse_cmd PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_CommonCase PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_NoZeroDurationChunkAtVideoEnd PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_make_concat_opts_NoZeroDurationChunkAtVideoStart PASSED [ 63%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_QuotesAtEnd PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_QuotesAtStart PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_quote_for_concat_RunsOfQuotes PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CanGetThroughUnaltered PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithAdjacentCuts PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithAdjacentSponsors PASSED [ 64%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithCutHidingSponsor PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithOverlappingCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithOverlappingSponsors PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingCuts PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingSponsors PASSED [ 65%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithRunsOfOverlappingSponsorsAndCuts PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsorCutInTheMiddle PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsors PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChapterWithSponsorsAndCuts PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChaptersAfterLastCut PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_ChaptersAfterLastSponsor PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutCoincidesWithChapters PASSED [ 66%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutEndsAtChapterEnd PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutOverlapsMultipleChapters PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutStartsAtChapterStart PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsAtVideoBoundaries PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsOverlapChaptersAtVideoBoundaries PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_CutsWithinSomeChaptersAndOverlappingOthers PASSED [ 67%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_EverythingCut PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_EverythingSponsored PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_OverlappingSponsorsDifferentTitlesAfterCut PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SingleTinyChapterIsPreserved PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SmallestSponsorInTheOverlapGetsNamed PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorBlockChapters PASSED [ 68%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorCoincidesWithChapters PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorEndsAtChapterEnd PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorOverlapsMultipleChapters PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorStartsAtChapterStart PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsAtVideoBoundaries PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsNoLongerOverlapAfterCut PASSED [ 69%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsOverlapChaptersAtVideoBoundaries PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsStillOverlapAfterCut PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_SponsorsWithinSomeChaptersAndOverlappingOthers PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChapterAtTheStartPrependedToTheNext PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersInTheOriginalArePreserved PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersResultingFromCutsAreIgnored PASSED [ 70%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinyChaptersResultingFromSponsorOverlapAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsOverlapsAreIgnored PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_TinySponsorsPrependedToTheNextSponsor PASSED [ 71%]\ntest/test_postprocessors.py::TestModifyChaptersPP::test_remove_marked_arrange_sponsors_UniqueNamesForOverlappingSponsors PASSED [ 71%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_no_auth[Urllib-http] PASSED [ 71%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_no_auth[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_auth[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_auth[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_ipv4_target[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_ipv4_target[Requests-http] SKIPPED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_domain_target[Urllib-http] PASSED [ 72%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4a_domain_target[Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv4_client_source_address[Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv4_client_source_address[Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[91-Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[91-Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[92-Urllib-http] PASSED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[92-Requests-http] SKIPPED [ 73%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[93-Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_socks4_errors[93-Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv6_socks4_proxy[Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_ipv6_socks4_proxy[Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_timeout[Urllib-http] PASSED [ 74%]\ntest/test_socks.py::TestSocks4Proxy::test_timeout[Requests-http] SKIPPED [ 74%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_no_auth[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_no_auth[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_user_pass[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_user_pass[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv4_target[Urllib-http] PASSED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv4_target[Requests-http] SKIPPED [ 75%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_domain_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_domain_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_domain_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_domain_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_ip_target[Urllib-http] PASSED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5h_ip_target[Requests-http] SKIPPED [ 76%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv6_destination[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_ipv6_destination[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv6_socks5_proxy[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv6_socks5_proxy[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv4_client_source_address[Urllib-http] PASSED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_ipv4_client_source_address[Requests-http] SKIPPED [ 77%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[1-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[1-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[2-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[2-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[3-Urllib-http] PASSED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[3-Requests-http] SKIPPED [ 78%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[4-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[4-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[5-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[5-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[6-Urllib-http] PASSED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[6-Requests-http] SKIPPED [ 79%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[7-Urllib-http] PASSED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[7-Requests-http] SKIPPED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[8-Urllib-http] PASSED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_socks5_errors[8-Requests-http] SKIPPED [ 80%]\ntest/test_socks.py::TestSocks5Proxy::test_timeout[Urllib-http] PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_LazyList PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_LazyList_laziness PASSED [ 80%]\ntest/test_utils.py::TestUtil::test_Popen_windows_escaping SKIPPED (O...) [ 81%]\ntest/test_utils.py::TestUtil::test_age_restricted PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_args_to_str PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_base_url PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_caesar PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_clean_html PASSED [ 81%]\ntest/test_utils.py::TestUtil::test_clean_podcast_url PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_bool_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_cli_valueless_option PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_date_from_str PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_daterange PASSED [ 82%]\ntest/test_utils.py::TestUtil::test_datetime_from_str PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_detect_exe_version PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_determine_ext PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_determine_file_encoding PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_dfxp2srt PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_dict_get PASSED [ 83%]\ntest/test_utils.py::TestUtil::test_encode_base_n PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_encode_compat_str PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_escape_rfc3986 PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_expand_path PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_extract_attributes PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_extract_basic_auth PASSED [ 84%]\ntest/test_utils.py::TestUtil::test_find_xpath_attr PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_fix_xml_ampersands PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_float_or_none PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_format_bytes PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_compatible_ext PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_element_by_attribute PASSED [ 85%]\ntest/test_utils.py::TestUtil::test_get_element_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_html_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_html_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_element_text_and_html_by_tag PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_by_class PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_html_by_attribute PASSED [ 86%]\ntest/test_utils.py::TestUtil::test_get_elements_html_by_class PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_get_elements_text_and_html_by_attribute PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_hide_login_info PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_http_header_dict PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_int_or_none PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_intlist_to_bytes PASSED [ 87%]\ntest/test_utils.py::TestUtil::test_iri_to_uri PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_is_html PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_common_constructors PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_edgecases PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_malformed PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_realworld PASSED [ 88%]\ntest/test_utils.py::TestUtil::test_js_to_json_template_literal PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_js_to_json_vars_strings PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_limit_length PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_locked_file PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_lowercase_escape PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_match_str PASSED [ 89%]\ntest/test_utils.py::TestUtil::test_merge_dicts PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_mimetype2ext PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_month_by_name PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_multipart_encode PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_normalize_url PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_ohdave_rsa_encrypt PASSED [ 90%]\ntest/test_utils.py::TestUtil::test_ordered_set PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_paged_list PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_age_limit PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_bitrate PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_codecs PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_count PASSED [ 91%]\ntest/test_utils.py::TestUtil::test_parse_dfxp_time_expr PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_duration PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_filesize PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_iso8601 PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_parse_resolution PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_pkcs1pad PASSED [ 92%]\ntest/test_utils.py::TestUtil::test_prepend_extension PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_read_batch_urls PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_dot_segments PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_end PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_quotes PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_remove_start PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_render_table PASSED [ 93%]\ntest/test_utils.py::TestUtil::test_replace_extension PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_rot47 PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_filename PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_filename_restricted PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_ids PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_path PASSED [ 94%]\ntest/test_utils.py::TestUtil::test_sanitize_url PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_shell_quote PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_smuggle_url PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_str_to_int PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_strip_jsonp PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_strip_or_none PASSED [ 95%]\ntest/test_utils.py::TestUtil::test_subtitles_filename PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_timeconvert PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_traverse_obj PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_try_call PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unescape_html PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unified_dates PASSED [ 96%]\ntest/test_utils.py::TestUtil::test_unified_timestamps PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_update_url_query PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_uppercase_escape PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_url_basename PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_url_or_none PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_urlencode_postdata PASSED [ 97%]\ntest/test_utils.py::TestUtil::test_urljoin PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_urshift PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_variadic PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_version_tuple PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_attr PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_element PASSED [ 98%]\ntest/test_utils.py::TestUtil::test_xpath_text PASSED [ 99%]\ntest/test_utils.py::TestUtil::test_xpath_with_ns PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_arg PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_eq PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_shortarg PASSED [ 99%]\ntest/test_verbose_output.py::TestVerboseOutput::test_private_info_shortarg_eq PASSED [ 99%]\ntest/test_youtube_misc.py::TestYoutubeMisc::test_youtube_extract PASSED [100%]\n\n=================================== FAILURES ===================================\n___________________________ TestHttpFD.test_chunked ____________________________\n\nself = <test.test_downloader_http.TestHttpFD testMethod=test_chunked>\n\n def test_chunked(self):\n> self.download_all({\n 'http_chunk_size': 1000,\n })\n\ntest/test_downloader_http.py:100: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \ntest/test_downloader_http.py:94: in download_all\n self.download(params, ep)\ntest/test_downloader_http.py:86: in download\n self.assertTrue(downloader.real_download(filename, {\nyt_dlp/downloader/http.py:365: in real_download\n return download()\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\n def download():\n data_len = ctx.data.headers.get('Content-length')\n \n if ctx.data.headers.get('Content-encoding'):\n # Content-encoding is present, Content-length is not reliable anymore as we are\n # doing auto decompression. (See: https://github.com/yt-dlp/yt-dlp/pull/6176)\n data_len = None\n \n # Range HTTP header may be ignored/unsupported by a webserver\n # (e.g. extractor/scivee.py, extractor/bambuser.py).\n # However, for a test we still would like to download just a piece of a file.\n # To achieve this we limit data_len to _TEST_FILE_SIZE and manually control\n # block size when downloading a file.\n if is_test and (data_len is None or int(data_len) > self._TEST_FILE_SIZE):\n data_len = self._TEST_FILE_SIZE\n \n if data_len is not None:\n data_len = int(data_len) + ctx.resume_len\n min_data_len = self.params.get('min_filesize')\n max_data_len = self.params.get('max_filesize')\n if min_data_len is not None and data_len < min_data_len:\n self.to_screen(\n f'\\r[download] File is smaller than min-filesize ({data_len} bytes < {min_data_len} bytes). Aborting.')\n return False\n if max_data_len is not None and data_len > max_data_len:\n self.to_screen(\n f'\\r[download] File is larger than max-filesize ({data_len} bytes > {max_data_len} bytes). Aborting.')\n return False\n \n byte_counter = 0 + ctx.resume_len\n block_size = ctx.block_size\n start = time.time()\n \n # measure time over whole while-loop, so slow_down() and best_block_size() work together properly\n now = None # needed for slow_down() in the first loop run\n before = start # start measuring\n \n def retry(e):\n close_stream()\n ctx.resume_len = (byte_counter if ctx.tmpfilename == '-'\n else os.path.getsize(encodeFilename(ctx.tmpfilename)))\n raise RetryDownload(e)\n \n while True:\n try:\n # Download and write\n data_block = ctx.data.read(block_size if not is_test else min(block_size, data_len - byte_counter))\n except TransportError as err:\n retry(err)\n \n byte_counter += len(data_block)\n \n # exit loop when download is finished\n if len(data_block) == 0:\n break\n \n # Open destination file just in time\n if ctx.stream is None:\n try:\n ctx.stream, ctx.tmpfilename = self.sanitize_open(\n ctx.tmpfilename, ctx.open_mode)\n assert ctx.stream is not None\n ctx.filename = self.undo_temp_name(ctx.tmpfilename)\n self.report_destination(ctx.filename)\n except OSError as err:\n self.report_error('unable to open for writing: %s' % str(err))\n return False\n \n if self.params.get('xattr_set_filesize', False) and data_len is not None:\n try:\n write_xattr(ctx.tmpfilename, 'user.ytdl.filesize', str(data_len).encode())\n except (XAttrUnavailableError, XAttrMetadataError) as err:\n self.report_error('unable to set filesize xattr: %s' % str(err))\n \n try:\n ctx.stream.write(data_block)\n except OSError as err:\n self.to_stderr('\\n')\n self.report_error('unable to write data: %s' % str(err))\n return False\n \n # Apply rate limit\n self.slow_down(start, now, byte_counter - ctx.resume_len)\n \n # end measuring of one loop run\n now = time.time()\n after = now\n \n # Adjust block size\n if not self.params.get('noresizebuffer', False):\n block_size = self.best_block_size(after - before, len(data_block))\n \n before = after\n \n # Progress message\n> speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\nE AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n\nyt_dlp/downloader/http.py:296: AttributeError\n___________________________ TestHttpFD.test_regular ____________________________\n\nself = <test.test_downloader_http.TestHttpFD testMethod=test_regular>\n\n def test_regular(self):\n> self.download_all({})\n\ntest/test_downloader_http.py:97: \n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \ntest/test_downloader_http.py:94: in download_all\n self.download(params, ep)\ntest/test_downloader_http.py:86: in download\n self.assertTrue(downloader.real_download(filename, {\nyt_dlp/downloader/http.py:365: in real_download\n return download()\n_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ \n\n def download():\n data_len = ctx.data.headers.get('Content-length')\n \n if ctx.data.headers.get('Content-encoding'):\n # Content-encoding is present, Content-length is not reliable anymore as we are\n # doing auto decompression. (See: https://github.com/yt-dlp/yt-dlp/pull/6176)\n data_len = None\n \n # Range HTTP header may be ignored/unsupported by a webserver\n # (e.g. extractor/scivee.py, extractor/bambuser.py).\n # However, for a test we still would like to download just a piece of a file.\n # To achieve this we limit data_len to _TEST_FILE_SIZE and manually control\n # block size when downloading a file.\n if is_test and (data_len is None or int(data_len) > self._TEST_FILE_SIZE):\n data_len = self._TEST_FILE_SIZE\n \n if data_len is not None:\n data_len = int(data_len) + ctx.resume_len\n min_data_len = self.params.get('min_filesize')\n max_data_len = self.params.get('max_filesize')\n if min_data_len is not None and data_len < min_data_len:\n self.to_screen(\n f'\\r[download] File is smaller than min-filesize ({data_len} bytes < {min_data_len} bytes). Aborting.')\n return False\n if max_data_len is not None and data_len > max_data_len:\n self.to_screen(\n f'\\r[download] File is larger than max-filesize ({data_len} bytes > {max_data_len} bytes). Aborting.')\n return False\n \n byte_counter = 0 + ctx.resume_len\n block_size = ctx.block_size\n start = time.time()\n \n # measure time over whole while-loop, so slow_down() and best_block_size() work together properly\n now = None # needed for slow_down() in the first loop run\n before = start # start measuring\n \n def retry(e):\n close_stream()\n ctx.resume_len = (byte_counter if ctx.tmpfilename == '-'\n else os.path.getsize(encodeFilename(ctx.tmpfilename)))\n raise RetryDownload(e)\n \n while True:\n try:\n # Download and write\n data_block = ctx.data.read(block_size if not is_test else min(block_size, data_len - byte_counter))\n except TransportError as err:\n retry(err)\n \n byte_counter += len(data_block)\n \n # exit loop when download is finished\n if len(data_block) == 0:\n break\n \n # Open destination file just in time\n if ctx.stream is None:\n try:\n ctx.stream, ctx.tmpfilename = self.sanitize_open(\n ctx.tmpfilename, ctx.open_mode)\n assert ctx.stream is not None\n ctx.filename = self.undo_temp_name(ctx.tmpfilename)\n self.report_destination(ctx.filename)\n except OSError as err:\n self.report_error('unable to open for writing: %s' % str(err))\n return False\n \n if self.params.get('xattr_set_filesize', False) and data_len is not None:\n try:\n write_xattr(ctx.tmpfilename, 'user.ytdl.filesize', str(data_len).encode())\n except (XAttrUnavailableError, XAttrMetadataError) as err:\n self.report_error('unable to set filesize xattr: %s' % str(err))\n \n try:\n ctx.stream.write(data_block)\n except OSError as err:\n self.to_stderr('\\n')\n self.report_error('unable to write data: %s' % str(err))\n return False\n \n # Apply rate limit\n self.slow_down(start, now, byte_counter - ctx.resume_len)\n \n # end measuring of one loop run\n now = time.time()\n after = now\n \n # Adjust block size\n if not self.params.get('noresizebuffer', False):\n block_size = self.best_block_size(after - before, len(data_block))\n \n before = after\n \n # Progress message\n> speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\nE AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n\nyt_dlp/downloader/http.py:296: AttributeError\n=========================== short test summary info ============================\nSKIPPED [1] test/test_downloader_external.py:115: ffmpeg not found\nSKIPPED [1] test/test_jsinterp.py:111: Not implemented\nSKIPPED [1] test/test_jsinterp.py:373: Not implemented\nSKIPPED [1] test/test_jsinterp.py:335: Not implemented\nSKIPPED [1] test/test_networking.py:309: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:320: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:338: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:350: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:365: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:372: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:382: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:393: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:448: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:480: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:486: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:492: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:509: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:525: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:541: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:549: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:555: brotli support is not installed\nSKIPPED [1] test/test_networking.py:566: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:576: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:586: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:597: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:607: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:637: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:663: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:673: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:681: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:719: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:725: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:732: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:739: Requests request handler is not available\nSKIPPED [11] test/test_networking.py:824: Requests request handler is not available\nSKIPPED [8] test/test_networking.py:855: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:996: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:1005: Requests request handler is not available\nSKIPPED [2] test/test_networking.py:1010: Requests request handler is not available\nSKIPPED [6] test/test_networking.py:1019: Requests request handler is not available\nSKIPPED [1] test/test_networking.py:1028: Requests request handler is not available\nSKIPPED [4] test/test_networking.py:1033: Requests request handler is not available\nSKIPPED [5] test/test_networking.py:1038: Requests request handler is not available\nSKIPPED [1] test/test_networking_utils.py:96: certifi is not installed\nSKIPPED [1] test/test_socks.py:266: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:274: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:284: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:292: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:301: Requests request handler is not available\nSKIPPED [3] test/test_socks.py:311: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:323: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:332: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:342: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:350: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:363: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:371: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:379: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:388: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:397: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:405: Requests request handler is not available\nSKIPPED [1] test/test_socks.py:416: Requests request handler is not available\nSKIPPED [8] test/test_socks.py:425: Requests request handler is not available\nSKIPPED [1] test/test_utils.py:2398: Only relevant on Windows\nFAILED test/test_downloader_http.py::TestHttpFD::test_chunked - AttributeError: 'HttpFD' object has no attribute 'calc_speed'\nFAILED test/test_downloader_http.py::TestHttpFD::test_regular - AttributeError: 'HttpFD' object has no attribute 'calc_speed'\n=== 2 failed, 509 passed, 104 skipped, 5705 deselected in 120.03s (0:02:00) ====\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/yt_dlp/downloader/http.py b/yt_dlp/downloader/http.py\nindex f5237443e..a1f95db29 100644\n--- a/yt_dlp/downloader/http.py\n+++ b/yt_dlp/downloader/http.py\n@@ -11,6 +11,7 @@\n )\n from ..utils import (\n ContentTooShortError,\n+ FormatProgressInfos,\n RetryManager,\n ThrottledDownload,\n XAttrMetadataError,\n@@ -293,11 +294,11 @@ def retry(e):\n before = after\n \n # Progress message\n- speed = self.calc_speed(start, now, byte_counter - ctx.resume_len)\n+ speed = FormatProgressInfos.calc_speed(start, now, byte_counter - ctx.resume_len)\n if ctx.data_len is None:\n eta = None\n else:\n- eta = self.calc_eta(start, time.time(), ctx.data_len - ctx.resume_len, byte_counter - ctx.resume_len)\n+ eta = FormatProgressInfos.calc_eta(start, time.time(), ctx.data_len - ctx.resume_len, byte_counter - ctx.resume_len)\n \n self._hook_progress({\n 'status': 'downloading',\ndiff --git a/yt_dlp/downloader/rtmp.py b/yt_dlp/downloader/rtmp.py\nindex 0e0952599..9373f3585 100644\n--- a/yt_dlp/downloader/rtmp.py\n+++ b/yt_dlp/downloader/rtmp.py\n@@ -5,6 +5,7 @@\n \n from .common import FileDownloader\n from ..utils import (\n+ FormatProgressInfos,\n Popen,\n check_executable,\n encodeArgument,\n@@ -50,8 +51,8 @@ def run_rtmpdump(args):\n resume_percent = percent\n resume_downloaded_data_len = downloaded_data_len\n time_now = time.time()\n- eta = self.calc_eta(start, time_now, 100 - resume_percent, percent - resume_percent)\n- speed = self.calc_speed(start, time_now, downloaded_data_len - resume_downloaded_data_len)\n+ eta = FormatProgressInfos.calc_eta(start, time_now, 100 - resume_percent, percent - resume_percent)\n+ speed = FormatProgressInfos.calc_speed(start, time_now, downloaded_data_len - resume_downloaded_data_len)\n data_len = None\n if percent > 0:\n data_len = int(downloaded_data_len * 100 / percent)\n", "difficulty": 2, "changed_files": ["yt_dlp/downloader/http.py", "yt_dlp/downloader/rtmp.py"], "commit_link": "https://github.com/yt-dlp/yt-dlp/tree/43dd59c3137df77f5dd22cef4cb7bedfe9f6b12e"}
|
|
|
|
data/python/4410203.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 96, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "error_logging_callback", "workflow_name": "Smoketest", "workflow_filename": "smoketest.yaml", "workflow_path": ".github/workflows/smoketest.yaml", "contributor": "bmosaicml", "sha_fail": "4410203c56984c613d23f29a81ecd1b96c57b1ee", "sha_success": "e4eb7eeac2e76c3287bdd650407e4067ad3a7625", "workflow": "name: Smoketest\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n smoketest:\n runs-on: ubuntu-20.04\n timeout-minutes: 10\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .\n python -m pip install pytest==7.2.1 pytest_codeblocks==0.16.1\n - name: Run checks\n run: |\n pytest tests/test_smoketest.py\n", "logs": [{"step_name": "smoketest (3.8)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}, {"step_name": "smoketest (3.9)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}, {"step_name": "smoketest (3.10)/5_Run checks.txt", "log": "##[group]Run pytest tests/test_smoketest.py\n\u001b[36;1mpytest tests/test_smoketest.py\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nImportError while loading conftest '/home/runner/work/composer/composer/tests/conftest.py'.\ntests/conftest.py:9: in <module>\n from composer.utils import reproducibility\ncomposer/__init__.py:10: in <module>\n from composer.trainer import Trainer\ncomposer/trainer/__init__.py:6: in <module>\n from composer.trainer.trainer import Trainer\ncomposer/trainer/trainer.py:37: in <module>\n from composer.callbacks import CheckpointSaver, OptimizerMonitor\ncomposer/callbacks/__init__.py:12: in <module>\n from composer.callbacks.eval_output_logging_callback import EvalOutputLogging\ncomposer/callbacks/eval_output_logging_callback.py:16: in <module>\n from composer.datasets.in_context_learning_evaluation import (InContextLearningCodeEvalDataset,\ncomposer/datasets/in_context_learning_evaluation.py:13: in <module>\n import transformers\nE ModuleNotFoundError: No module named 'transformers'\n##[error]Process completed with exit code 4.\n"}], "diff": "diff --git a/composer/callbacks/eval_output_logging_callback.py b/composer/callbacks/eval_output_logging_callback.py\nindex 48e4f76d..65cbdebb 100644\n--- a/composer/callbacks/eval_output_logging_callback.py\n+++ b/composer/callbacks/eval_output_logging_callback.py\n@@ -7,7 +7,7 @@ import hashlib\n import os\n import random\n import shutil\n-import time\n+import tempfile\n from typing import Callable, Optional\n \n from torch.utils.data import DataLoader\n@@ -58,6 +58,8 @@ class EvalOutputLogging(Callback):\n self.hash = hashlib.sha256()\n self.destination_file = None\n \n+ # with tempfile.NamedTemporaryFile\n+ # tmp_dir =\n def _write_tables_to_output_dir(self, state: State):\n try:\n import pandas as pd\n@@ -66,16 +68,9 @@ class EvalOutputLogging(Callback):\n conda_package='pandas',\n conda_channel='conda-forge') from e\n # write tmp files\n- self.hash.update((str(time.time()) + str(random.randint(0, 1_000_000))).encode('utf-8'))\n- tmp_dir = os.getcwd() + '/' + self.hash.hexdigest()\n-\n- if not os.path.exists(tmp_dir):\n- with dist.local_rank_zero_download_and_wait(tmp_dir):\n- if dist.get_local_rank() == 0:\n- os.mkdir(tmp_dir)\n \n full_df = pd.DataFrame()\n- file_name = f'eval-outputs-ba{state.timestamp.batch.value}.tsv'\n+ upload_file_name = f'eval-outputs-ba{state.timestamp.batch.value}.tsv'\n \n for benchmark in self.table:\n cols, rows = self.table[benchmark]\n@@ -84,18 +79,14 @@ class EvalOutputLogging(Callback):\n df['benchmark'] = benchmark\n full_df = pd.concat([full_df, df], ignore_index=True)\n \n- with dist.local_rank_zero_download_and_wait(f'{tmp_dir}/{file_name}'):\n- if dist.get_local_rank() == 0:\n- with open(f'{tmp_dir}/{file_name}', 'wb') as f:\n- full_df.to_csv(f, sep='\\t', index=False)\n+ tmp_file = ''\n+ with tempfile.NamedTemporaryFile('wb') as f:\n+ full_df.to_csv(f, sep='\\t', index=False)\n+ tmp_file = f.name\n \n # copy/upload tmp files\n- _write(destination_path=f'{self.output_directory}/{file_name}', src_file=f'{tmp_dir}/{file_name}')\n- os.remove(f'{tmp_dir}/{file_name}')\n- self.destination_file = f'{self.output_directory}/{file_name}'\n-\n- # delete tmp files\n- os.rmdir(tmp_dir)\n+ _write(destination_path=f'{self.output_directory}/{upload_file_name}', src_file=tmp_file)\n+ self.destination_file = f'{self.output_directory}/{upload_file_name}'\n \n def _prep_response_cache(self, state, cache):\n benchmark = state.dataloader_label\ndiff --git a/composer/datasets/in_context_learning_evaluation.py b/composer/datasets/in_context_learning_evaluation.py\nindex e1566509..1a53f75c 100644\n--- a/composer/datasets/in_context_learning_evaluation.py\n+++ b/composer/datasets/in_context_learning_evaluation.py\n@@ -10,7 +10,6 @@ import random\n from typing import TYPE_CHECKING, Any, Dict, List, Optional, Union\n \n import torch\n-import transformers\n from torch.utils.data import DataLoader, Dataset\n from tqdm import tqdm\n \n", "difficulty": 1, "changed_files": ["composer/callbacks/eval_output_logging_callback.py", "composer/datasets/in_context_learning_evaluation.py"], "commit_link": "https://github.com/mosaicml/composer/tree/4410203c56984c613d23f29a81ecd1b96c57b1ee"}
|
|
|
|
data/python/44b56e0.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 127, "repo_owner": "huggingface", "repo_name": "accelerate", "head_branch": "main", "workflow_name": "Quality Check", "workflow_filename": "quality.yml", "workflow_path": ".github/workflows/quality.yml", "contributor": "PhilJd", "sha_fail": "44b56e01683771fb4ca583f9ea57c67dcee8e779", "sha_success": "90bb882770fafc8db2fabc7fe58f446e12b55d77", "workflow": "name: Quality Check\n\non: [pull_request]\n\njobs:\n quality:\n runs-on: ubuntu-latest\n steps:\n - uses: actions/[email protected]\n - name: Set up Python 3.8\n uses: actions/setup-python@v3\n with:\n python-version: 3.8\n - name: Install Python dependencies\n run: pip install -e .[quality]\n - name: Run Quality check\n run: make quality\n - name: Check if failure\n if: ${{ failure() }}\n run: |\n echo \"Quality check failed. Please ensure the right dependency versions are installed with 'pip install -e .[quality]' and rerun 'make style; make quality;'\" >> $GITHUB_STEP_SUMMARY\n\n", "logs": [{"step_name": "quality/5_Run Quality check.txt", "log": "##[group]Run make quality\n\u001b[36;1mmake quality\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nblack --required-version 23 --check tests src examples benchmarks utils\nAll done! \u2728 \ud83c\udf70 \u2728\n118 files would be left unchanged.\nruff tests src examples benchmarks utils\ndoc-builder style src/accelerate docs/source --max_len 119 --check_only\nTraceback (most recent call last):\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/bin/doc-builder\", line 8, in <module>\n sys.exit(main())\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/doc_builder/commands/doc_builder_cli.py\", line 47, in main\n args.func(args)\n File \"/opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/doc_builder/commands/style.py\", line 28, in style_command\n raise ValueError(f\"{len(changed)} files should be restyled!\")\nValueError: 1 files should be restyled!\nmake: *** [Makefile:17: quality] Error 1\n##[error]Process completed with exit code 2.\n"}], "diff": "diff --git a/src/accelerate/big_modeling.py b/src/accelerate/big_modeling.py\nindex 5fd2bd5..6a2392a 100644\n--- a/src/accelerate/big_modeling.py\n+++ b/src/accelerate/big_modeling.py\n@@ -73,7 +73,9 @@ def init_empty_weights(include_buffers: bool = None):\n \n Any model created under this context manager has no weights. As such you can't do something like\n `model.to(some_device)` with it. To load weights inside your empty model, see [`load_checkpoint_and_dispatch`].\n- Make sure to overwrite the default device_map param, otherwise dispatch is not called.\n+ Make sure to overwrite the default device_map param for [`load_checkpoint_and_dispatch`], otherwise dispatch is not\n+ called.\n+\n </Tip>\n \"\"\"\n if include_buffers is None:\n@@ -480,7 +482,7 @@ def load_checkpoint_and_dispatch(\n \n To have Accelerate compute the most optimized `device_map` automatically, set `device_map=\"auto\"`. For more\n information about each option see [here](../concept_guides/big_model_inference#designing-a-device-map).\n- Defaults to None, which means `dispatch_model` will not be called.\n+ Defaults to None, which means [`dispatch_model`] will not be called.\n max_memory (`Dict`, *optional*):\n A dictionary device identifier to maximum memory. Will default to the maximum memory available for each GPU\n and the available CPU RAM if unset.\n", "difficulty": 0, "changed_files": ["src/accelerate/big_modeling.py"], "commit_link": "https://github.com/huggingface/accelerate/tree/44b56e01683771fb4ca583f9ea57c67dcee8e779"}
|
|
|
|
data/python/4541644.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 41, "repo_owner": "beetbox", "repo_name": "beets", "head_branch": "lb", "workflow_name": "Check code formatting", "workflow_filename": "formatting_check.yml", "workflow_path": ".github/workflows/formatting_check.yml", "contributor": "arsaboo", "sha_fail": "454164496177fd8b9d6aad4f106e68e816becb6c", "sha_success": "0ed65564b30e002d7f5884e32132098542c3c7dd", "workflow": "name: formatting_check\nrun-name: Check code formatting\non: [push, pull_request]\njobs:\n formatting_check:\n runs-on: ubuntu-latest\n steps:\n - name: Install dependencies\n uses: actions/checkout@v3\n\n - name: Run formatting check\n uses: paolorechia/[email protected]\n with:\n tox_env: \"format_check\"\n", "logs": [{"step_name": "formatting_check/3_Run formatting check.txt", "log": "##[group]Run paolorechia/[email protected]\nwith:\n tox_env: format_check\n##[endgroup]\nstdout: Defaulting to user installation because normal site-packages is not writeable\n\nstdout: Collecting tox\n\nstdout: Downloading tox-4.11.4-py3-none-any.whl (153 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 154.0/154.0 KB 5.2 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting pyproject-api>=1.6.1\n\nstdout: Downloading pyproject_api-1.6.1-py3-none-any.whl (12 kB)\n\nstdout: Collecting colorama>=0.4.6\n\nstdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB)\n\nstdout: Collecting filelock>=3.12.3\n\nstdout: Downloading filelock-3.13.1-py3-none-any.whl (11 kB)\n\nstdout: Requirement already satisfied: tomli>=2.0.1 in /usr/local/lib/python3.10/dist-packages (from tox) (2.0.1)\n\nstdout: Collecting pluggy>=1.3\n\nstdout: Downloading pluggy-1.3.0-py3-none-any.whl (18 kB)\n\nstdout: Collecting chardet>=5.2\n\nstdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 199.4/199.4 KB 9.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Requirement already satisfied: packaging>=23.1 in /usr/local/lib/python3.10/dist-packages (from tox) (23.2)\n\nstdout: Collecting virtualenv>=20.24.3\n\nstdout: Downloading virtualenv-20.25.0-py3-none-any.whl (3.8 MB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 3.8/3.8 MB 18.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Collecting cachetools>=5.3.1\n\nstdout: Downloading cachetools-5.3.2-py3-none-any.whl (9.3 kB)\n\nstdout: Requirement already satisfied: platformdirs>=3.10 in /usr/local/lib/python3.10/dist-packages (from tox) (4.1.0)\n\nstdout: Collecting distlib<1,>=0.3.7\n\nstdout: Downloading distlib-0.3.8-py2.py3-none-any.whl (468 kB)\n\nstdout: \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 468.9/468.9 KB 30.1 MB/s eta 0:00:00\nstdout: \n\nstdout: Installing collected packages: distlib, pyproject-api, pluggy, filelock, colorama, chardet, cachetools, virtualenv, tox\n\nstdout: Successfully installed cachetools-5.3.2 chardet-5.2.0 colorama-0.4.6 distlib-0.3.8 filelock-3.13.1 pluggy-1.3.0 pyproject-api-1.6.1 tox-4.11.4 virtualenv-20.25.0\n\nReceived 6565789 of 6565789 (100.0%), 7.4 MBs/sec\nCache Size: ~6 MB (6565789 B)\n[command]/usr/bin/tar -z -xf /home/runner/work/_temp/2ecbc63f-4679-45d0-8b2e-0acac9db60db/cache.tgz -P -C /home/runner/work/beets/beets\nCache restored successfully\nHit cache on tox\nstdout: format_check: recreate env because python changed virtualenv version='20.24.5'->'20.25.0'\n\nstdout: format_check: remove tox env folder /home/runner/work/beets/beets/.tox/format_check\n\nstdout: format_check: install_deps> python -I -m pip install black isort\n\nstdout: format_check: commands[0]> isort beets beetsplug test --check\n\nstderr: ERROR: /home/runner/work/beets/beets/beetsplug/listenbrainz.py Imports are incorrectly sorted and/or formatted.\n\nstdout: format_check: exit 1 (0.57 seconds) /home/runner/work/beets/beets> isort beets beetsplug test --check pid=1921\n\nstdout: format_check: FAIL code 1 (4.17=setup[3.60]+cmd[0.57] seconds)\n evaluation failed :( (4.22 seconds)\n\nchild process exited with code 1\n##[warning]The `set-output` command is deprecated and will be disabled soon. Please upgrade to using Environment Files. For more information see: https://github.blog/changelog/2022-10-11-github-actions-deprecating-save-state-and-set-output-commands/\n##[error]Tox failed\n"}], "diff": "diff --git a/beetsplug/listenbrainz.py b/beetsplug/listenbrainz.py\nindex bb5020902..a5595642f 100644\n--- a/beetsplug/listenbrainz.py\n+++ b/beetsplug/listenbrainz.py\n@@ -4,13 +4,15 @@ import datetime\n \n import musicbrainzngs\n import requests\n+\n from beets import config, ui\n from beets.plugins import BeetsPlugin\n from beetsplug.lastimport import process_tracks\n \n \n class ListenBrainzPlugin(BeetsPlugin):\n- \"\"\" A Beets plugin for interacting with ListenBrainz.\"\"\"\n+ \"\"\"A Beets plugin for interacting with ListenBrainz.\"\"\"\n+\n data_source = \"ListenBrainz\"\n ROOT = \"http://api.listenbrainz.org/1/\"\n \n@@ -101,7 +103,6 @@ class ListenBrainzPlugin(BeetsPlugin):\n else:\n return None\n \n-\n def get_tracks_from_listens(self, listens):\n \"\"\"Returns a list of tracks from a list of listens.\"\"\"\n tracks = []\n", "difficulty": 0, "changed_files": ["beetsplug/listenbrainz.py"], "commit_link": "https://github.com/beetbox/beets/tree/454164496177fd8b9d6aad4f106e68e816becb6c"}
|
|
|
|
data/python/483a1db.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 139, "repo_owner": "burnash", "repo_name": "gspread", "head_branch": "simplify-get-records", "workflow_name": "lint_python", "workflow_filename": "main.yaml", "workflow_path": ".github/workflows/main.yaml", "contributor": "burnash", "sha_fail": "483a1db177fd4c33b68d122640c32034d5dbb8ae", "sha_success": "74980b8fba2d73c8446a2664103c52a9c914a4ca", "workflow": "name: lint_python\non:\n push:\n branches:\n - \"**\" # run all branches\n tags-ignore:\n - \"*\" # ignore all tags, release.yaml will trigger the CI\n pull_request: # run on all pull requests\n\nconcurrency:\n cancel-in-progress: true\n group: group-${{ github.ref_name }}\n\njobs:\n lint_python:\n runs-on: ubuntu-latest\n strategy:\n fail-fast: false\n matrix:\n python: [\"3.8\", \"3.9\", \"3.10\", \"3.11\", \"3.x\"]\n steps:\n - uses: actions/checkout@v4\n - uses: actions/setup-python@v5\n with:\n python-version: ${{ matrix.python }}\n - run: pip install -U pip\n - run: pip install -U bandit mypy pyupgrade safety tox setuptools\n - run: bandit --recursive --skip B105,B110,B311,B605,B607 --exclude ./.tox .\n if: ${{ matrix.python >= '3.8' }}\n - run: tox -e lint\n - run: tox -e py\n - run: shopt -s globstar && pyupgrade --py3-only **/*.py # --py36-plus\n - run: safety check -i 42559 -i 62044 # pip <= 20.1.1, we upgrade it just above\n - run: tox -e build\n - run: tox -e doc\n", "logs": [{"step_name": "lint_python (3.8)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:11.002645 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:38.759397 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.18 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1934\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.8.18/x64/lib/python3.8/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (11.96=setup[10.78]+cmd[1.18] seconds)\n evaluation failed :( (12.11 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.9)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.9.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.9.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.9.18/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:12.019448 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:43.591179 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.31 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1903\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.9.18/x64/lib/python3.9/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (12.96=setup[11.66]+cmd[1.31] seconds)\n evaluation failed :( (13.20 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.10)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.10.13/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.10.13/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.10.13/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:10.826736 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:40.023159 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.43 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1907\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.10.13/x64/lib/python3.10/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (13.37=setup[11.94]+cmd[1.43] seconds)\n evaluation failed :( (13.62 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.11)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.11.7/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.11.7/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.11.7/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:12.077791 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:39.355057 +0000\n@@ -517,13 +517,14 @@\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (1.19 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1935\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.11.7/x64/lib/python3.11/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (12.92=setup[11.72]+cmd[1.19] seconds)\n evaluation failed :( (12.98 seconds)\n##[error]Process completed with exit code 1.\n"}, {"step_name": "lint_python (3.x)/7_Run tox -e lint.txt", "log": "##[group]Run tox -e lint\n\u001b[36;1mtox -e lint\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.12.1/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.12.1/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.12.1/x64/lib\n##[endgroup]\nlint: install_deps> python -I -m pip install -r lint-requirements.txt\n.pkg: install_requires> python -I -m pip install 'flit_core<4,>=3.2'\n.pkg: _optional_hooks> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: get_requires_for_build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n.pkg: build_sdist> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\nlint: install_package_deps> python -I -m pip install StrEnum==0.4.15 'google-auth-oauthlib>=0.4.1' 'google-auth>=1.12.0'\nlint: install_package> python -I -m pip install --force-reinstall --no-deps /home/runner/work/gspread/gspread/.tox/.pkg/dist/gspread-6.0.0.tar.gz\nlint: commands[0]> black --check --diff --extend-exclude ./env .\n--- /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:11.804783 +0000\n+++ /home/runner/work/gspread/gspread/gspread/worksheet.py\t2023-12-29 18:49:49.027795 +0000\n@@ -517,13 +517,14 @@\n \n if expected_headers is None:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n set(expected_headers)\n )\nwould reformat /home/runner/work/gspread/gspread/gspread/worksheet.py\n\nOh no! \ud83d\udca5 \ud83d\udc94 \ud83d\udca5\n1 file would be reformatted, 17 files would be left unchanged.\nlint: exit 1 (2.35 seconds) /home/runner/work/gspread/gspread> black --check --diff --extend-exclude ./env . pid=1932\n.pkg: _exit> python /opt/hostedtoolcache/Python/3.12.1/x64/lib/python3.12/site-packages/pyproject_api/_backend.py True flit_core.buildapi\n lint: FAIL code 1 (15.07=setup[12.72]+cmd[2.35] seconds)\n evaluation failed :( (15.37 seconds)\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/gspread/worksheet.py b/gspread/worksheet.py\nindex d6268ea..8b4f551 100644\n--- a/gspread/worksheet.py\n+++ b/gspread/worksheet.py\n@@ -519,9 +519,10 @@ class Worksheet:\n # all headers must be unique\n header_row_is_unique = len(keys) == len(set(keys))\n if not header_row_is_unique:\n- raise GSpreadException(\"the header row in the worksheet is not unique\"\n- \"try passing 'expected_headers' to get_all_records\"\n- )\n+ raise GSpreadException(\n+ \"the header row in the worksheet is not unique\"\n+ \"try passing 'expected_headers' to get_all_records\"\n+ )\n else:\n # all expected headers must be unique\n expected_headers_are_unique = len(expected_headers) == len(\n", "difficulty": 0, "changed_files": ["gspread/worksheet.py"], "commit_link": "https://github.com/burnash/gspread/tree/483a1db177fd4c33b68d122640c32034d5dbb8ae"}
|
|
|
|
data/python/4921af9.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
data/python/4b7e472.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 94, "repo_owner": "mosaicml", "repo_name": "composer", "head_branch": "composer_lora", "workflow_name": "Code Quality Checks", "workflow_filename": "code-quality.yaml", "workflow_path": ".github/workflows/code-quality.yaml", "contributor": "dakinggg", "sha_fail": "4b7e4724b2f92741fb094421ebfa60be5bfb18d9", "sha_success": "09346e7f3367ae17409d2a153ec02db1869e09e0", "workflow": "name: Code Quality Checks\non:\n push:\n branches:\n - dev\n - main\n - release/**\n pull_request:\n workflow_call:\n workflow_dispatch:\n# Cancel old runs when a new commit is pushed to the same branch if not on main or dev\nconcurrency:\n group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}\n cancel-in-progress: ${{ github.ref != 'refs/heads/main' && github.ref != 'refs/heads/dev' }}\ndefaults:\n run:\n working-directory: .\njobs:\n code-quality:\n runs-on: ubuntu-20.04\n timeout-minutes: 10\n strategy:\n matrix:\n python_version:\n - \"3.8\"\n - \"3.9\"\n - \"3.10\"\n pip_deps:\n - \"[dev]\"\n steps:\n - uses: actions/checkout@v3\n - uses: actions/setup-python@v4\n with:\n python-version: ${{ matrix.python_version }}\n - name: Setup\n run: |\n set -ex\n python -m pip install --upgrade 'pip<23' wheel\n python -m pip install --upgrade .${{ matrix.pip_deps }}\n - name: Run checks\n run: |\n pre-commit run --all-files\n", "logs": [{"step_name": "code-quality (3.8, [dev])/5_Run checks.txt", "log": "##[group]Run pre-commit run --all-files\n\u001b[36;1mpre-commit run --all-files\u001b[0m\nshell: /usr/bin/bash -e {0}\nenv:\n pythonLocation: /opt/hostedtoolcache/Python/3.8.18/x64\n PKG_CONFIG_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib/pkgconfig\n Python_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python2_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n Python3_ROOT_DIR: /opt/hostedtoolcache/Python/3.8.18/x64\n LD_LIBRARY_PATH: /opt/hostedtoolcache/Python/3.8.18/x64/lib\n##[endgroup]\n[INFO] Initializing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Initializing environment for https://github.com/google/yapf.\n[INFO] Initializing environment for https://github.com/google/yapf:toml.\n[INFO] Initializing environment for https://github.com/pycqa/isort.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Initializing environment for https://github.com/PyCQA/pydocstyle:toml.\n[INFO] Initializing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Initializing environment for https://github.com/kynan/nbstripout.\n[INFO] Initializing environment for local:[email protected].\n[INFO] Initializing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Installing environment for https://github.com/astral-sh/ruff-pre-commit.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/google/yapf.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pycqa/isort.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/PyCQA/pydocstyle.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/pre-commit/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/Lucas-C/pre-commit-hooks.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/kynan/nbstripout.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for local.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\n[INFO] Installing environment for https://github.com/trufflesecurity/trufflehog.git.\n[INFO] Once installed this environment will be reused.\n[INFO] This may take a few minutes...\nruff.....................................................................Passed\nyapf.....................................................................Failed\n- hook id: yapf\n- files were modified by this hook\n\nReformatting composer/datasets/imagenet.py\nReformatting tests/loggers/test_cometml_logger.py\nReformatting tests/algorithms/test_gradient_clipping.py\nReformatting tests/fixtures/__init__.py\nReformatting composer/callbacks/runtime_estimator.py\nReformatting composer/__init__.py\nReformatting composer/trainer/mosaic_fsdp.py\nReformatting composer/models/efficientnetb0/__init__.py\nReformatting composer/callbacks/export_for_inference.py\nReformatting composer/algorithms/cutmix/__init__.py\nReformatting composer/profiler/profiler_action.py\nReformatting tests/datasets/test_mnist.py\nReformatting composer/utils/inference.py\nReformatting tests/trainer/test_predict.py\nReformatting composer/algorithms/randaugment/__init__.py\nReformatting tests/test_notebooks.py\nReformatting composer/algorithms/alibi/__init__.py\nReformatting composer/models/tasks/__init__.py\nReformatting composer/utils/object_store/gcs_object_store.py\nReformatting composer/models/unet/unet.py\nReformatting composer/utils/object_store/object_store.py\nReformatting composer/loggers/remote_uploader_downloader.py\nReformatting composer/utils/eval_client/local_eval_client.py\nReformatting composer/devices/__init__.py\nReformatting tests/utils/test_file_helpers.py\nReformatting composer/algorithms/ema/__init__.py\nReformatting tests/algorithms/test_cutmix.py\nReformatting tests/fixtures/autouse_fixtures.py\nReformatting composer/_version.py\nReformatting composer/core/time.py\nReformatting tests/common/models.py\nReformatting composer/utils/device.py\nReformatting tests/test_passes.py\nReformatting tests/datasets/test_synthetic_data.py\nReformatting tests/utils/__init__.py\nReformatting composer/trainer/__init__.py\nReformatting tests/utils/test_module_surgery.py\nReformatting composer/models/unet/__init__.py\nReformatting tests/callbacks/test_inference.py\nReformatting composer/loggers/wandb_logger.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/_bert.py\nReformatting tests/algorithms/test_channels_last.py\nReformatting tests/common/__init__.py\nReformatting tests/test_split_batch.py\nReformatting composer/algorithms/factorize/factorize_core.py\nReformatting composer/utils/string_enum.py\nReformatting composer/profiler/profiler_schedule.py\nReformatting tests/models/test_bert.py\nReformatting tests/algorithms/test_blurpool_layers.py\nReformatting composer/devices/device_gpu.py\nReformatting docs/source/doctest_cleanup.py\nReformatting composer/datasets/utils.py\nReformatting docs/source/tables/utils.py\nReformatting composer/cli/launcher.py\nReformatting tests/utils/test_dist.py\nReformatting composer/loggers/mosaicml_logger.py\nReformatting composer/profiler/system_profiler.py\nReformatting composer/optim/scheduler.py\nReformatting composer/callbacks/threshold_stopper.py\nReformatting tests/models/test_hf_model.py\nReformatting tests/utils/eval_client/__init__.py\nReformatting composer/core/data_spec.py\nReformatting composer/models/timm/model.py\nReformatting composer/utils/module_surgery.py\nReformatting composer/models/efficientnetb0/efficientnets.py\nReformatting composer/datasets/in_context_learning_evaluation.py\nReformatting tests/callbacks/test_optimizer_monitor.py\nReformatting composer/callbacks/health_checker.py\nReformatting composer/core/__init__.py\nReformatting composer/loggers/progress_bar_logger.py\nReformatting composer/models/gpt2/__init__.py\nReformatting composer/algorithms/gated_linear_units/__init__.py\nReformatting composer/models/bert/__init__.py\nReformatting tests/algorithms/test_factorize_core.py\nReformatting composer/models/resnet_cifar/__init__.py\nReformatting tests/test_precision.py\nReformatting composer/algorithms/randaugment/randaugment.py\nReformatting composer/algorithms/mixup/mixup.py\nReformatting composer/algorithms/stochastic_depth/stochastic_layers.py\nReformatting composer/devices/device_mps.py\nReformatting composer/utils/object_store/mlflow_object_store.py\nReformatting composer/core/callback.py\nReformatting composer/cli/__init__.py\nReformatting composer/utils/eval_client/lambda_eval_client.py\nReformatting composer/models/resnet_cifar/model.py\nReformatting tests/optim/__init__.py\nReformatting composer/utils/auto_log_hparams.py\nReformatting tests/utils/object_store/test_integration_gs_object_store.py\nReformatting composer/algorithms/factorize/factorize.py\nReformatting tests/callbacks/test_mlperf_callback.py\nReformatting composer/callbacks/activation_monitor.py\nReformatting composer/core/engine.py\nReformatting tests/test_engine.py\nReformatting composer/models/classify_mnist/__init__.py\nReformatting tests/utils/eval_client/test_mcli_eval_client.py\nReformatting tests/utils/test_retrying.py\nReformatting tests/utils/object_store/test_uc_object_store.py\nReformatting composer/algorithms/squeeze_excite/squeeze_excite.py\nReformatting tests/profiler/test_json_trace_handler.py\nReformatting composer/models/base.py\nReformatting tests/callbacks/test_callbacks.py\nReformatting tests/models/test_composer_model.py\nReformatting composer/algorithms/warnings.py\nReformatting tests/trainer/test_fsdp_checkpoint.py\nReformatting tests/loggers/test_progress_bar_logger.py\nReformatting composer/metrics/metrics.py\nReformatting composer/loggers/mlflow_logger.py\nReformatting docs/source/conf.py\nReformatting composer/algorithms/mixup/__init__.py\nReformatting composer/models/unet/model.py\nReformatting composer/models/resnet/__init__.py\nReformatting composer/trainer/dist_strategy.py\nReformatting composer/functional/__init__.py\nReformatting composer/utils/misc.py\nReformatting composer/algorithms/cutout/__init__.py\nReformatting tests/datasets/test_ffcv_utils.py\nReformatting tests/algorithms/test_low_precision_groupnorm.py\nReformatting composer/algorithms/cutout/cutout.py\nReformatting tests/loggers/test_mlflow_logger.py\nReformatting tests/common/markers.py\nReformatting tests/trainer/test_checkpoint.py\nReformatting tests/utils/test_autolog_hparams.py\nReformatting tests/test_time.py\nReformatting composer/algorithms/low_precision_layernorm/__init__.py\nReformatting composer/utils/object_store/uc_object_store.py\nReformatting composer/utils/eval_client/__init__.py\nReformatting composer/algorithms/stochastic_depth/stochastic_depth.py\nReformatting tests/utils/test_dynamic_import.py\nReformatting composer/algorithms/layer_freezing/layer_freezing.py\nReformatting tests/utils/object_store/test_gs_object_store.py\nReformatting docs/source/tables/update_alg_tables.py\nReformatting tests/utils/test_batch_helpers.py\nReformatting composer/trainer/_deepspeed.py\nReformatting composer/utils/reproducibility.py\nReformatting composer/loggers/logger_destination.py\nReformatting tests/callbacks/test_threshold_stopper.py\nReformatting tests/algorithms/test_squeeze_excite.py\nReformatting composer/loggers/tensorboard_logger.py\nReformatting tests/trainer/test_trainer_eval.py\nReformatting tests/loggers/test_file_logger.py\nReformatting tests/utils/object_store/test_s3_object_store.py\nReformatting tests/algorithms/test_selective_backprop.py\nReformatting tests/datasets/test_cifar.py\nReformatting composer/algorithms/weight_standardization/weight_standardization.py\nReformatting composer/callbacks/memory_monitor.py\nReformatting composer/utils/batch_helpers.py\nReformatting tests/metrics/test_current_metrics.py\nReformatting composer/core/evaluator.py\nReformatting composer/algorithms/blurpool/blurpool_layers.py\nReformatting composer/core/types.py\nReformatting composer/algorithms/gyro_dropout/__init__.py\nReformatting tests/trainer/test_scale_schedule.py\nReformatting composer/models/deeplabv3/model.py\nReformatting composer/algorithms/gated_linear_units/gated_linear_unit_layers.py\nReformatting composer/algorithms/augmix/__init__.py\nReformatting composer/algorithms/channels_last/channels_last.py\nReformatting composer/models/unet/_layers.py\nReformatting composer/utils/__init__.py\nReformatting composer/core/state.py\nReformatting composer/algorithms/seq_length_warmup/__init__.py\nReformatting docs/source/tables/generate_cost_graphs.py\nReformatting composer/algorithms/label_smoothing/label_smoothing.py\nReformatting tests/utils/object_store/test_libcloud_object_store.py\nReformatting composer/algorithms/alibi/alibi.py\nReformatting composer/algorithms/progressive_resizing/__init__.py\nReformatting tests/models/test_gpt2.py\nReformatting tests/__init__.py\nReformatting tests/loggers/test_mosaicml_logger.py\nReformatting tests/metrics/__init__.py\nReformatting tests/test_smoketest.py\nReformatting composer/datasets/brats.py\nReformatting composer/algorithms/low_precision_groupnorm/low_precision_groupnorm.py\nReformatting tests/trainer/test_ddp_sync_strategy.py\nReformatting tests/trainer/test_trainer.py\nReformatting composer/profiler/torch_profiler.py\nReformatting composer/utils/object_store/sftp_object_store.py\nReformatting composer/algorithms/selective_backprop/__init__.py\nReformatting composer/algorithms/low_precision_layernorm/low_precision_layernorm.py\nReformatting tests/callbacks/test_memory_monitor.py\nReformatting tests/datasets/test_add_dataset_transform.py\nReformatting composer/algorithms/weight_standardization/__init__.py\nReformatting composer/algorithms/ghost_batchnorm/__init__.py\nReformatting composer/loggers/console_logger.py\nReformatting tests/test_docs.py\nReformatting tests/algorithms/test_alibi.py\nReformatting composer/__main__.py\nReformatting docker/pillow_stub/setup.py\nReformatting composer/models/resnet_cifar/resnets.py\nReformatting composer/datasets/c4.py\nReformatting tests/loggers/test_wandb_logger.py\nReformatting composer/algorithms/no_op_model/no_op_model.py\nReformatting tests/algorithms/test_torch_export.py\nReformatting tests/common/state.py\nReformatting composer/models/initializers.py\nReformatting composer/models/gpt2/model.py\nReformatting tests/metrics/test_miou.py\nReformatting tests/metrics/test_map.py\nReformatting docs/source/tables/__init__.py\nReformatting tests/test_state.py\nReformatting composer/loss/loss.py\nReformatting composer/callbacks/generate.py\nReformatting tests/metrics/metric_setter_callback.py\nReformatting composer/algorithms/channels_last/__init__.py\nReformatting composer/utils/retrying.py\nReformatting composer/algorithms/stochastic_depth/__init__.py\nReformatting composer/core/algorithm.py\nReformatting composer/models/vit_small_patch16/__init__.py\nReformatting tests/optim/test_scheduler.py\nReformatting composer/models/efficientnetb0/model.py\nReformatting composer/models/__init__.py\nReformatting composer/algorithms/factorize/__init__.py\nReformatting composer/callbacks/lr_monitor.py\nReformatting composer/loggers/__init__.py\nReformatting composer/callbacks/__init__.py\nReformatting composer/callbacks/image_visualizer.py\nReformatting composer/utils/eval_client/mosaicml_lambda_eval_client.py\nReformatting tests/test_simple_nlp.py\nReformatting tests/utils/eval_client/test_local_eval_client.py\nReformatting composer/metrics/nlp.py\nReformatting tests/trainer/test_fsdp_param_groups.py\nReformatting tests/algorithms/test_stochastic_depth.py\nReformatting composer/utils/fx_utils.py\nReformatting composer/algorithms/squeeze_excite/__init__.py\nReformatting examples/profiler_demo.py\nReformatting tests/algorithms/test_progressive_resizing.py\nReformatting tests/callbacks/test_runtime_estimator.py\nReformatting composer/loss/__init__.py\nReformatting tests/common/events.py\nReformatting composer/algorithms/gyro_dropout/gyro_dropout.py\nReformatting tests/callbacks/test_nan_monitor.py\nReformatting composer/models/timm/__init__.py\nReformatting composer/algorithms/utils/augmentation_common.py\nReformatting composer/profiler/__init__.py\nReformatting composer/utils/eval_client/eval_client.py\nReformatting composer/utils/object_store/s3_object_store.py\nReformatting tests/datasets/__init__.py\nReformatting composer/models/bert/model.py\nReformatting tests/algorithms/test_colout.py\nReformatting composer/utils/dist.py\nReformatting composer/loggers/cometml_logger.py\nReformatting composer/utils/object_store/oci_object_store.py\nReformatting composer/algorithms/no_op_model/__init__.py\nReformatting tests/trainer/__init__.py\nReformatting docs/source/tables/update_model_tables.py\nReformatting composer/utils/import_helpers.py\nReformatting tests/datasets/test_segmentation_transforms.py\nReformatting tests/loggers/__init__.py\nReformatting tests/algorithms/test_ghost_batchnorm.py\nReformatting tests/test_device.py\nReformatting composer/utils/file_helpers.py\nReformatting tests/loggers/test_slack_logger.py\nReformatting composer/devices/device.py\nReformatting composer/algorithms/__init__.py\nReformatting composer/callbacks/speed_monitor.py\nReformatting tests/algorithms/test_required_on_load.py\nReformatting composer/utils/object_store/libcloud_object_store.py\nReformatting tests/datasets/test_in_context_learning_datasets.py\nReformatting tests/callbacks/__init__.py\nReformatting tests/callbacks/test_system_metrics_monitor.py\nReformatting tests/test_cli.py\nReformatting tests/algorithms/test_low_precision_layernorm.py\nReformatting setup.py\nReformatting composer/metrics/__init__.py\nReformatting composer/callbacks/mlperf.py\nReformatting composer/loggers/logger.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/__init__.py\nReformatting tests/callbacks/test_loggers_across_callbacks.py\nReformatting composer/algorithms/swa/swa.py\nReformatting tests/utils/test_iter_helpers.py\nReformatting examples/custom_models.py\nReformatting tests/trainer/test_dataspec.py\nReformatting tests/metrics/test_nlp_metrics.py\nReformatting examples/segmentation/train_deeplabv3_ade20k.py\nReformatting tests/algorithms/test_weight_standardization.py\nReformatting examples/imagenet/train_resnet_imagenet1k.py\nReformatting composer/algorithms/progressive_resizing/progressive_resizing.py\nReformatting docker/generate_build_matrix.py\nReformatting composer/datasets/__init__.py\nReformatting composer/algorithms/cutmix/cutmix.py\nReformatting tests/utils/test_inference.py\nReformatting composer/devices/device_cpu.py\nReformatting composer/core/serializable.py\nReformatting tests/utils/object_store/test_mlflow_object_store.py\nReformatting composer/callbacks/early_stopper.py\nReformatting tests/common/compare.py\nReformatting composer/callbacks/optimizer_monitor.py\nReformatting tests/algorithms/test_factorize_algorithm.py\nReformatting composer/algorithms/sam/__init__.py\nReformatting docs/source/doctest_fixtures.py\nReformatting composer/loss/utils.py\nReformatting composer/profiler/trace_handler.py\nReformatting tests/algorithms/__init__.py\nReformatting tests/callbacks/test_health_checker.py\nReformatting composer/trainer/_scaler.py\nReformatting tests/algorithms/algorithm_settings.py\nReformatting composer/devices/device_tpu.py\nReformatting composer/models/deeplabv3/__init__.py\nReformatting tests/algorithms/test_blurpool.py\nReformatting tests/algorithms/test_layer_freezing.py\nReformatting tests/algorithms/test_augmentations_functional.py\nReformatting composer/algorithms/augmix/augmix.py\nReformatting composer/profiler/json_trace_handler.py\nReformatting tests/fixtures/fixtures.py\nReformatting tests/callbacks/test_early_stopper.py\nReformatting composer/algorithms/blurpool/__init__.py\nReformatting composer/profiler/utils.py\nReformatting composer/algorithms/selective_backprop/selective_backprop.py\nReformatting composer/datasets/lm_dataset.py\nReformatting composer/callbacks/checkpoint_saver.py\nReformatting tests/utils/object_store/object_store_settings.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/utils.py\nReformatting tests/callbacks/test_speed_monitor.py\nReformatting tests/models/test_mmdet_model.py\nReformatting composer/algorithms/gradient_clipping/gradient_clipping.py\nReformatting .github/mcli/mcli_pytest.py\nReformatting tests/algorithms/test_algorithms_train.py\nReformatting composer/datasets/ade20k.py\nReformatting composer/algorithms/colout/colout.py\nReformatting composer/models/mmdetection.py\nReformatting composer/algorithms/ghost_batchnorm/ghost_batchnorm.py\nReformatting composer/loggers/file_logger.py\nReformatting composer/algorithms/label_smoothing/__init__.py\nReformatting tests/datasets/test_dataset_utils.py\nReformatting composer/algorithms/utils/__init__.py\nReformatting tests/algorithms/test_label_smoothing.py\nReformatting composer/datasets/synthetic.py\nReformatting tests/loggers/test_in_memory_logger.py\nReformatting composer/models/efficientnetb0/_layers.py\nReformatting composer/algorithms/utils/augmentation_primitives.py\nReformatting tests/algorithms/test_cutout.py\nReformatting tests/loggers/test_console_logger.py\nReformatting tests/utils/object_store/__init__.py\nReformatting composer/algorithms/gradient_clipping/__init__.py\nReformatting tests/utils/eval_client/test_lambda_eval_client.py\nReformatting tests/callbacks/test_image_visualizer.py\nReformatting tests/algorithms/test_mixup.py\nReformatting composer/utils/object_store/__init__.py\nReformatting composer/profiler/marker.py\nReformatting tests/conftest.py\nReformatting composer/cli/__main__.py\nReformatting tests/utils/object_store/test_oci_object_store.py\nReformatting tests/algorithms/test_gated_linear_units.py\nReformatting composer/callbacks/system_metrics_monitor.py\nReformatting composer/algorithms/blurpool/blurpool.py\nReformatting tests/models/test_efficientnet.py\nReformatting tests/callbacks/test_generate.py\nReformatting tests/utils/test_fx_utils.py\nReformatting composer/loggers/in_memory_logger.py\nReformatting .github/bin/gen_docker_matrix.py\nReformatting tests/algorithms/test_sam.py\nReformatting composer/optim/decoupled_weight_decay.py\nReformatting composer/core/event.py\nReformatting tests/algorithms/test_seq_length_warmup.py\nReformatting composer/algorithms/ema/ema.py\nReformatting tests/utils/object_store/test_object_store.py\nReformatting composer/models/tasks/classification.py\nReformatting composer/algorithms/layer_freezing/__init__.py\nReformatting composer/core/passes.py\nReformatting composer/utils/iter_helpers.py\nReformatting tests/test_events.py\nReformatting examples/checkpoint_with_wandb.py\nReformatting composer/models/resnet/model.py\nReformatting composer/core/precision.py\nReformatting docs/source/tables/update_methods_overview.py\nReformatting tests/callbacks/test_activation_monitor.py\nReformatting tests/trainer/test_ddp.py\nReformatting composer/callbacks/utils.py\nReformatting tests/algorithms/test_ema.py\nReformatting tests/profiler/test_profiler.py\nReformatting composer/algorithms/sam/sam.py\nReformatting tests/trainer/test_fsdp.py\nReformatting composer/models/classify_mnist/model.py\nReformatting composer/algorithms/seq_length_warmup/seq_length_warmup.py\nReformatting composer/datasets/cifar.py\nReformatting composer/trainer/mosaic_fsdp_utils.py\nReformatting tests/callbacks/callback_settings.py\nReformatting composer/trainer/_scale_schedule.py\nReformatting tests/test_full_nlp.py\nReformatting composer/algorithms/factorize/factorize_modules.py\nReformatting composer/models/vit_small_patch16/model.py\nReformatting composer/devices/device_hpu.py\nReformatting examples/gyro_dropout_example.py\nReformatting composer/algorithms/low_precision_groupnorm/__init__.py\nReformatting composer/utils/checkpoint.py\nReformatting tests/common/datasets.py\nReformatting composer/algorithms/swa/__init__.py\nReformatting tests/algorithms/test_algorithm_resumption.py\nReformatting tests/utils/test_string_enum.py\nReformatting tests/algorithms/test_factorized_modules.py\nReformatting tests/algorithms/algorithm_test_template.py\nReformatting composer/datasets/mnist.py\nReformatting composer/models/huggingface.py\nReformatting composer/callbacks/free_outputs.py\nReformatting tests/test_loss.py\nReformatting composer/algorithms/colout/__init__.py\nReformatting composer/metrics/map.py\nReformatting composer/utils/collect_env.py\nReformatting composer/profiler/profiler.py\nReformatting tests/test_docker.py\nReformatting tests/loggers/test_remote_uploader_downloader.py\nReformatting tests/algorithms/test_gyro_dropout.py\nReformatting tests/datasets/test_streaming_datasets_train.py\nReformatting composer/algorithms/gated_linear_units/gated_linear_units.py\nReformatting composer/datasets/ffcv_utils.py\nReformatting composer/callbacks/nan_monitor.py\nReformatting scripts/ffcv/create_ffcv_datasets.py\nReformatting composer/algorithms/alibi/attention_surgery_functions/_gpt2.py\nReformatting tests/profiler/__init__.py\nReformatting composer/loggers/slack_logger.py\nReformatting composer/optim/__init__.py\nReformatting tests/loggers/test_logger.py\nReformatting composer/trainer/meta_safe_apply.py\nReformatting composer/profiler/json_trace_merger.py\nReformatting composer/trainer/trainer.py\n\nisort....................................................................Passed\npydocstyle...............................................................Passed\ncheck for added large files..............................................Passed\ncheck python ast.........................................................Passed\ncheck builtin type constructor use.......................................Passed\ncheck for case conflicts.................................................Passed\ncheck docstring is first.................................................Passed\ncheck that executables have shebangs.....................................Passed\ncheck json...............................................................Passed\ncheck that scripts with shebangs are executable..........................Passed\npretty format json.......................................................Passed\ncheck for merge conflicts................................................Passed\ncheck for broken symlinks................................................Passed\ncheck toml...............................................................Passed\ncheck vcs permalinks.....................................................Passed\ncheck xml................................................................Passed\ncheck yaml...............................................................Passed\ndebug statements (python)................................................Passed\ndetect destroyed symlinks................................................Passed\nfix double quoted strings................................................Passed\nfix end of files.........................................................Passed\nfix utf-8 byte order marker..............................................Passed\nmixed line ending........................................................Passed\ntrim trailing whitespace.................................................Passed\nInsert license in comments...............................................Passed\nnbstripout...............................................................Passed\npyright..................................................................Passed\nsecret scan..............................................................Passed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/composer/models/huggingface.py b/composer/models/huggingface.py\nindex 2498c5f3..d7fb0404 100644\n--- a/composer/models/huggingface.py\n+++ b/composer/models/huggingface.py\n@@ -169,13 +169,13 @@ class HuggingFaceModel(ComposerModel):\n def state_dict(self, *args, **kwargs) -> Dict[str, Any]:\n \"\"\"Returns the state dict of the model.\"\"\"\n full_state_dict = super().state_dict(*args, **kwargs)\n- \n+\n if self.peft_filter_state_dict_trainable:\n- full_state_dict = filter_state_dict_peft(full_state_dict, self.model.peft_config[self.model.active_adapter], False)\n+ full_state_dict = filter_state_dict_peft(full_state_dict, self.model.peft_config[self.model.active_adapter],\n+ False)\n \n return full_state_dict\n \n-\n @staticmethod\n def load_huggingface_tokenizer_from_saved_state(\n hf_state: Dict[str, Any],\n@@ -803,7 +803,10 @@ def write_huggingface_pretrained_from_composer_checkpoint(\n else:\n torch.save(weights_state_dict, Path(output_folder) / 'pytorch_model.bin')\n \n-def filter_state_dict_peft(state_dict: Dict[str, Any], peft_config: 'PeftConfig', remove_adapter_names: bool = True) -> Dict[str, Any]:\n+\n+def filter_state_dict_peft(state_dict: Dict[str, Any],\n+ peft_config: 'PeftConfig',\n+ remove_adapter_names: bool = True) -> Dict[str, Any]:\n # Filtering copied from https://github.com/huggingface/peft/blob/4186c9b104644fd247a4cc0dc2dfc1ede4665204/src/peft/utils/save_and_load.py#L68C1-L86C116\n bias = peft_config.bias\n if bias == 'none':\n@@ -821,7 +824,7 @@ def filter_state_dict_peft(state_dict: Dict[str, Any], peft_config: 'PeftConfig'\n else:\n raise NotImplementedError\n to_return = {k: v for k, v in to_return.items() if (('lora_' in k and 'default' in k) or ('bias' in k))}\n- \n+\n if remove_adapter_names:\n to_return = {k.replace(f'.default', ''): v for k, v in to_return.items()}\n- return to_return\n\\n+ return to_return\ndiff --git a/tests/models/test_hf_model.py b/tests/models/test_hf_model.py\nindex d7b7081d..0cf13fb3 100644\n--- a/tests/models/test_hf_model.py\n+++ b/tests/models/test_hf_model.py\n@@ -1223,6 +1223,7 @@ def test_peft_init_not_installed(tiny_gpt2_model, gpt2_peft_config):\n from composer.models import HuggingFaceModel\n _ = HuggingFaceModel(tiny_gpt2_model, peft_config=gpt2_peft_config)\n \n+\n @pytest.mark.parametrize('just_lora', [True, False])\n def test_peft_trains_and_loads(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path, just_lora):\n pytest.importorskip('peft')\n@@ -1283,6 +1284,7 @@ def test_peft_metadata(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config):\n \n assert loaded_peft_config == gpt2_peft_config\n \n+\n @pytest.mark.parametrize('just_lora', [True, False])\n def test_peft_write_hf_from_composer(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path, just_lora):\n peft = pytest.importorskip('peft')\n@@ -1369,11 +1371,14 @@ def test_peft_fsdp_trains(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config\n for p1, p2 in zip(trainer.state.model.parameters(), load_trainer.state.model.parameters()):\n torch.testing.assert_close(p1, p2)\n \n+\n def test_filtered_state_dict(tiny_gpt2_model, tiny_gpt2_tokenizer, gpt2_peft_config, tmp_path):\n pytest.importorskip('peft')\n \n- hf_model = HuggingFaceModel(tiny_gpt2_model, tokenizer=tiny_gpt2_tokenizer, peft_config=gpt2_peft_config, peft_filter_state_dict_trainable=True)\n+ hf_model = HuggingFaceModel(tiny_gpt2_model,\n+ tokenizer=tiny_gpt2_tokenizer,\n+ peft_config=gpt2_peft_config,\n+ peft_filter_state_dict_trainable=True)\n state_dict = hf_model.state_dict()\n \n assert len(state_dict.keys()) == 4\n-\n", "difficulty": 2, "changed_files": ["composer/models/huggingface.py", "tests/models/test_hf_model.py"], "commit_link": "https://github.com/mosaicml/composer/tree/4b7e4724b2f92741fb094421ebfa60be5bfb18d9"}
|
|
|
|
data/python/4d5898b.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"language": "Python", "id": 27, "repo_owner": "canonical", "repo_name": "cloud-init", "head_branch": "holman/dhcpcd", "workflow_name": "Lint Tests", "workflow_filename": "check_format.yml", "workflow_path": ".github/workflows/check_format.yml", "contributor": "holmanb", "sha_fail": "4d5898b8a73c93e1ed4434744c2fa7c3f7fbd501", "sha_success": "2bd296ef8983a716a7b8d107571181333f093109", "workflow": "name: Lint Tests\non:\n pull_request:\n push:\n branches:\n - main\n\nconcurrency:\n group: 'ci-${{ github.workflow }}-${{ github.ref }}'\n cancel-in-progress: true\ndefaults:\n run:\n shell: sh -ex {0}\n\njobs:\n check_format:\n strategy:\n fail-fast: false\n matrix:\n env: [ruff, mypy, pylint, black, isort]\n lint-with:\n - {tip-versions: false, os: ubuntu-20.04}\n - {tip-versions: true, os: ubuntu-latest}\n name: Check ${{ matrix.lint-with.tip-versions && 'tip-' || '' }}${{ matrix.env }}\n runs-on: ${{ matrix.lint-with.os }}\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n\n - name: Dependencies\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n\n - name: Print version\n run: python3 --version\n\n - name: Test\n if: ${{ !matrix.lint-with.tip-versions }}\n env:\n # matrix env: not to be confused w/environment variables or testenv\n TOXENV: ${{ matrix.env }}\n run: tox\n - name: Test (tip versions)\n if: matrix.lint-with.tip-versions\n continue-on-error: true\n env:\n TOXENV: tip-${{ matrix.env }}\n run: tox\n schema-format:\n strategy:\n fail-fast: false\n name: Check json format\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Test format\"\n run: |\n tools/check_json_format.sh cloudinit/config/schemas/schema-cloud-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/schema-network-config-v1.json\n tools/check_json_format.sh cloudinit/config/schemas/versions.schema.cloud-config.json\n\n doc:\n strategy:\n fail-fast: false\n name: Check docs\n runs-on: ubuntu-22.04\n steps:\n - name: \"Checkout #1\"\n uses: actions/[email protected]\n\n - name: \"Checkout #2 (for tools/read-version)\"\n run: |\n git fetch --unshallow\n git remote add upstream https://git.launchpad.net/cloud-init\n - name: \"Install Python 3.10\"\n uses: actions/setup-python@v4\n with:\n python-version: '3.10.8'\n - name: \"Install dependencies\"\n run: |\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy update\n sudo DEBIAN_FRONTEND=noninteractive apt-get -qy install tox\n - name: \"Spellcheck\"\n run: |\n tox\n env:\n TOXENV: doc-spelling\n - name: \"Build docs\"\n env:\n TOXENV: doc\n run: |\n tox\n\n shell-lint:\n name: Shell Lint\n runs-on: ubuntu-22.04\n steps:\n - name: Checkout\n uses: actions/[email protected]\n\n - name: Install ShellCheck\n run: |\n sudo apt-get update\n sudo apt install shellcheck\n\n - name: Run ShellCheck\n run: |\n shellcheck ./tools/ds-identify\n\n check-cla-signers:\n runs-on: ubuntu-22.04\n steps:\n - uses: actions/[email protected]\n\n - name: Check CLA signers file\n run: tools/check-cla-signers\n", "logs": [{"step_name": "Check pylint/6_Test.txt", "log": "##[group]Run tox\n\u001b[36;1mtox\u001b[0m\nshell: /usr/bin/sh -ex {0}\nenv:\n TOXENV: pylint\n##[endgroup]\n+ tox\nGLOB sdist-make: /home/runner/work/cloud-init/cloud-init/setup.py\npylint create: /home/runner/work/cloud-init/cloud-init/.tox/pylint\npylint installdeps: pylint===2.13.9, -r/home/runner/work/cloud-init/cloud-init/test-requirements.txt, -r/home/runner/work/cloud-init/cloud-init/integration-requirements.txt\npylint inst: /home/runner/work/cloud-init/cloud-init/.tox/.tmp/package/1/cloud-init-23.4+88.g0e8fe220.zip\npylint installed: adal==1.2.7,anyio==4.2.0,appdirs==1.4.4,applicationinsights==0.11.10,argcomplete==3.1.6,astroid==2.11.7,attrs==23.2.0,autopage==0.5.2,azure-cli-core==2.56.0,azure-cli-telemetry==1.1.0,azure-common==1.1.28,azure-core==1.29.6,azure-identity==1.15.0,azure-mgmt-compute==30.4.0,azure-mgmt-core==1.4.0,azure-mgmt-network==25.2.0,azure-mgmt-resource==23.0.1,bcrypt==4.1.2,boto3==1.34.17,botocore==1.34.17,cachetools==5.3.2,certifi==2023.11.17,cffi==1.16.0,charset-normalizer==3.3.2,circuitbreaker==1.4.0,cliff==4.5.0,cloud-init==23.4+88.g0e8fe220,cmd2==2.4.3,configobj==5.0.8,coverage==7.4.0,cryptography==41.0.7,debtcollector==2.5.0,decorator==5.1.1,dill==0.3.7,distro==1.9.0,dogpile.cache==1.3.0,exceptiongroup==1.2.0,google-api-core==2.15.0,google-api-python-client==2.113.0,google-auth==2.26.2,google-auth-httplib2==0.2.0,googleapis-common-protos==1.62.0,httplib2==0.22.0,humanfriendly==10.0,ibm-cloud-sdk-core==3.18.2,ibm-platform-services==0.49.0,ibm-vpc==0.20.0,idna==3.6,importlib-metadata==7.0.1,importlib-resources==6.1.1,iniconfig==2.0.0,iso8601==2.1.0,isodate==0.6.1,isort==5.13.2,Jinja2==3.1.3,jmespath==1.0.1,jsonpatch==1.33,jsonpointer==2.4,jsonschema==4.20.0,jsonschema-specifications==2023.12.1,keystoneauth1==5.5.0,knack==0.11.0,lazy-object-proxy==1.10.0,MarkupSafe==2.1.3,mccabe==0.7.0,msal==1.24.0b2,msal-extensions==1.0.0,msgpack==1.0.7,msrest==0.7.1,msrestazure==0.6.4,netaddr==0.10.1,netifaces==0.11.0,oauthlib==3.2.2,oci==2.118.2,openstacksdk==1.4.0,os-service-types==1.7.0,osc-lib==3.0.0,oslo.config==9.3.0,oslo.i18n==6.2.0,oslo.serialization==5.3.0,oslo.utils==6.3.0,packaging==23.2,paramiko==3.4.0,passlib==1.7.4,pbr==6.0.0,pkginfo==1.9.6,pkgutil-resolve-name==1.3.10,platformdirs==4.1.0,pluggy==1.3.0,portalocker==2.8.2,prettytable==3.9.0,protobuf==3.19.6,psutil==5.9.7,pyasn1==0.5.1,pyasn1-modules==0.3.0,pycloudlib==1!5.12.1,pycparser==2.21,pygments==2.17.2,PyJWT==2.8.0,pylint==2.13.9,PyNaCl==1.5.0,pyOpenSSL==23.3.0,pyparsing==2.4.7,pyperclip==1.8.2,pyserial==3.5,pytest==7.4.4,pytest-cov==4.1.0,pytest-mock==3.12.0,python-cinderclient==9.4.0,python-dateutil==2.8.2,python-keystoneclient==5.3.0,python-novaclient==18.4.0,python-openstackclient==6.4.0,pytz==2023.3.post1,PyYAML==6.0.1,qemu.qmp==0.0.3,referencing==0.32.1,requests==2.31.0,requests-oauthlib==1.3.1,requestsexceptions==1.4.0,responses==0.24.1,rfc3986==2.0.0,rpds-py==0.17.1,rsa==4.9,s3transfer==0.10.0,simplejson==3.19.2,six==1.16.0,sniffio==1.3.0,stevedore==5.1.0,tabulate==0.9.0,toml==0.10.0,tomli==2.0.1,typing-extensions==4.9.0,tzdata==2023.4,uritemplate==4.1.1,urllib3==2.1.0,wcwidth==0.2.13,wrapt==1.16.0,zipp==3.17.0\npylint run-test-pre: PYTHONHASHSEED='1345909868'\npylint run-test: commands[0] | /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py\n************* Module cloudinit.net.dhcp\ncloudinit/net/dhcp.py:538: [W0235(useless-super-delegation), Dhcpcd.__init__] Useless super delegation in method '__init__'\n\n------------------------------------\nYour code has been rated at 10.00/10\n\nERROR: InvocationError for command /home/runner/work/cloud-init/cloud-init/.tox/pylint/bin/python -m pylint cloudinit/ tests/ tools/ conftest.py setup.py (exited with code 4)\n___________________________________ summary ____________________________________\nERROR: pylint: commands failed\n##[error]Process completed with exit code 1.\n"}], "diff": "diff --git a/cloudinit/distros/__init__.py b/cloudinit/distros/__init__.py\nindex 5c891f26e..c02166ad2 100644\n--- a/cloudinit/distros/__init__.py\n+++ b/cloudinit/distros/__init__.py\n@@ -170,6 +170,7 @@ class Distro(persistence.CloudInitPickleMixin, metaclass=abc.ABCMeta):\n self._runner = helpers.Runners(paths)\n self.package_managers: List[PackageManager] = []\n self._dhcp_client = None\n+ self._fallback_interface = None\n \n def _unpickle(self, ci_pkl_version: int) -> None:\n \"\"\"Perform deserialization fixes for Distro.\"\"\"\n@@ -1274,6 +1275,13 @@ class Distro(persistence.CloudInitPickleMixin, metaclass=abc.ABCMeta):\n \"/bin/true\",\n ] + ([\"-cf\", config_file, interface] if config_file else [interface])\n \n+ @property\n+ def fallback_interface(self):\n+ \"\"\"Determine the network interface used during local network config.\"\"\"\n+ if self._fallback_interface is None:\n+ self._fallback_interface = net.find_fallback_nic()\n+ return self._fallback_interface\n+\n \n def _apply_hostname_transformations_to_url(url: str, transformations: list):\n \"\"\"\ndiff --git a/cloudinit/net/dhcp.py b/cloudinit/net/dhcp.py\nindex 9c94b3f92..7b2d41df1 100644\n--- a/cloudinit/net/dhcp.py\n+++ b/cloudinit/net/dhcp.py\n@@ -19,8 +19,6 @@ import configobj\n \n from cloudinit import subp, temp_utils, util\n from cloudinit.net import (\n- find_fallback_nic,\n- get_devicelist,\n get_ib_interface_hwaddr,\n get_interface_mac,\n is_ib_interface,\n@@ -98,17 +96,8 @@ def maybe_perform_dhcp_discovery(distro, nic=None, dhcp_log_func=None):\n from the dhclient discovery if run, otherwise an empty list is\n returned.\n \"\"\"\n- if nic is None:\n- nic = find_fallback_nic()\n- if nic is None:\n- LOG.debug(\"Skip dhcp_discovery: Unable to find fallback nic.\")\n- raise NoDHCPLeaseInterfaceError()\n- elif nic not in get_devicelist():\n- LOG.debug(\n- \"Skip dhcp_discovery: nic %s not found in get_devicelist.\", nic\n- )\n- raise NoDHCPLeaseInterfaceError()\n- return distro.dhcp_client.dhcp_discovery(nic, dhcp_log_func, distro)\n+ interface = nic or distro.fallback_interface\n+ return distro.dhcp_client.dhcp_discovery(interface, dhcp_log_func, distro)\n \n \n def networkd_parse_lease(content):\n@@ -185,7 +174,7 @@ class DhcpClient(abc.ABC):\n def parse_static_routes(routes: str) -> List[Tuple[str, str]]:\n return []\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n return {}\n \n \n@@ -218,15 +207,15 @@ class IscDhclient(DhcpClient):\n dhcp_leases.append(dict(lease_options))\n return dhcp_leases\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n \"\"\"Get the most recent lease from the ephemeral phase as a dict.\n \n Return a dict of dhcp options. The dict contains key value\n pairs from the most recent lease.\n \n- @param distro: a distro object - not used in this class, but required\n- for function signature compatibility with other classes\n- that require a distro object\n+ @param interface: an interface string - not used in this class, but\n+ required for function signature compatibility with other classes\n+ that require a distro object\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\n@@ -355,7 +344,7 @@ class IscDhclient(DhcpClient):\n )\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n- lease = self.get_newest_lease(distro)\n+ lease = self.get_newest_lease(interface)\n if lease:\n return lease\n raise InvalidDHCPLeaseFileError()\n@@ -535,9 +524,6 @@ class IscDhclient(DhcpClient):\n class Dhcpcd(DhcpClient):\n client_name = \"dhcpcd\"\n \n- def __init__(self):\n- super().__init__()\n-\n def dhcp_discovery(\n self,\n interface: str,\n@@ -547,12 +533,12 @@ class Dhcpcd(DhcpClient):\n \"\"\"Run dhcpcd on the interface without scripts/filesystem artifacts.\n \n @param interface: Name of the network interface on which to send a\n- dhcp request\n+ dhcp request\n @param dhcp_log_func: A callable accepting the client output and\n- error streams.\n+ error streams.\n @param distro: a distro object for network interface manipulation\n @return: dict of lease options representing the most recent dhcp lease\n- parsed from the dhclient.lease file\n+ parsed from the dhclient.lease file\n \"\"\"\n LOG.debug(\"Performing a dhcp discovery on %s\", interface)\n \n@@ -580,7 +566,7 @@ class Dhcpcd(DhcpClient):\n )\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n- lease = self.get_newest_lease(distro)\n+ lease = self.get_newest_lease(interface)\n if lease:\n return lease\n raise NoDHCPLeaseError(\"No lease found\")\n@@ -644,12 +630,10 @@ class Dhcpcd(DhcpClient):\n # for compatibility we still expect a list of leases\n return lease\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n- \"\"\"Return a lease\n-\n- Return a list of dicts of dhcp options. Each dict contains key value\n- pairs a specific lease in order from oldest to newest.\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n+ \"\"\"Return a dict of dhcp options.\n \n+ @param interface: which interface to dump the lease from\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\n@@ -659,11 +643,11 @@ class Dhcpcd(DhcpClient):\n [\n \"dhcpcd\",\n \"--dumplease\",\n- distro.fallback_interface,\n+ interface,\n ],\n rcs=[0, 1],\n ).stdout,\n- distro.fallback_interface,\n+ interface,\n )\n \n except subp.ProcessExecutionError as error:\n@@ -781,17 +765,17 @@ class Udhcpc(DhcpClient):\n if dhcp_log_func is not None:\n dhcp_log_func(out, err)\n \n- return self.get_newest_lease(distro)\n+ return self.get_newest_lease(interface)\n \n- def get_newest_lease(self, distro) -> Dict[str, Any]:\n+ def get_newest_lease(self, interface: str) -> Dict[str, Any]:\n \"\"\"Get the most recent lease from the ephemeral phase as a dict.\n \n Return a dict of dhcp options. The dict contains key value\n pairs from the most recent lease.\n \n- @param distro: a distro object - not used in this class, but required\n- for function signature compatibility with other classes\n- that require a distro object\n+ @param interface: an interface name - not used in this class, but\n+ required for function signature compatibility with other classes\n+ that require a distro object\n @raises: InvalidDHCPLeaseFileError on empty or unparseable leasefile\n content.\n \"\"\"\ndiff --git a/cloudinit/sources/__init__.py b/cloudinit/sources/__init__.py\nindex c207b5ed6..f9650268a 100644\n--- a/cloudinit/sources/__init__.py\n+++ b/cloudinit/sources/__init__.py\n@@ -610,13 +610,11 @@ class DataSource(CloudInitPickleMixin, metaclass=abc.ABCMeta):\n @property\n def fallback_interface(self):\n \"\"\"Determine the network interface used during local network config.\"\"\"\n- if self._fallback_interface is None:\n- self._fallback_interface = net.find_fallback_nic()\n- if self._fallback_interface is None:\n- LOG.warning(\n- \"Did not find a fallback interface on %s.\", self.cloud_name\n- )\n- return self._fallback_interface\n+ if self.distro.fallback_interface is None:\n+ LOG.warning(\n+ \"Did not find a fallback interface on %s.\", self.cloud_name\n+ )\n+ return self.distro.fallback_interface\n \n @property\n def platform_type(self):\ndiff --git a/tests/unittests/net/test_dhcp.py b/tests/unittests/net/test_dhcp.py\nindex ffa3eab17..9af5f93aa 100644\n--- a/tests/unittests/net/test_dhcp.py\n+++ b/tests/unittests/net/test_dhcp.py\n@@ -372,20 +372,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n with_logs = True\n ib_address_prefix = \"00:00:00:00:00:00:00:00:00:00:00:00\"\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n- def test_no_fallback_nic_found(self, m_fallback_nic):\n- \"\"\"Log and do nothing when nic is absent and no fallback is found.\"\"\"\n- m_fallback_nic.return_value = None # No fallback nic found\n-\n- with pytest.raises(NoDHCPLeaseInterfaceError):\n- maybe_perform_dhcp_discovery(MockDistro())\n-\n- self.assertIn(\n- \"Skip dhcp_discovery: Unable to find fallback nic.\",\n- self.logs.getvalue(),\n- )\n-\n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=\"eth9\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\", return_value=\"eth9\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n @mock.patch(\"cloudinit.net.dhcp.subp.subp\")\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n@@ -406,7 +393,7 @@ class TestDHCPDiscoveryClean(CiTestCase):\n self.logs.getvalue(),\n )\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=\"eth9\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\", return_value=\"eth9\")\n @mock.patch(\"cloudinit.net.dhcp.os.remove\")\n @mock.patch(\"cloudinit.net.dhcp.subp.subp\")\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n@@ -435,17 +422,6 @@ class TestDHCPDiscoveryClean(CiTestCase):\n self.logs.getvalue(),\n )\n \n- @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\", return_value=None)\n- def test_provided_nic_does_not_exist(self, m_fallback_nic):\n- \"\"\"When the provided nic doesn't exist, log a message and no-op.\"\"\"\n- with pytest.raises(NoDHCPLeaseInterfaceError):\n- maybe_perform_dhcp_discovery(MockDistro(), \"idontexist\")\n-\n- self.assertIn(\n- \"Skip dhcp_discovery: nic idontexist not found in get_devicelist.\",\n- self.logs.getvalue(),\n- )\n-\n @mock.patch(\"cloudinit.net.dhcp.subp.which\")\n @mock.patch(\"cloudinit.net.dhcp.find_fallback_nic\")\n def test_absent_dhclient_command(self, m_fallback, m_which):\ndiff --git a/tests/unittests/sources/test_cloudstack.py b/tests/unittests/sources/test_cloudstack.py\nindex 8ee04b3e1..b33c5e492 100644\n--- a/tests/unittests/sources/test_cloudstack.py\n+++ b/tests/unittests/sources/test_cloudstack.py\n@@ -244,6 +244,13 @@ class TestCloudStackHostname(CiTestCase):\n )\n )\n \n+ self.patches.enter_context(\n+ mock.patch(\n+ \"cloudinit.distros.net.find_fallback_nic\",\n+ return_value=\"eth0\",\n+ )\n+ )\n+\n self.patches.enter_context(\n mock.patch(\n MOD_PATH\n@@ -304,7 +311,7 @@ class TestCloudStackHostname(CiTestCase):\n )\n \n ds = DataSourceCloudStack(\n- {}, MockDistro(), helpers.Paths({\"run_dir\": self.tmp})\n+ {}, ubuntu.Distro(\"\", {}, {}), helpers.Paths({\"run_dir\": self.tmp})\n )\n ds._fallback_interface = \"eth0\"\n with mock.patch(MOD_PATH + \".util.load_file\"):\ndiff --git a/tests/unittests/sources/test_ec2.py b/tests/unittests/sources/test_ec2.py\nindex 5f60ad737..32fed89da 100644\n--- a/tests/unittests/sources/test_ec2.py\n+++ b/tests/unittests/sources/test_ec2.py\n@@ -10,6 +10,7 @@ import requests\n import responses\n \n from cloudinit import helpers\n+from cloudinit.distros import ubuntu\n from cloudinit.sources import DataSourceEc2 as ec2\n from tests.unittests import helpers as test_helpers\n \n@@ -342,9 +343,11 @@ class TestEc2(test_helpers.ResponsesTestCase):\n p.start()\n self.addCleanup(p.stop)\n \n- def _setup_ds(self, sys_cfg, platform_data, md, md_version=None):\n+ def _setup_ds(\n+ self, sys_cfg, platform_data, md, md_version=None, distro=None\n+ ):\n self.uris = []\n- distro = mock.MagicMock()\n+ distro = distro or mock.MagicMock()\n distro.get_tmp_exec_path = self.tmp_dir\n paths = helpers.Paths({\"run_dir\": self.tmp})\n if sys_cfg is None:\n@@ -846,7 +849,7 @@ class TestEc2(test_helpers.ResponsesTestCase):\n \n @mock.patch(\"cloudinit.net.ephemeral.EphemeralIPv6Network\")\n @mock.patch(\"cloudinit.net.ephemeral.EphemeralIPv4Network\")\n- @mock.patch(\"cloudinit.net.find_fallback_nic\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\")\n @mock.patch(\"cloudinit.net.ephemeral.maybe_perform_dhcp_discovery\")\n @mock.patch(\"cloudinit.sources.DataSourceEc2.util.is_FreeBSD\")\n def test_ec2_local_performs_dhcp_on_non_bsd(\n@@ -873,6 +876,7 @@ class TestEc2(test_helpers.ResponsesTestCase):\n platform_data=self.valid_platform_data,\n sys_cfg={\"datasource\": {\"Ec2\": {\"strict_id\": False}}},\n md={\"md\": DEFAULT_METADATA},\n+ distro=ubuntu.Distro(\"\", {}, {}),\n )\n \n ret = ds.get_data()\ndiff --git a/tests/unittests/sources/test_init.py b/tests/unittests/sources/test_init.py\nindex 44d63b816..b3764945d 100644\n--- a/tests/unittests/sources/test_init.py\n+++ b/tests/unittests/sources/test_init.py\n@@ -6,6 +6,7 @@ import os\n import stat\n \n from cloudinit import importer, util\n+from cloudinit.distros import ubuntu\n from cloudinit.event import EventScope, EventType\n from cloudinit.helpers import Paths\n from cloudinit.sources import (\n@@ -73,7 +74,7 @@ class TestDataSource(CiTestCase):\n def setUp(self):\n super(TestDataSource, self).setUp()\n self.sys_cfg = {\"datasource\": {\"_undef\": {\"key1\": False}}}\n- self.distro = \"distrotest\" # generally should be a Distro object\n+ self.distro = ubuntu.Distro(\"\", {}, {})\n self.paths = Paths({})\n self.datasource = DataSource(self.sys_cfg, self.distro, self.paths)\n \n@@ -201,7 +202,7 @@ class TestDataSource(CiTestCase):\n for log in expected_logs:\n self.assertIn(log, logs)\n \n- @mock.patch(\"cloudinit.sources.net.find_fallback_nic\")\n+ @mock.patch(\"cloudinit.distros.net.find_fallback_nic\")\n def test_fallback_interface_is_discovered(self, m_get_fallback_nic):\n \"\"\"The fallback_interface is discovered via find_fallback_nic.\"\"\"\n m_get_fallback_nic.return_value = \"nic9\"\n@@ -221,7 +222,7 @@ class TestDataSource(CiTestCase):\n @mock.patch(\"cloudinit.sources.net.find_fallback_nic\")\n def test_wb_fallback_interface_is_cached(self, m_get_fallback_nic):\n \"\"\"The fallback_interface is cached and won't be rediscovered.\"\"\"\n- self.datasource._fallback_interface = \"nic10\"\n+ self.datasource.distro._fallback_interface = \"nic10\"\n self.assertEqual(\"nic10\", self.datasource.fallback_interface)\n m_get_fallback_nic.assert_not_called()\n \ndiff --git a/tests/unittests/sources/test_openstack.py b/tests/unittests/sources/test_openstack.py\nindex 127123cff..97cc8c94e 100644\n--- a/tests/unittests/sources/test_openstack.py\n+++ b/tests/unittests/sources/test_openstack.py\n@@ -338,7 +338,7 @@ class TestOpenStackDataSource(test_helpers.ResponsesTestCase):\n ds_os_local = ds.DataSourceOpenStackLocal(\n settings.CFG_BUILTIN, distro, helpers.Paths({\"run_dir\": self.tmp})\n )\n- ds_os_local._fallback_interface = \"eth9\" # Monkey patch for dhcp\n+ distro.fallback_interface = \"eth9\" # Monkey patch for dhcp\n m_dhcp.return_value = {\n \"interface\": \"eth9\",\n \"fixed-address\": \"192.168.2.9\",\n", "difficulty": 0, "changed_files": ["cloudinit/distros/__init__.py", "cloudinit/net/dhcp.py", "cloudinit/sources/__init__.py", "tests/unittests/net/test_dhcp.py", "tests/unittests/sources/test_cloudstack.py", "tests/unittests/sources/test_ec2.py", "tests/unittests/sources/test_init.py", "tests/unittests/sources/test_openstack.py"], "commit_link": "https://github.com/canonical/cloud-init/tree/4d5898b8a73c93e1ed4434744c2fa7c3f7fbd501"}
|
|
|
|