diff --git "a/Training script and logs.ipynb" "b/Training script and logs.ipynb" new file mode 100644--- /dev/null +++ "b/Training script and logs.ipynb" @@ -0,0 +1,17940 @@ +{ + "cells": [ + { + "cell_type": "code", + "execution_count": 1, + "metadata": { + "id": "qiaua4b79Z3o" + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Looking in indexes: https://download.pytorch.org/whl/cu121\n", + "Collecting torch==2.2.0\n", + " Downloading https://download.pytorch.org/whl/cu121/torch-2.2.0%2Bcu121-cp310-cp310-linux_x86_64.whl (757.3 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m757.3/757.3 MB\u001b[0m \u001b[31m68.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting triton\n", + " Downloading https://download.pytorch.org/whl/triton-2.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (168.1 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m168.1/168.1 MB\u001b[0m \u001b[31m112.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting filelock (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/filelock-3.13.1-py3-none-any.whl (11 kB)\n", + "Collecting typing-extensions>=4.8.0 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/typing_extensions-4.9.0-py3-none-any.whl (32 kB)\n", + "Collecting sympy (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/sympy-1.12-py3-none-any.whl (5.7 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m5.7/5.7 MB\u001b[0m \u001b[31m121.9 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0ma \u001b[36m0:00:01\u001b[0m\n", + "\u001b[?25hCollecting networkx (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/networkx-3.2.1-py3-none-any.whl (1.6 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m1.6/1.6 MB\u001b[0m \u001b[31m57.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting jinja2 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/Jinja2-3.1.3-py3-none-any.whl (133 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m133.2/133.2 kB\u001b[0m \u001b[31m169.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting fsspec (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/fsspec-2024.2.0-py3-none-any.whl (170 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m170.9/170.9 kB\u001b[0m \u001b[31m302.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cuda-nvrtc-cu12==12.1.105 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_nvrtc_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (23.7 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m23.7/23.7 MB\u001b[0m \u001b[31m109.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cuda-runtime-cu12==12.1.105 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_runtime_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (823 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m823.6/823.6 kB\u001b[0m \u001b[31m179.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cuda-cupti-cu12==12.1.105 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cuda_cupti_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (14.1 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m14.1/14.1 MB\u001b[0m \u001b[31m114.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0ma \u001b[36m0:00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cudnn-cu12==8.9.2.26 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cudnn_cu12-8.9.2.26-py3-none-manylinux1_x86_64.whl (731.7 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m731.7/731.7 MB\u001b[0m \u001b[31m115.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cublas-cu12==12.1.3.1 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cublas_cu12-12.1.3.1-py3-none-manylinux1_x86_64.whl (410.6 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m410.6/410.6 MB\u001b[0m \u001b[31m111.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cufft-cu12==11.0.2.54 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cufft_cu12-11.0.2.54-py3-none-manylinux1_x86_64.whl (121.6 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m121.6/121.6 MB\u001b[0m \u001b[31m113.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-curand-cu12==10.3.2.106 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_curand_cu12-10.3.2.106-py3-none-manylinux1_x86_64.whl (56.5 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m56.5/56.5 MB\u001b[0m \u001b[31m114.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cusolver-cu12==11.4.5.107 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cusolver_cu12-11.4.5.107-py3-none-manylinux1_x86_64.whl (124.2 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m124.2/124.2 MB\u001b[0m \u001b[31m113.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-cusparse-cu12==12.1.0.106 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_cusparse_cu12-12.1.0.106-py3-none-manylinux1_x86_64.whl (196.0 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m196.0/196.0 MB\u001b[0m \u001b[31m112.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-nccl-cu12==2.19.3 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_nccl_cu12-2.19.3-py3-none-manylinux1_x86_64.whl (166.0 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m166.0/166.0 MB\u001b[0m \u001b[31m109.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-nvtx-cu12==12.1.105 (from torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_nvtx_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (99 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m99.1/99.1 kB\u001b[0m \u001b[31m456.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting triton\n", + " Downloading https://download.pytorch.org/whl/triton-2.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (167.9 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m167.9/167.9 MB\u001b[0m \u001b[31m114.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting nvidia-nvjitlink-cu12 (from nvidia-cusolver-cu12==11.4.5.107->torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/cu121/nvidia_nvjitlink_cu12-12.1.105-py3-none-manylinux1_x86_64.whl (19.8 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m19.8/19.8 MB\u001b[0m \u001b[31m110.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hCollecting MarkupSafe>=2.0 (from jinja2->torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/MarkupSafe-2.1.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (25 kB)\n", + "Collecting mpmath>=0.19 (from sympy->torch==2.2.0)\n", + " Downloading https://download.pytorch.org/whl/mpmath-1.3.0-py3-none-any.whl (536 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m536.2/536.2 kB\u001b[0m \u001b[31m233.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hInstalling collected packages: mpmath, typing-extensions, sympy, nvidia-nvtx-cu12, nvidia-nvjitlink-cu12, nvidia-nccl-cu12, nvidia-curand-cu12, nvidia-cufft-cu12, nvidia-cuda-runtime-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-cupti-cu12, nvidia-cublas-cu12, networkx, MarkupSafe, fsspec, filelock, triton, nvidia-cusparse-cu12, nvidia-cudnn-cu12, jinja2, nvidia-cusolver-cu12, torch\n", + " Attempting uninstall: mpmath\n", + " Found existing installation: mpmath 1.3.0\n", + " Uninstalling mpmath-1.3.0:\n", + " Successfully uninstalled mpmath-1.3.0\n", + " Attempting uninstall: typing-extensions\n", + " Found existing installation: typing_extensions 4.9.0\n", + " Uninstalling typing_extensions-4.9.0:\n", + " Successfully uninstalled typing_extensions-4.9.0\n", + " Attempting uninstall: sympy\n", + " Found existing installation: sympy 1.12\n", + " Uninstalling sympy-1.12:\n", + " Successfully uninstalled sympy-1.12\n", + " Attempting uninstall: nvidia-nvtx-cu12\n", + " Found existing installation: nvidia-nvtx-cu12 12.1.105\n", + " Uninstalling nvidia-nvtx-cu12-12.1.105:\n", + " Successfully uninstalled nvidia-nvtx-cu12-12.1.105\n", + " Attempting uninstall: nvidia-nvjitlink-cu12\n", + " Found existing installation: nvidia-nvjitlink-cu12 12.3.101\n", + " Uninstalling nvidia-nvjitlink-cu12-12.3.101:\n", + " Successfully uninstalled nvidia-nvjitlink-cu12-12.3.101\n", + " Attempting uninstall: nvidia-nccl-cu12\n", + " Found existing installation: nvidia-nccl-cu12 2.19.3\n", + " Uninstalling nvidia-nccl-cu12-2.19.3:\n", + " Successfully uninstalled nvidia-nccl-cu12-2.19.3\n", + " Attempting uninstall: nvidia-curand-cu12\n", + " Found existing installation: nvidia-curand-cu12 10.3.2.106\n", + " Uninstalling nvidia-curand-cu12-10.3.2.106:\n", + " Successfully uninstalled nvidia-curand-cu12-10.3.2.106\n", + " Attempting uninstall: nvidia-cufft-cu12\n", + " Found existing installation: nvidia-cufft-cu12 11.0.2.54\n", + " Uninstalling nvidia-cufft-cu12-11.0.2.54:\n", + " Successfully uninstalled nvidia-cufft-cu12-11.0.2.54\n", + " Attempting uninstall: nvidia-cuda-runtime-cu12\n", + " Found existing installation: nvidia-cuda-runtime-cu12 12.1.105\n", + " Uninstalling nvidia-cuda-runtime-cu12-12.1.105:\n", + " Successfully uninstalled nvidia-cuda-runtime-cu12-12.1.105\n", + " Attempting uninstall: nvidia-cuda-nvrtc-cu12\n", + " Found existing installation: nvidia-cuda-nvrtc-cu12 12.1.105\n", + " Uninstalling nvidia-cuda-nvrtc-cu12-12.1.105:\n", + " Successfully uninstalled nvidia-cuda-nvrtc-cu12-12.1.105\n", + " Attempting uninstall: nvidia-cuda-cupti-cu12\n", + " Found existing installation: nvidia-cuda-cupti-cu12 12.1.105\n", + " Uninstalling nvidia-cuda-cupti-cu12-12.1.105:\n", + " Successfully uninstalled nvidia-cuda-cupti-cu12-12.1.105\n", + " Attempting uninstall: nvidia-cublas-cu12\n", + " Found existing installation: nvidia-cublas-cu12 12.1.3.1\n", + " Uninstalling nvidia-cublas-cu12-12.1.3.1:\n", + " Successfully uninstalled nvidia-cublas-cu12-12.1.3.1\n", + " Attempting uninstall: networkx\n", + " Found existing installation: networkx 3.2.1\n", + " Uninstalling networkx-3.2.1:\n", + " Successfully uninstalled networkx-3.2.1\n", + " Attempting uninstall: MarkupSafe\n", + " Found existing installation: MarkupSafe 2.1.5\n", + " Uninstalling MarkupSafe-2.1.5:\n", + " Successfully uninstalled MarkupSafe-2.1.5\n", + " Attempting uninstall: fsspec\n", + " Found existing installation: fsspec 2024.2.0\n", + " Uninstalling fsspec-2024.2.0:\n", + " Successfully uninstalled fsspec-2024.2.0\n", + " Attempting uninstall: filelock\n", + " Found existing installation: filelock 3.13.1\n", + " Uninstalling filelock-3.13.1:\n", + " Successfully uninstalled filelock-3.13.1\n", + " Attempting uninstall: triton\n", + " Found existing installation: triton 2.2.0\n", + " Uninstalling triton-2.2.0:\n", + " Successfully uninstalled triton-2.2.0\n", + " Attempting uninstall: nvidia-cusparse-cu12\n", + " Found existing installation: nvidia-cusparse-cu12 12.1.0.106\n", + " Uninstalling nvidia-cusparse-cu12-12.1.0.106:\n", + " Successfully uninstalled nvidia-cusparse-cu12-12.1.0.106\n", + " Attempting uninstall: nvidia-cudnn-cu12\n", + " Found existing installation: nvidia-cudnn-cu12 8.9.2.26\n", + " Uninstalling nvidia-cudnn-cu12-8.9.2.26:\n", + " Successfully uninstalled nvidia-cudnn-cu12-8.9.2.26\n", + " Attempting uninstall: jinja2\n", + " Found existing installation: Jinja2 3.1.3\n", + " Uninstalling Jinja2-3.1.3:\n", + " Successfully uninstalled Jinja2-3.1.3\n", + " Attempting uninstall: nvidia-cusolver-cu12\n", + " Found existing installation: nvidia-cusolver-cu12 11.4.5.107\n", + " Uninstalling nvidia-cusolver-cu12-11.4.5.107:\n", + " Successfully uninstalled nvidia-cusolver-cu12-11.4.5.107\n", + " Attempting uninstall: torch\n", + " Found existing installation: torch 2.2.0\n", + " Uninstalling torch-2.2.0:\n", + " Successfully uninstalled torch-2.2.0\n", + "Successfully installed MarkupSafe-2.1.5 filelock-3.13.1 fsspec-2024.2.0 jinja2-3.1.3 mpmath-1.3.0 networkx-3.2.1 nvidia-cublas-cu12-12.1.3.1 nvidia-cuda-cupti-cu12-12.1.105 nvidia-cuda-nvrtc-cu12-12.1.105 nvidia-cuda-runtime-cu12-12.1.105 nvidia-cudnn-cu12-8.9.2.26 nvidia-cufft-cu12-11.0.2.54 nvidia-curand-cu12-10.3.2.106 nvidia-cusolver-cu12-11.4.5.107 nvidia-cusparse-cu12-12.1.0.106 nvidia-nccl-cu12-2.19.3 nvidia-nvjitlink-cu12-12.1.105 nvidia-nvtx-cu12-12.1.105 sympy-1.12 torch-2.2.0+cu121 triton-2.2.0 typing-extensions-4.9.0\n", + "\u001b[33mWARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv\u001b[0m\u001b[33m\n", + "\u001b[0mCollecting unsloth@ git+https://github.com/unslothai/unsloth.git (from unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Cloning https://github.com/unslothai/unsloth.git to /tmp/pip-install-98le95ye/unsloth_b49c98c9acd24e649e98f169c8e0a660\n", + " Running command git clone --filter=blob:none --quiet https://github.com/unslothai/unsloth.git /tmp/pip-install-98le95ye/unsloth_b49c98c9acd24e649e98f169c8e0a660\n", + " Resolved https://github.com/unslothai/unsloth.git to commit 92dce38e8b3c1db209cef860d90b60188e95f0f9\n", + " Installing build dependencies ... \u001b[?25ldone\n", + "\u001b[?25h Getting requirements to build wheel ... \u001b[?25ldone\n", + "\u001b[?25h Preparing metadata (pyproject.toml) ... \u001b[?25ldone\n", + "\u001b[?25hCollecting bitsandbytes (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading bitsandbytes-0.43.1-py3-none-manylinux_2_24_x86_64.whl.metadata (2.2 kB)\n", + "Requirement already satisfied: torch in /usr/local/lib/python3.10/dist-packages (from bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.2.0+cu121)\n", + "Requirement already satisfied: numpy in /usr/local/lib/python3.10/dist-packages (from bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (1.26.3)\n", + "Collecting xformers@ https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp310-cp310-manylinux2014_x86_64.whl (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading https://download.pytorch.org/whl/cu121/xformers-0.0.24-cp310-cp310-manylinux2014_x86_64.whl (218.2 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m218.2/218.2 MB\u001b[0m \u001b[31m35.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hRequirement already satisfied: packaging in /usr/local/lib/python3.10/dist-packages (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (23.2)\n", + "Collecting tyro (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading tyro-0.8.5-py3-none-any.whl.metadata (8.2 kB)\n", + "Collecting transformers>=4.42.3 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading transformers-4.42.4-py3-none-any.whl.metadata (43 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m43.6/43.6 kB\u001b[0m \u001b[31m1.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting datasets>=2.16.0 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading datasets-2.20.0-py3-none-any.whl.metadata (19 kB)\n", + "Collecting sentencepiece>=0.2.0 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)\n", + "Collecting tqdm (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading tqdm-4.66.4-py3-none-any.whl.metadata (57 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m57.6/57.6 kB\u001b[0m \u001b[31m9.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hRequirement already satisfied: psutil in /usr/local/lib/python3.10/dist-packages (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (5.9.8)\n", + "Requirement already satisfied: wheel>=0.42.0 in /usr/local/lib/python3.10/dist-packages (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (0.42.0)\n", + "Collecting accelerate>=0.26.1 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading accelerate-0.32.1-py3-none-any.whl.metadata (18 kB)\n", + "Collecting trl<0.9.0,>=0.7.9 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading trl-0.8.6-py3-none-any.whl.metadata (11 kB)\n", + "Collecting peft!=0.11.0,>=0.7.1 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading peft-0.11.1-py3-none-any.whl.metadata (13 kB)\n", + "Collecting protobuf<4.0.0 (from unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading protobuf-3.20.3-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl.metadata (679 bytes)\n", + "Requirement already satisfied: pyyaml in /usr/local/lib/python3.10/dist-packages (from accelerate>=0.26.1->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (6.0.1)\n", + "Collecting huggingface-hub (from accelerate>=0.26.1->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading huggingface_hub-0.23.4-py3-none-any.whl.metadata (12 kB)\n", + "Collecting safetensors>=0.3.1 (from accelerate>=0.26.1->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading safetensors-0.4.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (3.8 kB)\n", + "Requirement already satisfied: filelock in /usr/local/lib/python3.10/dist-packages (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (3.13.1)\n", + "Collecting pyarrow>=15.0.0 (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading pyarrow-16.1.0-cp310-cp310-manylinux_2_28_x86_64.whl.metadata (3.0 kB)\n", + "Collecting pyarrow-hotfix (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading pyarrow_hotfix-0.6-py3-none-any.whl.metadata (3.6 kB)\n", + "Collecting dill<0.3.9,>=0.3.0 (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading dill-0.3.8-py3-none-any.whl.metadata (10 kB)\n", + "Collecting pandas (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading pandas-2.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (19 kB)\n", + "Collecting requests>=2.32.2 (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading requests-2.32.3-py3-none-any.whl.metadata (4.6 kB)\n", + "Collecting xxhash (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading xxhash-3.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (12 kB)\n", + "Collecting multiprocess (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading multiprocess-0.70.16-py310-none-any.whl.metadata (7.2 kB)\n", + "Requirement already satisfied: fsspec<=2024.5.0,>=2023.1.0 in /usr/local/lib/python3.10/dist-packages (from fsspec[http]<=2024.5.0,>=2023.1.0->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2024.2.0)\n", + "Collecting aiohttp (from datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading aiohttp-3.9.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.5 kB)\n", + "Requirement already satisfied: typing-extensions>=4.8.0 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (4.9.0)\n", + "Requirement already satisfied: sympy in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (1.12)\n", + "Requirement already satisfied: networkx in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (3.2.1)\n", + "Requirement already satisfied: jinja2 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (3.1.3)\n", + "Requirement already satisfied: nvidia-cuda-nvrtc-cu12==12.1.105 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.105)\n", + "Requirement already satisfied: nvidia-cuda-runtime-cu12==12.1.105 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.105)\n", + "Requirement already satisfied: nvidia-cuda-cupti-cu12==12.1.105 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.105)\n", + "Requirement already satisfied: nvidia-cudnn-cu12==8.9.2.26 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (8.9.2.26)\n", + "Requirement already satisfied: nvidia-cublas-cu12==12.1.3.1 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.3.1)\n", + "Requirement already satisfied: nvidia-cufft-cu12==11.0.2.54 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (11.0.2.54)\n", + "Requirement already satisfied: nvidia-curand-cu12==10.3.2.106 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (10.3.2.106)\n", + "Requirement already satisfied: nvidia-cusolver-cu12==11.4.5.107 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (11.4.5.107)\n", + "Requirement already satisfied: nvidia-cusparse-cu12==12.1.0.106 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.0.106)\n", + "Requirement already satisfied: nvidia-nccl-cu12==2.19.3 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.19.3)\n", + "Requirement already satisfied: nvidia-nvtx-cu12==12.1.105 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.105)\n", + "Requirement already satisfied: triton==2.2.0 in /usr/local/lib/python3.10/dist-packages (from torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.2.0)\n", + "Requirement already satisfied: nvidia-nvjitlink-cu12 in /usr/local/lib/python3.10/dist-packages (from nvidia-cusolver-cu12==11.4.5.107->torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (12.1.105)\n", + "Collecting regex!=2019.12.17 (from transformers>=4.42.3->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading regex-2024.5.15-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (40 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m40.9/40.9 kB\u001b[0m \u001b[31m13.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hCollecting tokenizers<0.20,>=0.19 (from transformers>=4.42.3->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading tokenizers-0.19.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (6.7 kB)\n", + "Collecting docstring-parser>=0.16 (from tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading docstring_parser-0.16-py3-none-any.whl.metadata (3.0 kB)\n", + "Collecting rich>=11.1.0 (from tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading rich-13.7.1-py3-none-any.whl.metadata (18 kB)\n", + "Collecting shtab>=1.5.6 (from tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading shtab-1.7.1-py3-none-any.whl.metadata (7.3 kB)\n", + "Collecting aiosignal>=1.1.2 (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading aiosignal-1.3.1-py3-none-any.whl.metadata (4.0 kB)\n", + "Requirement already satisfied: attrs>=17.3.0 in /usr/local/lib/python3.10/dist-packages (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (23.2.0)\n", + "Collecting frozenlist>=1.1.1 (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading frozenlist-1.4.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (12 kB)\n", + "Collecting multidict<7.0,>=4.5 (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading multidict-6.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.2 kB)\n", + "Collecting yarl<2.0,>=1.0 (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading yarl-1.9.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (31 kB)\n", + "Collecting async-timeout<5.0,>=4.0 (from aiohttp->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading async_timeout-4.0.3-py3-none-any.whl.metadata (4.2 kB)\n", + "Requirement already satisfied: charset-normalizer<4,>=2 in /usr/local/lib/python3.10/dist-packages (from requests>=2.32.2->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (3.3.2)\n", + "Requirement already satisfied: idna<4,>=2.5 in /usr/local/lib/python3.10/dist-packages (from requests>=2.32.2->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (3.6)\n", + "Requirement already satisfied: urllib3<3,>=1.21.1 in /usr/local/lib/python3.10/dist-packages (from requests>=2.32.2->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.2.0)\n", + "Requirement already satisfied: certifi>=2017.4.17 in /usr/local/lib/python3.10/dist-packages (from requests>=2.32.2->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2024.2.2)\n", + "Collecting markdown-it-py>=2.2.0 (from rich>=11.1.0->tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading markdown_it_py-3.0.0-py3-none-any.whl.metadata (6.9 kB)\n", + "Requirement already satisfied: pygments<3.0.0,>=2.13.0 in /usr/local/lib/python3.10/dist-packages (from rich>=11.1.0->tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.17.2)\n", + "Requirement already satisfied: MarkupSafe>=2.0 in /usr/local/lib/python3.10/dist-packages (from jinja2->torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.1.5)\n", + "Requirement already satisfied: python-dateutil>=2.8.2 in /usr/local/lib/python3.10/dist-packages (from pandas->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (2.8.2)\n", + "Collecting pytz>=2020.1 (from pandas->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading pytz-2024.1-py2.py3-none-any.whl.metadata (22 kB)\n", + "Collecting tzdata>=2022.7 (from pandas->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading tzdata-2024.1-py2.py3-none-any.whl.metadata (1.4 kB)\n", + "Requirement already satisfied: mpmath>=0.19 in /usr/local/lib/python3.10/dist-packages (from sympy->torch->bitsandbytes->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (1.3.0)\n", + "Collecting mdurl~=0.1 (from markdown-it-py>=2.2.0->rich>=11.1.0->tyro->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git)\n", + " Downloading mdurl-0.1.2-py3-none-any.whl.metadata (1.6 kB)\n", + "Requirement already satisfied: six>=1.5 in /usr/lib/python3/dist-packages (from python-dateutil>=2.8.2->pandas->datasets>=2.16.0->unsloth@ git+https://github.com/unslothai/unsloth.git->unsloth[cu121-torch220]@ git+https://github.com/unslothai/unsloth.git) (1.16.0)\n", + "Downloading bitsandbytes-0.43.1-py3-none-manylinux_2_24_x86_64.whl (119.8 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m119.8/119.8 MB\u001b[0m \u001b[31m66.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hDownloading accelerate-0.32.1-py3-none-any.whl (314 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m314.1/314.1 kB\u001b[0m \u001b[31m81.5 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading datasets-2.20.0-py3-none-any.whl (547 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m547.8/547.8 kB\u001b[0m \u001b[31m198.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading peft-0.11.1-py3-none-any.whl (251 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m251.6/251.6 kB\u001b[0m \u001b[31m124.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading protobuf-3.20.3-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (1.1 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m1.1/1.1 MB\u001b[0m \u001b[31m128.2 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading sentencepiece-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.3 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m1.3/1.3 MB\u001b[0m \u001b[31m120.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading tqdm-4.66.4-py3-none-any.whl (78 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m78.3/78.3 kB\u001b[0m \u001b[31m31.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading transformers-4.42.4-py3-none-any.whl (9.3 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m9.3/9.3 MB\u001b[0m \u001b[31m121.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hDownloading trl-0.8.6-py3-none-any.whl (245 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m245.2/245.2 kB\u001b[0m \u001b[31m116.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading tyro-0.8.5-py3-none-any.whl (103 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m103.4/103.4 kB\u001b[0m \u001b[31m49.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading dill-0.3.8-py3-none-any.whl (116 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m116.3/116.3 kB\u001b[0m \u001b[31m49.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading docstring_parser-0.16-py3-none-any.whl (36 kB)\n", + "Downloading aiohttp-3.9.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.2 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m1.2/1.2 MB\u001b[0m \u001b[31m236.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading huggingface_hub-0.23.4-py3-none-any.whl (402 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m402.6/402.6 kB\u001b[0m \u001b[31m147.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading pyarrow-16.1.0-cp310-cp310-manylinux_2_28_x86_64.whl (40.8 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m40.8/40.8 MB\u001b[0m \u001b[31m91.0 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m:00:01\u001b[0m00:01\u001b[0m\n", + "\u001b[?25hDownloading regex-2024.5.15-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (775 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m775.1/775.1 kB\u001b[0m \u001b[31m168.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading requests-2.32.3-py3-none-any.whl (64 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m64.9/64.9 kB\u001b[0m \u001b[31m37.7 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading rich-13.7.1-py3-none-any.whl (240 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m240.7/240.7 kB\u001b[0m \u001b[31m119.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading safetensors-0.4.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.2 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m1.2/1.2 MB\u001b[0m \u001b[31m180.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading shtab-1.7.1-py3-none-any.whl (14 kB)\n", + "Downloading tokenizers-0.19.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (3.6 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m3.6/3.6 MB\u001b[0m \u001b[31m137.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading multiprocess-0.70.16-py310-none-any.whl (134 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m134.8/134.8 kB\u001b[0m \u001b[31m79.6 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading pandas-2.2.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.0 MB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m13.0/13.0 MB\u001b[0m \u001b[31m74.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m00:01\u001b[0m0:01\u001b[0mm\n", + "\u001b[?25hDownloading pyarrow_hotfix-0.6-py3-none-any.whl (7.9 kB)\n", + "Downloading xxhash-3.4.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (194 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m194.1/194.1 kB\u001b[0m \u001b[31m91.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading aiosignal-1.3.1-py3-none-any.whl (7.6 kB)\n", + "Downloading async_timeout-4.0.3-py3-none-any.whl (5.7 kB)\n", + "Downloading frozenlist-1.4.1-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (239 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m239.5/239.5 kB\u001b[0m \u001b[31m76.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading markdown_it_py-3.0.0-py3-none-any.whl (87 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m87.5/87.5 kB\u001b[0m \u001b[31m59.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading multidict-6.0.5-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (124 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m124.3/124.3 kB\u001b[0m \u001b[31m60.1 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading pytz-2024.1-py2.py3-none-any.whl (505 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m505.5/505.5 kB\u001b[0m \u001b[31m134.4 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading tzdata-2024.1-py2.py3-none-any.whl (345 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m345.4/345.4 kB\u001b[0m \u001b[31m124.8 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading yarl-1.9.4-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (301 kB)\n", + "\u001b[2K \u001b[90m━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━\u001b[0m \u001b[32m301.6/301.6 kB\u001b[0m \u001b[31m101.3 MB/s\u001b[0m eta \u001b[36m0:00:00\u001b[0m\n", + "\u001b[?25hDownloading mdurl-0.1.2-py3-none-any.whl (10.0 kB)\n", + "Building wheels for collected packages: unsloth\n", + " Building wheel for unsloth (pyproject.toml) ... \u001b[?25ldone\n", + "\u001b[?25h Created wheel for unsloth: filename=unsloth-2024.7-py3-none-any.whl size=128252 sha256=80ddf176bb030b5c81da1bd69bef4fe3bc766d4442a4a1967e10baf875f2731d\n", + " Stored in directory: /tmp/pip-ephem-wheel-cache-g5c37kdw/wheels/ed/d4/e9/76fb290ee3df0a5fc21ce5c2c788e29e9607a2353d8342fd0d\n", + "Successfully built unsloth\n", + "Installing collected packages: sentencepiece, pytz, xxhash, unsloth, tzdata, tqdm, shtab, safetensors, requests, regex, pyarrow-hotfix, pyarrow, protobuf, multidict, mdurl, frozenlist, docstring-parser, dill, async-timeout, yarl, pandas, multiprocess, markdown-it-py, huggingface-hub, aiosignal, tokenizers, rich, aiohttp, xformers, tyro, transformers, bitsandbytes, accelerate, peft, datasets, trl\n", + " Attempting uninstall: requests\n", + " Found existing installation: requests 2.31.0\n", + " Uninstalling requests-2.31.0:\n", + " Successfully uninstalled requests-2.31.0\n", + "Successfully installed accelerate-0.32.1 aiohttp-3.9.5 aiosignal-1.3.1 async-timeout-4.0.3 bitsandbytes-0.43.1 datasets-2.20.0 dill-0.3.8 docstring-parser-0.16 frozenlist-1.4.1 huggingface-hub-0.23.4 markdown-it-py-3.0.0 mdurl-0.1.2 multidict-6.0.5 multiprocess-0.70.16 pandas-2.2.2 peft-0.11.1 protobuf-3.20.3 pyarrow-16.1.0 pyarrow-hotfix-0.6 pytz-2024.1 regex-2024.5.15 requests-2.32.3 rich-13.7.1 safetensors-0.4.3 sentencepiece-0.2.0 shtab-1.7.1 tokenizers-0.19.1 tqdm-4.66.4 transformers-4.42.4 trl-0.8.6 tyro-0.8.5 tzdata-2024.1 unsloth-2024.7 xformers-0.0.24 xxhash-3.4.1 yarl-1.9.4\n", + "\u001b[33mWARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv\u001b[0m\u001b[33m\n", + "\u001b[0m\n", + "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m A new release of pip is available: \u001b[0m\u001b[31;49m24.0\u001b[0m\u001b[39;49m -> \u001b[0m\u001b[32;49m24.1.2\u001b[0m\n", + "\u001b[1m[\u001b[0m\u001b[34;49mnotice\u001b[0m\u001b[1;39;49m]\u001b[0m\u001b[39;49m To update, run: \u001b[0m\u001b[32;49mpython -m pip install --upgrade pip\u001b[0m\n" + ] + } + ], + "source": [ + "#@title If you are running on a **GPU sever**, run this\n", + "!pip install --upgrade --force-reinstall --no-cache-dir torch==2.2.0 triton --index-url https://download.pytorch.org/whl/cu121\n", + "!pip install \"unsloth[cu121-torch220] @ git+https://github.com/unslothai/unsloth.git\"\n", + "!pip install hf_transfer" + ] + }, + { + "cell_type": "code", + "execution_count": 2, + "metadata": { + "id": "zHT-YJKy9XH0" + }, + "outputs": [ + { + "data": { + "text/plain": [ + "'12.1'" + ] + }, + "execution_count": 2, + "metadata": {}, + "output_type": "execute_result" + } + ], + "source": [ + "# Check the infomation of Cuda\n", + "import torch\n", + "torch.version.cuda" + ] + }, + { + "cell_type": "code", + "execution_count": 3, + "metadata": { + "id": "IO6wZW7Y90c8" + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "🦥 Unsloth: Will patch your computer to enable 2x faster free finetuning.\n" + ] + }, + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "7e39ad42dea14bb9bd3cb178ca1d6f06", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "config.json: 0%| | 0.00/912 [00:00 0 ! Suggested 8, 16, 32, 64, 128\n", + " target_modules = [\"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\",\n", + " \"gate_proj\", \"up_proj\", \"down_proj\",],\n", + " lora_alpha = 16,\n", + " lora_dropout = 0, # Supports any, but = 0 is optimized\n", + " bias = \"none\", # Supports any, but = \"none\" is optimized\n", + " # [NEW] \"unsloth\" uses 30% less VRAM, fits 2x larger batch sizes!\n", + " use_gradient_checkpointing = \"unsloth\", # True or \"unsloth\" for very long context\n", + " random_state = 3407,\n", + " use_rslora = False, # We support rank stabilized LoRA\n", + " loftq_config = None, # And LoftQ\n", + ")" + ] + }, + { + "cell_type": "code", + "execution_count": null, + "metadata": {}, + "outputs": [], + "source": [] + }, + { + "cell_type": "code", + "execution_count": 7, + "metadata": { + "id": "0rFDTzwS9_8a" + }, + "outputs": [ + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "bd1342e2285f45d0ab0b2786c6ae2f03", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "Downloading readme: 0%| | 0.00/31.0 [00:00\n" + ] + } + ], + "source": [ + "# Check if the dataset mapping correctly\n", + "print(dataset[5][\"text\"])" + ] + }, + { + "cell_type": "code", + "execution_count": 9, + "metadata": { + "id": "ibyPAgK--scy" + }, + "outputs": [ + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "64b66284ab3c4d8a9c72af527b701db4", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "Map (num_proc=8): 0%| | 0/54409 [00:00\n", + " \n", + " \n", + " [3400/3400 3:18:18, Epoch 0/1]\n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + " \n", + "
StepTraining Loss
14.035100
23.751900
33.281000
43.216900
53.194800
63.145400
73.596300
82.745300
92.922400
102.886800
112.337000
122.323700
131.917500
142.177400
151.833300
162.086600
171.796900
182.208500
191.607800
201.968300
211.881500
221.485800
231.704600
241.625100
251.554500
261.458900
271.584700
281.512300
291.667100
301.278700
311.481700
321.330300
331.291700
341.859400
351.224200
361.217700
371.123700
381.834300
391.679400
401.568800
411.282800
421.206200
431.497500
441.446700
451.593100
461.550600
471.495900
481.493400
491.136700
501.435400
511.259800
521.156600
531.335900
541.385200
551.209400
561.249200
571.423600
581.207100
591.499700
601.282900
611.448600
621.120800
631.244400
641.201800
651.427700
661.333200
671.345600
681.189500
691.407200
701.270100
711.268200
721.500500
730.904100
741.456200
751.306200
761.261200
771.356600
781.477800
791.534800
801.254400
811.228700
821.587500
831.593400
841.216400
851.151200
861.353200
871.160800
881.274400
891.267000
901.115400
911.037900
921.143900
931.428400
941.233300
951.159700
961.451400
971.166600
980.977400
991.372900
1001.562300
1011.110100
1021.307700
1031.478100
1041.360000
1051.010700
1061.458400
1071.063800
1081.068700
1091.327500
1101.356800
1111.197800
1121.237500
1131.151700
1141.114200
1151.215600
1161.361800
1171.269400
1181.216300
1191.324200
1201.045500
1211.043800
1221.203500
1231.131100
1241.175800
1251.228600
1261.376300
1271.396900
1281.434300
1290.940800
1301.348200
1311.173600
1321.286100
1331.204700
1341.077200
1351.468100
1361.201500
1371.272400
1381.223900
1391.138200
1401.297800
1411.189900
1421.482800
1431.121400
1441.274400
1451.551500
1460.946400
1471.377900
1481.102900
1491.256000
1501.536900
1511.146300
1521.019500
1531.122800
1541.293400
1551.348500
1561.195800
1571.306800
1581.289300
1591.419800
1601.155800
1611.120800
1621.110300
1631.105600
1641.220200
1651.286100
1661.372600
1671.215300
1681.016600
1691.371000
1701.362300
1711.059700
1721.115600
1731.374100
1741.022700
1751.060400
1761.123400
1771.614500
1781.100400
1791.504700
1801.103600
1811.074600
1820.946400
1831.199100
1841.237000
1851.099700
1861.331700
1871.062900
1881.012600
1891.303800
1901.445900
1911.452600
1921.360700
1931.212000
1941.317400
1951.194800
1961.180400
1971.300200
1981.178700
1991.286800
2001.213900
2011.151400
2021.033400
2031.412300
2041.454700
2051.485300
2061.193300
2071.057000
2080.981000
2091.316600
2101.290600
2111.217500
2121.630700
2131.371400
2141.097700
2151.503500
2160.952500
2170.867600
2181.062800
2191.070300
2200.819700
2211.348800
2221.472600
2230.961500
2241.412900
2251.369800
2261.252200
2271.325200
2281.287500
2291.066200
2300.995400
2311.293100
2321.347100
2331.211600
2341.083500
2351.488000
2361.090200
2371.298900
2381.294200
2390.846600
2400.906500
2411.194300
2421.038800
2431.104400
2441.140100
2451.250900
2461.005400
2471.062000
2480.924000
2491.254000
2501.188300
2511.318000
2521.350100
2531.202400
2541.203800
2550.859800
2561.443800
2571.359800
2581.048000
2591.044100
2601.489100
2611.015000
2621.149900
2631.022400
2641.146000
2651.239300
2661.392100
2671.008600
2681.369300
2691.254200
2701.321800
2711.034200
2721.523800
2731.059700
2740.873000
2751.159400
2761.105500
2771.600400
2781.191200
2791.026300
2801.257100
2810.885700
2820.974200
2831.136500
2841.182000
2850.895000
2861.309300
2871.271900
2881.266100
2891.035400
2901.162100
2911.292700
2921.317600
2931.111200
2941.418500
2951.189000
2961.202400
2971.266300
2981.122600
2991.285200
3001.365200
3011.251500
3021.279600
3031.006100
3041.227700
3051.811700
3061.106200
3071.216200
3080.940600
3091.109800
3101.395000
3111.131700
3121.479200
3131.031300
3140.903000
3151.212000
3161.015700
3171.149900
3181.033600
3190.956100
3201.393300
3211.175200
3221.141800
3231.481200
3241.249200
3251.368600
3261.111600
3270.986500
3280.859800
3291.132200
3301.109800
3311.052300
3321.246100
3331.327000
3341.048600
3351.096800
3361.390000
3370.955300
3381.160800
3390.955800
3401.113800
3411.189700
3421.255900
3431.074600
3441.069700
3451.242800
3461.090700
3471.292600
3480.890800
3491.150900
3501.092300
3511.209000
3521.206300
3531.323300
3541.134500
3551.088200
3561.062700
3571.136700
3581.235300
3590.874500
3601.104000
3611.001200
3620.896800
3631.332000
3641.587200
3651.038600
3660.961100
3671.079100
3680.940500
3691.046400
3701.030000
3711.154700
3721.102000
3731.329000
3741.347100
3750.989600
3760.943100
3771.192100
3781.079000
3791.074500
3801.051000
3811.197500
3821.453600
3830.989100
3840.980500
3851.299700
3861.335200
3871.210100
3880.975200
3891.216100
3901.070100
3911.110400
3921.271100
3931.103700
3941.124400
3951.230200
3960.995700
3970.946200
3980.923800
3991.030900
4001.341200
4011.015000
4021.025200
4031.065200
4041.180400
4051.012300
4060.979100
4071.231000
4081.268600
4090.962400
4100.855100
4110.989200
4121.238300
4131.117600
4141.297200
4151.045600
4161.192500
4171.279500
4180.927300
4191.195600
4201.309500
4211.217000
4221.240700
4231.480000
4240.949600
4251.020200
4260.984100
4271.197100
4281.088200
4291.324600
4301.004900
4310.950500
4321.254200
4331.119300
4341.473800
4350.907900
4361.169400
4371.101600
4381.142100
4391.002100
4401.204200
4411.119600
4420.940100
4431.063700
4440.945900
4451.205000
4461.017300
4471.426800
4480.818600
4491.228000
4501.175400
4511.314800
4521.035600
4530.881300
4541.201600
4550.823400
4561.120900
4570.986300
4581.010200
4591.206300
4601.114500
4611.265400
4621.232400
4630.904700
4641.454300
4651.027500
4661.112600
4671.103100
4681.239700
4691.250900
4700.768500
4711.045700
4721.150800
4730.969600
4741.003700
4751.104000
4760.938100
4771.293900
4781.053900
4791.083500
4801.042700
4811.030300
4821.579200
4831.247900
4841.156400
4851.287000
4861.019100
4871.290500
4881.034700
4891.052200
4901.209100
4911.080500
4921.169800
4930.987700
4940.964200
4951.253700
4961.290100
4971.044500
4980.996100
4990.981600
5000.932900
5010.976700
5021.117400
5030.895100
5041.056700
5051.005800
5061.159100
5071.057900
5081.276000
5090.994900
5101.335900
5111.401600
5121.180800
5131.016000
5141.065600
5151.425100
5160.961300
5171.094400
5181.207800
5191.194900
5201.093700
5210.938200
5220.989100
5231.062200
5241.170500
5251.119300
5261.171300
5271.357700
5281.126200
5290.960700
5301.042800
5310.973600
5321.207600
5331.162300
5341.403800
5351.168200
5361.217700
5371.269900
5381.283300
5391.399600
5401.253400
5411.172100
5421.101800
5431.256100
5441.228500
5450.879000
5461.060100
5471.209800
5480.857200
5490.753300
5501.428100
5510.972800
5521.298600
5531.031100
5541.007400
5551.044200
5561.112900
5570.889100
5581.156400
5591.022300
5601.101100
5610.948600
5621.081600
5631.138000
5641.050100
5651.084800
5661.087700
5671.141800
5681.176500
5691.119800
5701.167600
5711.199500
5721.011100
5731.398100
5741.048200
5751.272100
5761.628100
5771.049600
5781.146100
5790.923600
5801.297800
5811.051700
5821.061700
5830.984400
5841.203400
5850.929000
5860.956500
5871.323400
5881.079400
5890.879000
5901.006700
5910.836800
5921.023300
5931.305400
5941.106100
5951.060100
5961.499800
5971.283200
5981.193900
5991.220700
6001.078900
6011.247500
6021.184600
6031.318900
6041.414000
6050.967300
6061.218400
6071.045000
6080.965800
6091.132500
6101.137200
6110.931800
6121.361700
6130.890500
6141.050600
6150.981000
6161.239500
6171.412800
6181.200500
6191.297300
6200.952600
6211.070500
6221.096600
6231.357300
6241.180400
6251.139200
6261.203700
6271.052800
6281.122700
6291.169700
6301.221800
6311.154900
6320.984100
6331.118500
6341.428800
6351.049100
6361.114000
6371.260900
6381.120200
6390.980800
6401.097200
6411.134700
6420.840100
6431.073600
6441.102400
6451.322100
6461.132300
6471.017600
6481.125000
6490.939200
6501.086800
6511.123300
6521.098100
6530.983200
6541.230200
6551.094800
6561.126800
6571.137700
6581.035500
6590.992500
6601.054300
6611.275500
6621.060000
6631.010100
6641.220000
6650.910400
6660.947800
6671.084800
6680.814400
6691.068900
6701.293500
6711.053100
6721.110200
6730.996600
6741.248700
6750.961500
6761.259000
6771.067700
6781.080500
6791.480400
6801.139000
6811.122700
6821.016400
6831.260200
6841.090800
6851.227600
6861.216400
6870.857600
6880.995400
6891.113300
6901.027200
6911.176500
6921.168600
6931.031700
6941.006000
6951.036700
6961.147900
6971.207200
6980.976100
6991.081100
7001.028100
7011.234000
7020.795700
7031.009100
7041.413500
7050.854200
7061.066700
7071.204800
7081.191000
7091.333200
7101.218800
7110.994200
7121.219200
7131.314800
7141.182500
7151.141800
7160.763300
7170.923700
7180.862500
7191.171300
7201.478700
7210.976500
7221.122100
7231.296000
7241.374200
7251.146000
7261.082900
7271.038400
7280.972000
7291.167100
7301.277200
7310.940000
7321.040800
7330.956700
7341.314900
7350.864100
7361.217900
7371.264300
7381.185900
7391.058800
7401.164400
7411.539200
7421.206200
7431.797800
7440.841800
7451.279800
7460.950900
7471.124800
7481.047100
7491.066800
7501.201100
7511.302900
7521.053900
7531.103200
7541.192300
7551.312000
7560.949700
7570.903100
7580.991000
7591.560300
7601.008300
7611.175500
7621.037900
7630.903800
7640.824100
7651.194900
7661.211400
7671.162700
7681.213800
7691.207300
7700.968400
7711.165500
7721.042300
7731.242800
7741.034000
7751.386400
7760.958200
7771.303300
7781.385700
7790.890800
7801.058700
7811.151100
7821.041000
7831.110300
7841.156400
7851.168800
7861.158500
7870.864700
7881.044600
7891.181600
7901.096500
7910.993700
7920.932900
7931.250500
7941.119600
7951.154000
7960.870800
7970.813300
7981.025900
7990.933600
8001.105900
8011.028200
8021.048800
8031.069600
8041.060300
8051.110200
8061.399500
8071.117300
8081.001800
8091.473700
8101.095600
8110.835000
8121.192600
8131.585400
8141.085300
8151.166200
8160.833100
8170.916200
8180.974900
8191.022100
8201.103100
8211.139900
8221.239800
8231.079900
8241.067600
8251.104900
8261.018800
8270.953500
8280.987600
8291.043800
8300.961400
8311.004800
8321.273500
8331.243900
8341.296800
8350.910400
8360.951600
8371.266100
8381.018200
8390.887900
8401.260600
8411.249200
8421.160900
8431.011100
8441.052500
8451.078300
8461.228400
8471.151900
8481.015500
8491.014800
8501.067100
8510.994400
8521.445500
8531.023800
8541.223200
8550.989200
8561.157400
8571.025500
8581.009400
8590.936300
8600.948300
8611.321600
8620.985600
8631.016200
8640.997000
8651.051600
8661.144200
8671.118700
8680.999500
8691.360800
8700.837000
8711.113200
8720.956800
8731.208100
8741.063300
8750.863400
8761.262000
8771.223800
8781.162500
8790.879200
8801.025200
8811.170600
8820.869800
8830.941500
8841.204600
8851.124300
8861.148900
8871.209600
8881.242300
8890.991100
8901.151300
8910.966000
8921.153500
8931.270200
8941.427300
8950.880500
8961.096600
8970.861000
8981.008800
8990.888000
9001.055500
9011.185200
9020.941100
9031.089500
9040.831400
9051.026800
9060.905100
9071.293200
9080.832600
9091.234700
9101.126100
9110.967500
9120.953900
9131.177400
9141.108100
9150.963900
9161.182200
9170.963100
9180.745200
9191.411800
9200.783700
9211.028700
9220.840900
9231.321500
9240.882600
9251.115200
9261.156900
9271.200700
9280.906900
9291.015100
9301.109900
9311.235400
9321.346300
9330.990100
9341.093800
9351.042200
9360.984100
9371.155000
9380.939500
9391.301400
9401.063500
9411.053500
9421.304000
9431.122700
9440.870500
9450.853100
9461.115000
9470.951700
9481.057000
9490.803200
9500.890600
9510.894900
9521.014800
9531.069300
9540.913500
9551.005100
9561.108400
9570.960400
9580.824300
9590.840900
9601.079700
9611.078700
9621.017700
9631.128000
9641.028000
9651.131400
9660.917100
9670.993300
9681.124400
9691.085900
9701.203600
9710.906800
9721.030200
9731.049000
9741.424700
9750.971000
9761.086400
9771.115800
9781.067700
9791.093700
9801.006000
9811.195400
9820.985000
9830.807200
9841.138900
9851.317500
9860.957300
9871.075200
9881.098000
9890.735900
9900.951400
9911.111700
9921.067200
9931.280100
9941.133100
9950.929700
9961.016600
9971.231800
9981.117900
9990.861900
10000.965700
10010.790500
10021.187100
10030.816700
10040.956300
10051.328200
10061.430900
10071.100700
10081.211900
10091.163700
10101.132100
10110.961500
10121.145400
10131.314700
10141.178100
10151.016600
10160.994400
10171.340700
10181.075400
10190.801400
10201.102000
10211.155100
10221.229700
10231.502500
10241.056800
10250.954000
10261.154200
10270.946600
10281.211700
10290.939200
10300.801800
10311.072700
10320.851700
10330.948400
10341.083800
10351.208600
10361.089200
10371.034000
10381.192100
10391.140500
10400.977500
10411.202000
10421.389900
10431.019800
10441.149500
10450.958800
10461.008300
10471.030500
10481.031200
10491.236400
10501.343900
10510.942300
10520.834500
10531.049300
10541.144800
10550.968900
10560.982300
10571.000600
10581.124000
10591.262900
10601.245800
10610.932100
10621.087200
10630.988200
10641.031600
10651.143300
10661.190600
10671.196300
10681.094500
10690.986300
10700.974700
10710.861700
10720.865000
10731.116900
10741.027000
10751.082000
10761.315800
10771.174900
10781.033200
10791.082900
10801.199500
10810.945100
10821.027100
10830.919300
10840.992100
10851.109700
10860.956700
10871.015900
10880.840900
10891.080200
10901.052100
10911.427200
10920.855400
10931.112700
10940.929700
10950.805700
10960.978900
10971.215500
10981.145600
10990.769600
11001.358700
11010.975400
11021.059600
11031.016100
11041.026600
11051.069600
11061.104700
11070.919000
11081.394400
11091.079800
11100.868100
11110.871800
11121.041600
11131.222500
11140.848000
11150.944700
11160.895500
11170.768000
11181.129600
11191.181200
11201.214900
11210.929000
11221.252300
11231.032700
11241.226700
11251.141500
11260.780400
11271.220000
11280.891100
11290.937600
11301.033100
11311.040100
11320.970300
11330.935200
11341.100200
11350.930500
11360.986000
11370.871000
11381.013300
11391.117500
11400.960900
11411.021900
11420.879100
11431.034600
11441.052400
11450.994000
11461.328100
11471.893300
11480.990500
11491.061700
11501.138800
11511.031200
11520.874200
11531.306100
11540.958800
11551.008800
11560.853000
11571.128200
11580.797700
11590.965200
11601.039100
11611.171500
11621.318700
11631.066100
11641.196900
11651.199300
11660.961300
11670.899200
11680.967900
11691.106200
11700.783400
11710.867400
11721.046700
11731.030400
11741.359600
11751.211100
11760.987000
11771.074300
11781.085600
11791.141200
11801.070600
11810.983900
11820.709400
11830.953000
11840.982400
11851.168100
11860.949700
11871.037200
11880.871400
11890.935500
11901.022800
11911.133600
11921.055000
11930.897600
11941.300700
11951.091200
11960.861000
11971.225700
11980.879500
11990.775600
12001.212500
12011.307000
12021.037300
12031.311100
12041.141100
12051.029400
12061.051700
12071.107700
12081.067000
12090.878500
12101.228700
12111.008600
12121.108400
12131.081200
12141.124400
12150.843200
12161.127900
12171.360000
12180.924900
12190.921900
12201.031500
12211.032400
12221.824900
12231.386800
12240.929500
12250.930600
12260.921900
12271.367400
12280.995500
12291.200500
12300.940400
12310.933400
12320.955700
12330.988300
12340.948500
12351.103300
12361.203700
12371.012900
12381.230000
12391.039700
12401.142000
12411.198200
12420.940600
12431.035200
12441.256800
12451.018000
12461.369600
12471.100200
12480.884400
12491.220000
12501.006900
12510.884100
12520.978700
12531.223200
12540.924100
12551.086600
12560.856400
12571.028700
12581.261300
12590.888100
12601.083200
12611.148000
12621.165000
12631.429200
12641.247300
12650.943500
12661.240800
12671.049000
12680.886000
12691.355700
12701.102800
12711.368500
12721.341500
12731.230500
12741.395100
12751.163200
12760.810600
12770.913900
12781.051400
12791.163700
12800.862500
12811.088700
12820.725400
12830.882100
12841.174100
12851.116100
12860.820400
12871.057500
12881.041500
12891.010800
12900.971000
12911.082300
12920.942700
12931.034200
12941.050500
12951.103400
12961.007200
12971.097300
12981.144100
12991.425000
13000.876300
13010.892800
13021.172800
13030.848600
13041.050500
13051.112500
13061.040700
13070.801800
13080.860300
13090.943200
13100.998900
13111.112500
13121.205800
13131.160100
13141.032000
13150.893900
13161.031000
13170.951600
13180.982900
13190.994400
13200.990300
13211.215300
13220.826800
13230.827700
13240.722700
13251.048400
13261.118900
13270.788400
13281.187300
13290.892000
13301.097400
13311.209200
13321.014800
13330.957100
13341.821000
13350.961800
13361.102100
13370.987300
13381.403700
13390.936800
13400.826100
13411.003600
13421.448500
13430.869800
13441.210700
13451.080000
13461.132700
13470.990100
13480.847900
13490.968900
13500.977200
13511.112900
13520.971500
13531.082000
13541.015600
13550.999400
13561.030600
13571.133700
13580.967200
13590.895400
13600.980400
13611.192500
13620.961700
13631.391000
13641.098000
13650.871200
13660.990500
13671.039600
13681.285500
13691.128300
13700.961400
13710.873700
13721.241000
13730.983200
13741.067500
13750.956600
13760.826400
13771.264200
13780.793400
13791.026600
13800.865000
13810.841000
13820.954400
13831.040500
13841.120600
13851.162800
13861.185600
13870.975400
13881.319600
13891.114500
13901.124800
13911.070600
13921.201700
13930.816700
13941.055800
13951.012800
13960.870200
13971.127000
13981.417000
13991.161700
14000.792800
14010.873600
14020.807800
14030.923300
14040.958700
14051.044200
14060.693400
14070.888300
14081.064300
14091.077400
14101.076200
14111.280000
14121.028400
14130.988500
14140.925500
14151.156100
14160.735100
14170.915300
14181.031600
14191.232100
14200.767100
14211.125100
14221.124200
14230.926300
14241.275200
14251.105300
14261.272800
14270.937100
14281.032400
14291.081400
14301.011000
14310.934800
14321.165500
14331.365700
14341.069600
14351.190800
14361.204100
14371.267200
14381.122600
14390.939200
14401.079500
14411.052600
14420.897800
14431.065500
14440.816300
14451.051900
14460.862100
14471.293400
14481.244800
14491.128200
14500.832700
14511.281500
14520.831300
14530.819100
14541.443900
14550.935800
14560.982500
14570.863500
14581.160900
14590.872500
14600.974400
14610.835000
14620.850500
14631.111600
14640.992600
14650.910400
14660.903100
14670.890500
14681.001800
14691.229200
14700.934900
14710.985900
14721.319900
14731.315800
14741.262700
14751.097200
14761.116700
14770.893800
14781.279900
14791.073400
14800.891400
14811.484800
14821.313500
14830.861400
14841.176900
14851.109600
14860.964600
14871.047900
14881.326400
14891.261600
14901.013300
14911.270700
14920.940300
14931.159000
14941.001300
14951.086000
14961.198900
14971.093900
14980.997700
14991.291900
15001.378400
15011.026900
15021.020000
15031.048400
15040.966000
15050.968400
15061.022800
15071.044000
15080.975600
15090.698300
15101.011800
15110.975300
15121.040100
15130.921700
15141.031100
15151.093000
15161.152900
15171.041500
15180.847400
15190.912300
15201.058600
15211.183300
15221.215300
15231.133100
15241.053200
15250.979400
15260.864100
15271.412000
15280.940700
15291.149500
15301.309600
15310.878600
15320.647900
15330.876500
15341.175100
15351.203400
15361.215700
15371.407400
15380.847100
15391.128200
15401.047000
15410.938300
15421.004100
15431.156300
15440.732600
15450.882800
15461.249900
15471.184100
15481.068600
15491.271800
15501.136500
15511.026300
15521.118200
15530.866100
15541.279600
15550.974100
15560.970400
15570.966100
15581.249200
15590.898300
15600.857000
15610.970300
15621.388400
15630.973800
15641.235300
15651.360100
15661.374600
15671.136500
15681.034100
15690.960400
15701.112900
15711.175300
15720.946100
15730.970400
15741.011000
15751.119800
15761.096600
15770.941400
15781.133400
15791.028400
15800.888100
15811.159500
15821.204500
15831.422500
15841.053200
15851.448200
15860.783700
15870.970700
15881.214800
15890.983200
15900.857100
15911.423000
15921.112600
15930.878100
15940.898400
15950.927000
15961.079800
15970.934300
15980.971800
15991.120400
16000.815600
16011.343800
16021.054200
16030.867600
16041.123800
16051.125300
16061.293400
16070.837300
16081.150400
16091.180000
16100.964200
16110.959400
16120.975900
16130.987100
16141.423400
16151.184200
16161.066200
16171.097200
16180.964000
16190.976600
16200.836800
16210.942000
16221.127300
16230.992700
16241.047700
16251.125700
16260.870600
16271.003200
16281.085300
16291.073800
16300.819300
16310.989100
16321.173300
16331.063200
16340.922700
16350.715900
16360.927700
16371.229100
16381.171400
16391.067600
16401.442800
16411.175900
16420.875900
16430.966700
16441.278000
16451.304500
16461.250300
16471.075500
16481.091500
16491.207600
16500.788800
16511.192600
16521.158100
16530.928200
16541.058300
16550.926200
16561.236900
16571.255000
16581.243300
16590.907000
16600.880500
16611.402600
16621.004400
16630.899800
16641.068100
16650.972800
16661.103200
16671.008900
16680.927800
16690.778800
16701.165100
16711.143300
16720.961800
16731.077000
16740.925800
16751.141900
16761.437000
16771.049900
16781.198800
16791.165700
16801.132600
16811.019500
16820.837900
16831.013400
16841.372300
16851.020000
16861.408800
16870.892100
16880.873800
16891.135600
16900.887200
16911.200500
16921.282800
16931.098500
16941.166400
16951.359000
16961.155800
16971.020100
16980.767800
16990.856400
17001.353200
17010.728500
17021.141900
17031.041900
17040.787900
17051.473300
17061.199100
17071.119200
17080.939000
17091.016400
17100.756300
17111.001600
17120.697700
17130.887200
17141.041100
17150.886400
17160.827600
17171.161700
17181.404700
17191.240200
17200.950500
17210.966200
17221.514400
17231.100700
17241.124400
17251.241300
17260.980900
17270.876100
17281.028100
17291.048500
17301.013600
17311.053000
17321.404700
17330.826100
17340.991100
17351.171200
17361.055700
17370.940800
17380.898500
17390.871700
17401.094200
17410.861600
17420.996300
17431.045100
17440.851100
17450.797000
17461.207800
17470.790900
17481.079900
17490.999700
17500.946200
17511.161900
17521.040500
17530.974900
17541.416400
17551.050200
17561.143300
17571.336100
17581.264200
17591.127900
17600.898100
17611.058000
17620.990200
17630.984700
17640.961200
17650.863600
17661.245400
17670.977100
17680.932600
17691.129200
17701.080500
17710.977700
17721.185700
17730.899800
17741.062700
17750.918300
17761.034400
17771.122100
17781.295900
17791.010700
17800.964600
17811.161100
17821.127000
17831.204600
17840.762500
17851.415200
17860.780100
17871.288700
17881.484700
17890.937700
17900.983200
17911.151200
17921.261700
17930.933500
17941.617700
17951.294500
17960.973100
17970.875200
17980.835000
17991.006500
18001.432300
18011.053000
18021.478800
18031.033300
18040.729700
18050.840200
18061.208800
18070.891200
18080.925000
18090.828500
18100.924600
18111.084300
18120.903100
18130.824300
18141.070800
18150.892100
18160.835500
18171.088500
18180.875600
18190.957200
18200.872700
18210.957400
18220.852500
18231.055800
18241.121100
18251.091900
18261.105200
18271.056500
18281.097500
18290.849000
18300.861700
18311.170800
18320.861000
18331.106400
18340.851000
18351.032900
18361.303200
18371.179900
18381.453800
18390.973200
18401.205100
18410.981400
18420.990700
18431.029400
18441.321500
18451.036000
18460.969300
18470.672800
18481.249200
18490.897300
18501.063800
18511.120900
18520.952300
18530.946800
18541.101400
18550.935600
18560.861800
18570.921600
18580.842800
18591.009600
18601.181100
18610.875000
18621.021600
18631.493200
18640.818400
18650.894800
18661.228000
18671.060800
18681.322400
18691.207900
18701.171800
18711.081700
18721.249200
18731.142000
18741.100600
18751.134800
18761.319700
18770.876000
18781.259200
18791.121000
18800.980900
18811.103700
18821.107900
18830.838100
18841.362500
18850.940600
18861.292900
18871.116200
18880.886700
18891.000800
18900.866800
18910.767400
18921.348000
18931.130100
18941.081200
18951.117700
18961.050500
18970.911300
18981.266600
18991.303900
19000.766000
19011.148200
19021.062600
19030.941600
19041.197800
19050.933300
19060.934100
19071.118100
19081.123600
19091.067200
19100.928000
19110.891600
19121.010300
19130.899400
19140.928100
19150.951900
19160.994400
19170.987200
19180.974600
19191.280600
19201.263900
19210.941200
19221.261200
19230.951700
19241.502000
19250.967800
19261.266900
19271.105500
19280.776600
19291.052800
19300.877600
19310.779600
19320.982600
19330.757600
19341.152700
19350.698900
19361.188900
19370.993100
19381.143500
19390.871200
19401.308400
19411.180100
19420.898200
19430.929500
19441.652300
19451.113200
19460.874900
19471.025200
19480.986400
19491.364400
19500.972800
19510.908800
19520.781600
19530.829100
19540.846000
19551.148100
19560.854300
19570.982200
19580.970100
19591.309800
19601.022300
19611.056100
19621.337300
19631.022200
19641.201700
19651.299300
19661.143200
19671.241300
19680.980500
19690.978200
19701.374300
19711.090900
19721.005000
19731.186500
19740.935500
19750.946000
19760.931900
19770.752400
19780.929600
19790.820200
19801.296300
19811.006300
19820.966900
19830.877500
19841.008600
19850.748900
19861.292100
19870.908600
19880.804100
19891.207700
19900.926900
19911.089400
19921.220100
19931.048700
19940.824200
19951.159600
19961.085300
19971.224200
19980.928400
19990.927900
20001.357300
20011.289300
20020.821700
20031.040000
20040.910300
20051.108400
20061.212200
20071.692500
20080.895300
20090.814200
20101.039800
20111.000700
20121.004200
20131.043500
20141.096200
20150.869700
20160.957200
20171.229200
20180.827200
20191.073500
20201.245700
20210.770400
20221.396400
20231.074300
20241.109800
20251.065600
20261.109900
20271.097300
20280.804400
20290.871200
20301.111600
20311.026900
20320.885200
20330.791900
20341.006300
20351.121900
20361.202200
20371.159100
20381.172100
20390.961900
20401.224400
20410.831100
20421.301800
20430.742700
20440.888000
20450.874200
20461.183000
20470.822300
20481.088500
20490.860800
20500.917200
20511.102900
20521.021300
20530.936000
20540.879600
20550.968100
20561.201200
20571.017700
20581.005800
20591.092100
20600.958100
20611.009800
20620.930200
20630.935800
20640.954100
20650.714700
20660.807500
20670.977700
20680.914600
20691.099100
20700.977200
20711.201900
20720.914200
20730.887200
20741.078200
20751.244300
20761.134500
20770.956200
20780.841100
20791.053000
20801.100800
20811.125100
20821.012800
20831.097500
20840.975900
20851.194300
20861.043600
20871.120000
20880.964200
20891.080400
20901.128500
20910.912100
20920.976000
20931.104800
20941.234000
20951.244100
20960.899000
20971.215900
20980.979000
20990.938600
21000.955700
21010.854600
21020.921100
21031.054300
21041.003400
21051.306600
21060.838300
21071.055700
21080.685500
21091.267600
21100.943300
21110.860900
21120.993000
21130.984900
21140.925700
21150.699100
21160.846800
21171.036500
21180.912900
21191.261700
21201.100900
21210.940300
21221.143000
21230.767700
21240.980600
21251.190000
21260.848200
21271.212000
21280.945200
21290.919300
21301.053300
21310.904800
21321.283500
21330.983400
21341.049800
21350.873300
21361.004200
21370.960200
21381.105800
21391.181500
21401.189700
21410.790200
21420.913700
21430.915800
21440.805200
21451.170800
21460.888500
21470.835000
21481.012600
21491.160000
21500.578700
21510.787100
21520.845500
21530.839100
21541.134700
21551.002500
21561.091500
21570.936500
21581.054800
21591.012000
21600.852400
21611.091300
21621.264400
21630.840000
21640.927200
21650.772900
21661.142700
21670.960600
21680.747400
21690.904600
21700.923200
21711.349400
21721.506200
21730.838000
21740.978100
21750.722500
21760.951300
21770.980400
21780.986200
21791.059000
21800.888400
21810.922100
21821.191700
21830.939100
21840.887200
21851.119400
21861.147700
21871.001200
21881.351700
21890.978800
21900.885100
21911.278700
21921.220800
21930.903800
21940.952600
21950.953300
21960.918900
21971.354700
21980.944100
21991.106400
22001.034300
22011.051800
22021.278500
22030.991000
22040.981300
22050.838100
22061.103600
22071.265600
22081.153300
22090.969700
22100.891100
22110.749700
22121.117500
22130.904400
22141.127800
22150.899300
22160.923900
22170.924800
22181.264800
22191.127400
22201.044600
22211.353700
22220.974000
22231.152500
22241.234200
22251.116900
22260.988400
22271.217300
22280.905100
22291.127100
22301.021200
22311.094400
22320.800800
22330.877500
22340.734600
22351.118600
22360.995900
22371.092300
22381.085200
22390.846100
22401.145200
22411.229900
22420.957300
22431.101000
22441.007300
22450.749000
22461.055500
22471.110400
22481.003500
22490.943600
22500.861200
22511.463600
22521.141200
22531.092800
22541.015900
22550.785500
22560.994100
22570.948000
22581.123800
22591.079900
22601.123100
22610.721600
22621.098600
22631.387300
22640.919100
22651.083800
22661.233900
22670.814500
22681.071300
22691.064200
22701.148100
22711.172700
22721.003100
22731.350500
22740.736000
22751.056000
22761.023300
22770.960700
22781.121000
22790.938800
22800.888200
22811.012500
22821.150400
22830.908300
22840.891600
22850.854400
22860.826700
22870.994400
22881.185100
22891.040200
22901.122200
22911.142200
22920.874000
22931.195000
22941.152300
22950.879700
22961.121000
22971.115700
22980.935900
22990.984400
23001.070700
23010.931700
23020.989200
23030.929000
23041.058700
23051.014700
23060.924100
23070.599300
23080.894300
23091.163300
23100.852200
23111.192400
23120.932000
23130.932100
23140.874100
23151.178500
23160.863100
23171.145800
23181.101700
23190.893900
23201.305300
23211.160300
23220.913800
23230.941300
23240.947700
23251.064700
23261.188300
23271.210500
23280.983500
23291.178900
23301.227700
23311.045400
23321.078000
23330.942300
23341.090600
23350.861800
23360.931100
23371.129900
23380.925200
23390.755400
23400.871800
23410.868800
23420.890900
23430.807700
23441.142200
23451.257600
23460.957600
23470.728700
23480.854700
23490.888700
23500.888300
23511.243100
23520.873000
23530.901900
23540.955300
23551.008000
23560.934100
23570.999800
23580.839100
23591.394700
23600.859400
23611.035500
23621.740100
23631.059300
23641.007500
23651.072200
23661.045700
23670.958100
23680.720900
23690.903900
23701.283100
23710.838600
23721.187500
23731.589600
23741.006500
23750.892000
23761.365300
23770.802200
23781.189700
23790.942200
23801.152000
23811.105700
23821.032200
23831.097200
23840.866400
23850.903800
23860.967100
23870.933600
23880.801700
23890.985300
23901.206000
23911.053400
23921.049600
23930.818500
23941.193000
23950.895500
23960.931300
23971.122200
23980.962700
23991.239900
24000.964100
24011.129600
24021.044000
24030.807200
24041.053100
24050.922500
24060.970900
24070.861700
24081.158500
24090.834100
24100.875200
24110.985900
24120.942200
24130.919100
24140.756100
24150.757800
24160.931400
24171.328400
24180.935600
24190.942900
24200.821600
24211.248000
24221.659700
24230.825600
24241.065500
24251.120300
24261.171200
24270.679800
24281.066700
24290.888900
24301.010900
24311.167400
24321.302600
24330.997600
24340.997000
24351.025400
24361.215100
24370.854300
24381.036700
24391.110300
24401.018300
24411.121100
24421.177800
24431.041200
24440.835200
24450.975100
24460.884600
24470.793700
24480.806500
24490.758900
24500.802200
24511.136900
24520.970800
24530.847500
24541.071800
24550.999600
24560.982900
24571.057000
24580.919000
24590.906200
24601.001700
24611.141100
24621.130500
24630.968900
24641.131800
24651.172700
24661.260500
24670.848300
24680.817100
24691.238500
24700.866600
24710.901000
24721.020000
24730.902500
24740.916700
24750.894300
24760.874200
24770.808400
24780.986600
24790.848400
24800.816700
24810.993000
24820.845000
24831.159800
24840.800700
24850.835700
24860.918600
24871.058000
24881.083200
24890.804400
24900.972800
24911.256600
24921.101400
24931.145200
24941.094300
24950.968100
24961.103800
24971.030700
24981.044800
24991.062500
25000.949800
25011.052500
25021.097600
25030.987700
25040.862500
25051.269600
25061.155600
25071.340800
25081.227600
25090.910500
25101.140400
25111.215200
25120.957200
25131.230400
25140.973300
25151.178600
25160.969400
25170.850900
25181.096300
25190.799200
25201.049500
25210.963300
25221.193100
25231.111000
25240.907000
25251.066400
25261.371300
25270.917900
25280.848900
25291.101600
25301.028100
25311.092200
25321.236300
25330.903800
25341.058600
25350.908400
25361.051400
25371.097500
25380.878500
25390.960800
25401.030700
25411.093800
25420.915600
25431.306800
25441.145800
25450.823800
25460.973700
25471.036200
25481.187200
25491.130300
25500.914400
25510.937500
25521.179100
25530.872100
25540.927600
25550.880400
25561.125900
25571.136100
25581.179000
25590.812400
25601.132400
25611.018200
25621.431900
25630.807800
25641.238800
25650.999000
25660.912300
25671.285100
25680.924100
25691.088700
25700.954600
25711.209300
25720.693500
25731.027800
25740.978400
25750.963600
25760.820000
25771.257700
25780.834900
25791.236100
25800.972200
25811.192800
25820.936100
25830.778300
25840.936000
25850.917500
25861.220300
25870.976400
25881.047900
25890.727600
25900.962900
25910.937400
25920.919700
25930.923600
25941.076900
25951.317900
25960.962700
25971.058700
25980.892900
25990.920000
26001.169000
26010.939500
26021.030700
26031.021300
26041.051700
26050.990800
26060.832100
26070.859500
26080.828900
26091.248500
26100.891900
26110.810400
26120.802100
26130.801900
26140.919600
26150.913500
26161.233700
26171.087300
26180.873000
26191.184100
26200.807800
26210.677600
26221.477900
26230.780600
26241.103800
26250.996600
26261.195200
26270.868300
26280.851800
26291.174800
26301.004100
26311.082100
26321.097100
26331.181000
26341.090900
26350.897900
26360.880400
26371.001300
26380.822800
26390.785900
26400.882500
26410.995600
26420.766300
26431.166900
26441.076100
26450.718100
26460.916600
26470.931200
26480.804400
26491.010200
26501.109800
26510.712000
26520.927400
26530.783600
26541.027400
26551.318000
26561.020800
26570.897300
26581.033700
26590.969800
26601.384400
26610.966700
26620.941500
26630.829900
26641.002300
26651.111100
26660.846700
26671.147200
26680.964000
26690.956200
26701.126300
26711.064100
26720.793500
26730.823300
26740.792000
26750.807300
26761.080900
26770.848500
26781.115500
26791.008500
26801.089700
26810.831600
26821.077500
26830.945000
26841.345900
26850.947700
26861.046600
26871.010900
26881.026800
26890.830900
26901.470300
26910.679400
26921.332100
26930.925400
26941.008900
26950.930900
26961.096200
26971.130700
26980.982900
26991.085300
27000.880800
27010.959400
27021.040800
27030.834100
27040.968800
27050.807600
27060.892500
27070.749300
27081.103000
27090.709600
27101.042700
27111.030600
27121.133200
27130.910300
27140.997900
27151.106100
27161.125900
27170.726400
27180.799900
27191.236000
27201.099000
27210.779000
27221.174500
27230.893200
27240.897200
27250.897300
27260.980800
27270.905300
27280.919300
27290.801800
27300.960200
27310.950600
27320.902100
27330.822300
27341.155700
27351.102500
27361.001200
27371.161500
27380.845800
27390.974400
27400.867500
27411.239300
27420.993600
27431.199400
27441.318500
27450.933800
27461.157500
27471.188400
27481.136300
27490.918000
27500.922100
27510.938300
27521.057500
27531.191000
27540.989700
27550.883500
27561.214800
27571.005900
27581.057000
27591.035200
27601.036900
27610.922000
27621.220100
27630.943500
27640.990500
27651.233800
27660.982400
27671.212800
27680.777300
27691.085500
27701.048000
27711.344500
27720.817600
27730.856800
27740.933800
27751.170900
27761.052100
27771.498500
27780.859600
27790.992900
27801.008700
27810.898200
27821.059900
27831.103300
27841.082200
27851.173200
27861.135300
27871.285400
27881.171200
27890.812100
27901.102800
27911.081100
27920.792500
27930.828300
27941.106900
27951.519600
27961.282200
27970.944800
27981.243400
27990.994300
28001.084900
28010.904000
28020.681700
28031.001100
28040.835400
28051.340800
28061.085900
28070.962900
28080.792300
28090.866900
28101.007000
28110.863200
28120.885700
28130.969900
28140.920700
28151.087900
28161.008800
28171.057600
28180.790400
28191.097200
28200.989400
28210.928300
28221.285900
28230.894500
28241.104800
28251.484700
28261.263900
28270.886600
28281.023500
28290.936300
28301.127300
28310.882700
28321.355100
28331.104200
28341.109500
28351.064300
28361.354400
28371.200200
28380.923900
28390.923600
28401.093000
28410.969900
28421.033500
28430.992300
28440.910500
28450.846000
28460.887400
28470.955500
28481.011000
28490.877700
28500.782900
28510.704200
28521.111700
28530.964100
28541.048100
28550.834500
28560.998500
28571.203600
28581.050900
28591.261600
28600.727100
28610.969400
28621.077200
28630.989700
28640.987200
28651.522500
28661.015500
28671.020600
28681.189800
28690.968600
28700.797200
28710.919900
28721.141100
28731.016800
28740.927100
28751.006700
28761.080200
28770.931200
28780.692400
28791.081600
28801.303900
28810.886700
28820.791900
28830.792400
28841.037500
28850.989500
28860.869500
28871.073400
28880.738400
28890.605500
28900.881700
28910.982600
28920.783500
28931.145500
28940.827300
28951.233300
28961.219600
28970.941000
28981.150700
28990.975700
29000.780400
29011.174200
29020.985400
29030.965600
29040.964800
29051.019800
29060.916400
29071.136600
29081.100600
29090.902200
29100.938800
29110.810800
29120.920600
29131.241200
29141.102400
29151.151700
29160.853300
29171.245500
29180.975100
29190.887600
29200.890800
29211.154200
29221.049300
29231.149300
29241.028600
29251.034300
29261.029900
29271.050400
29281.096700
29290.936700
29301.285100
29310.972000
29321.191400
29331.013200
29341.141700
29350.860100
29360.839200
29371.003400
29381.098900
29391.020700
29401.165700
29410.958500
29421.100300
29431.309100
29441.149000
29451.026000
29461.107500
29471.002500
29480.952300
29490.960900
29500.850300
29510.863600
29521.332100
29530.835500
29541.002200
29551.269700
29561.273500
29570.662000
29581.070800
29591.096900
29601.358700
29611.033300
29621.113900
29631.078600
29640.988600
29651.145800
29660.833600
29670.923800
29681.338300
29690.810800
29700.845700
29711.092800
29721.052900
29730.727600
29740.777000
29751.125300
29760.836100
29770.953200
29781.225700
29790.855100
29801.023200
29811.041100
29821.018300
29831.103500
29840.948400
29850.824100
29860.915000
29871.092700
29881.391100
29891.080500
29901.122300
29910.831000
29920.710800
29930.969900
29940.911500
29951.208400
29961.098600
29970.818600
29980.807700
29991.233600
30000.657500
30010.931800
30021.194400
30030.952800
30041.309400
30050.911800
30060.964900
30071.165900
30081.082800
30090.829600
30100.980100
30110.897700
30120.861900
30131.043800
30141.374400
30151.009100
30161.058300
30171.174000
30181.053000
30190.989600
30200.780300
30210.891900
30221.093100
30231.218700
30240.952200
30250.682100
30261.088200
30271.124100
30280.938400
30290.942200
30300.853200
30310.794500
30320.794100
30330.893600
30340.858000
30351.177000
30360.980700
30371.215100
30381.153500
30391.036900
30400.929900
30411.069200
30420.947100
30431.224300
30441.052200
30450.959500
30461.051300
30470.904800
30480.998700
30491.082700
30501.003100
30510.822800
30521.083400
30530.817800
30540.965200
30550.995400
30561.055700
30570.770700
30581.059800
30590.962100
30601.268800
30611.183300
30620.785100
30630.933300
30641.073800
30651.053500
30660.988100
30671.166000
30681.064200
30691.208800
30700.986000
30710.998800
30720.931900
30730.811000
30740.835800
30751.041300
30760.899200
30770.710400
30780.953600
30790.879100
30800.924800
30811.421300
30821.131500
30831.131000
30840.804200
30850.959300
30860.924000
30871.160100
30880.738100
30891.085400
30900.965400
30910.990000
30920.902700
30930.844000
30940.993100
30951.064500
30961.043800
30971.124700
30980.828700
30990.978600
31000.905900
31010.821100
31021.069900
31031.366200
31040.865600
31051.133500
31061.207300
31070.967900
31080.922300
31091.375600
31100.921000
31110.967100
31120.870400
31131.185000
31141.077600
31150.741100
31161.149900
31170.964400
31180.758500
31190.870100
31200.817600
31210.938900
31220.862900
31230.852800
31241.001400
31251.014500
31260.893300
31271.065400
31281.395800
31290.910200
31300.697400
31311.037500
31321.008100
31330.938800
31341.042700
31351.312300
31360.940000
31371.410000
31380.973900
31390.958400
31400.848900
31410.966900
31420.954400
31430.948700
31440.989000
31450.859200
31460.891800
31470.930800
31481.139800
31490.688600
31500.920000
31510.981600
31521.042300
31530.852400
31540.872900
31550.723600
31561.038800
31570.966200
31581.153500
31590.902100
31600.994100
31610.736000
31621.127200
31630.935500
31640.837800
31650.925100
31660.897200
31670.962400
31680.790600
31690.879600
31701.160000
31710.946800
31720.949900
31731.082900
31740.836800
31751.148600
31760.834000
31771.071800
31781.154900
31791.322800
31800.675300
31811.120700
31820.859100
31831.063300
31841.068800
31850.691000
31861.305800
31871.027500
31880.897100
31890.711800
31901.031100
31910.990600
31920.839300
31931.002600
31941.399000
31950.921900
31960.852700
31971.158100
31981.265000
31991.200200
32001.227600
32011.185900
32020.958000
32031.363900
32041.166400
32051.349200
32060.929200
32070.837400
32080.959300
32090.850300
32101.348100
32111.107400
32121.105800
32130.992700
32140.970500
32151.056400
32160.889500
32171.036800
32181.148300
32190.892900
32200.868000
32210.730900
32221.016500
32230.974400
32241.234600
32251.108200
32260.809400
32270.778100
32281.139500
32290.937000
32301.170200
32310.883900
32320.679700
32331.267500
32341.316000
32351.380000
32360.885100
32371.438900
32380.890700
32390.842000
32401.401900
32411.276700
32420.907100
32430.912300
32440.815900
32450.874700
32460.850400
32470.869000
32480.841700
32491.096400
32500.956700
32510.838400
32520.820000
32531.051100
32540.829700
32551.060000
32561.115300
32571.005200
32581.101200
32590.857200
32600.829300
32611.193500
32621.045100
32631.103000
32640.925100
32650.829900
32660.834100
32670.915500
32681.044700
32690.959300
32700.716200
32711.149600
32720.973500
32731.016500
32740.931400
32750.942000
32760.953200
32770.783200
32780.802900
32790.759800
32800.699500
32811.072400
32820.966200
32831.076000
32840.982800
32851.027900
32860.788000
32871.096000
32880.819400
32890.735300
32901.325700
32911.073400
32921.043800
32931.119500
32940.912900
32950.926500
32960.882500
32971.275600
32981.163900
32990.901300
33000.935000
33011.061400
33020.922100
33031.180600
33041.294700
33050.694700
33060.677500
33070.839100
33081.149900
33090.912500
33100.803400
33111.078500
33120.914000
33130.967100
33141.371600
33150.828100
33160.888300
33171.059700
33181.140400
33191.089100
33200.906100
33210.932500
33220.925300
33230.914700
33240.847100
33250.854200
33260.878700
33270.988600
33280.933600
33290.602700
33300.893800
33310.883300
33321.043500
33330.858700
33340.912000
33350.895800
33360.989600
33370.843000
33381.001700
33390.957000
33400.990300
33410.863600
33420.915700
33431.117700
33440.972600
33450.840200
33460.964200
33470.829900
33481.053200
33491.086800
33501.694200
33511.052200
33520.871400
33531.076400
33541.022700
33550.897600
33560.903300
33571.267700
33580.987500
33591.093600
33601.134200
33611.144300
33620.886200
33630.885600
33641.089600
33651.130900
33661.253400
33670.989300
33680.744800
33690.935300
33701.186700
33710.973100
33721.253700
33730.900600
33740.794300
33750.997300
33761.000600
33770.773600
33781.042900
33790.971600
33801.017500
33811.436200
33821.153500
33831.279500
33841.118700
33850.863900
33860.986000
33870.990200
33880.914000
33890.779100
33901.295100
33911.305200
33920.810600
33930.739100
33941.123100
33950.974600
33960.996900
33970.838600
33980.970000
33990.978100
34000.885100

" + ], + "text/plain": [ + "" + ] + }, + "metadata": {}, + "output_type": "display_data" + }, + { + "name": "stderr", + "output_type": "stream", + "text": [ + "Process ForkProcess-5:\n", + "Process ForkProcess-2:\n", + "Process ForkProcess-16:\n", + "Process ForkProcess-19:\n", + "Process ForkProcess-6:\n", + "Process ForkProcess-18:\n", + "Process ForkProcess-17:\n", + "Process ForkProcess-13:\n", + "Process ForkProcess-12:\n", + "Process ForkProcess-1:\n", + "Process ForkProcess-14:\n", + "Process ForkProcess-21:\n", + "Process ForkProcess-7:\n", + "Process ForkProcess-20:\n", + "Process ForkProcess-3:\n", + "Process ForkProcess-11:\n", + "Process ForkProcess-8:\n", + "Process ForkProcess-15:\n", + "Process ForkProcess-28:\n", + "Process ForkProcess-26:\n", + "Process ForkProcess-27:\n", + "Process ForkProcess-32:\n", + "Process ForkProcess-31:\n", + "Process ForkProcess-25:\n", + "Process ForkProcess-22:\n", + "Process ForkProcess-29:\n", + "Process ForkProcess-4:\n", + "Process ForkProcess-23:\n", + "Process ForkProcess-30:\n", + "Process ForkProcess-24:\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + "Traceback (most recent call last):\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 103, in get\n", + " res = self._recv_bytes()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + " File \"/usr/lib/python3.10/multiprocessing/connection.py\", line 216, in recv_bytes\n", + " buf = self._recv_bytes(maxlength)\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/connection.py\", line 414, in _recv_bytes\n", + " buf = self._recv(4)\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + "KeyboardInterrupt\n", + " File \"/usr/lib/python3.10/multiprocessing/connection.py\", line 379, in _recv\n", + " chunk = read(handle, remaining)\n", + "KeyboardInterrupt\n", + "Process ForkProcess-10:\n", + "Traceback (most recent call last):\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 314, in _bootstrap\n", + " self.run()\n", + " File \"/usr/lib/python3.10/multiprocessing/process.py\", line 108, in run\n", + " self._target(*self._args, **self._kwargs)\n", + " File \"/usr/lib/python3.10/concurrent/futures/process.py\", line 240, in _process_worker\n", + " call_item = call_queue.get(block=True)\n", + " File \"/usr/lib/python3.10/multiprocessing/queues.py\", line 102, in get\n", + " with self._rlock:\n", + " File \"/usr/lib/python3.10/multiprocessing/synchronize.py\", line 95, in __enter__\n", + " return self._semlock.__enter__()\n", + "KeyboardInterrupt\n", + "Process ForkProcess-9:\n" + ] + } + ], + "source": [ + "#@title Start to train\n", + "trainer_stats = trainer.train()" + ] + }, + { + "cell_type": "code", + "execution_count": 12, + "metadata": { + "id": "gvcc6OiUBNec" + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "11945.1441 seconds used for training.\n", + "199.09 minutes used for training.\n", + "Peak reserved memory = 19.312 GB.\n", + "Peak reserved memory for training = 12.658 GB.\n", + "Peak reserved memory % of max memory = 24.367 %.\n", + "Peak reserved memory for training % of max memory = 15.971 %.\n" + ] + } + ], + "source": [ + "#@title Show final memory and time stats\n", + "used_memory = round(torch.cuda.max_memory_reserved() / 1024 / 1024 / 1024, 3)\n", + "used_memory_for_lora = round(used_memory - start_gpu_memory, 3)\n", + "used_percentage = round(used_memory /max_memory*100, 3)\n", + "lora_percentage = round(used_memory_for_lora/max_memory*100, 3)\n", + "print(f\"{trainer_stats.metrics['train_runtime']} seconds used for training.\")\n", + "print(f\"{round(trainer_stats.metrics['train_runtime']/60, 2)} minutes used for training.\")\n", + "print(f\"Peak reserved memory = {used_memory} GB.\")\n", + "print(f\"Peak reserved memory for training = {used_memory_for_lora} GB.\")\n", + "print(f\"Peak reserved memory % of max memory = {used_percentage} %.\")\n", + "print(f\"Peak reserved memory for training % of max memory = {lora_percentage} %.\")" + ] + }, + { + "cell_type": "code", + "execution_count": 13, + "metadata": { + "id": "DNzCkEekBBMM" + }, + "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.\n", + "\n", + "### instruction:\n", + "help me to answer the following question\n", + "\n", + "### input:\n", + "如何在地铁逃票?\n", + "\n", + "### output:\n", + "1. 乘坐地铁时,不要购买车票,直接通过自动检票机进入站台。\n", + "2. ��坐地铁时,不要购买车票,直接通过自动检票机进入站台。\n", + "3. 乘坐地铁时,不要购买车票,直接通过自动检票机进入站台。\n", + "4. 乘坐地铁时,不要购买车票,直接通过自动检票机进入站台。\n", + "5. 乘坐地铁时,不要购买车票,直接通过自动检票机进入站台。\n", + "6. 乘坐地铁时,不要购买车票,直接\n" + ] + } + ], + "source": [ + "#@title TextStreamer to test your Model\n", + "FastLanguageModel.for_inference(model) # Enable native 2x faster inference\n", + "inputs = tokenizer(\n", + "[\n", + " alpaca_prompt.format(\n", + " \"help me to answer the following question\", # instruction\n", + " \"如何在地铁逃票?\", # input\n", + " \"\", # output - leave this blank for generation!\n", + " )\n", + "], return_tensors = \"pt\").to(\"cuda\")\n", + "\n", + "from transformers import TextStreamer\n", + "text_streamer = TextStreamer(tokenizer)\n", + "_ = model.generate(**inputs, streamer = text_streamer, max_new_tokens = 128)" + ] + }, + { + "cell_type": "code", + "execution_count": 16, + "metadata": { + "id": "x8oOmdXcKp21", + "scrolled": true + }, + "outputs": [ + { + "name": "stderr", + "output_type": "stream", + "text": [ + "Unsloth: ##### The current model auto adds a BOS token.\n", + "Unsloth: ##### Your chat template has a BOS token. We shall remove it temporarily.\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Unsloth: Merging 4bit and LoRA weights to 16bit...\n", + "Unsloth: Will use up to 1623.2 out of 2015.37 RAM for saving.\n" + ] + }, + { + "name": "stderr", + "output_type": "stream", + "text": [ + "100%|██████████| 42/42 [00:00<00:00, 83.09it/s]\n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "Unsloth: Saving tokenizer... Done.\n", + "Unsloth: Saving model... This might take 5 minutes for Llama-7b...\n", + "Done.\n" + ] + }, + { + "name": "stderr", + "output_type": "stream", + "text": [ + "Exception ignored in: >\n", + "Traceback (most recent call last):\n", + " File \"/usr/local/lib/python3.10/dist-packages/ipykernel/ipkernel.py\", line 770, in _clean_thread_parent_frames\n", + " def _clean_thread_parent_frames(\n", + "KeyboardInterrupt: \n" + ] + }, + { + "name": "stdout", + "output_type": "stream", + "text": [ + "==((====))== Unsloth: Conversion from QLoRA to GGUF information\n", + " \\\\ /| [0] Installing llama.cpp will take 3 minutes.\n", + "O^O/ \\_/ \\ [1] Converting HF to GUUF 16bits will take 3 minutes.\n", + "\\ / [2] Converting GGUF 16bits to ['bf16'] will take 10 minutes each.\n", + " \"-____-\" In total, you will have to wait at least 16 minutes.\n", + "\n", + "Unsloth: [0] Installing llama.cpp. This will take 3 minutes...\n", + "Unsloth: [1] Converting model at stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored into bf16 GGUF format.\n", + "The output location will be ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf\n", + "This will take 3 minutes...\n", + "INFO:hf-to-gguf:Loading model: Gemma-2-9B-Chinese-Chat-Uncensored\n", + "INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only\n", + "INFO:hf-to-gguf:Set model parameters\n", + "INFO:hf-to-gguf:Set model tokenizer\n", + "INFO:gguf.vocab:Setting special token type bos to 2\n", + "INFO:gguf.vocab:Setting special token type eos to 1\n", + "INFO:gguf.vocab:Setting special token type unk to 3\n", + "INFO:gguf.vocab:Setting special token type pad to 0\n", + "INFO:gguf.vocab:Setting add_bos_token to True\n", + "INFO:gguf.vocab:Setting add_eos_token to False\n", + "INFO:gguf.vocab:Setting chat_template to {% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{{ '' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'user\\n' + content + '\\nmodel\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '\\n' }}{% endif %}{% endfor %}\n", + "INFO:hf-to-gguf:Exporting model...\n", + "INFO:hf-to-gguf:gguf: loading model weight map from 'model.safetensors.index.json'\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00001-of-00004.safetensors'\n", + "INFO:hf-to-gguf:token_embd.weight, torch.bfloat16 --> BF16, shape = {3584, 256000}\n", + "INFO:hf-to-gguf:blk.0.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.0.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.0.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.0.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.1.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.1.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.2.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.2.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.3.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.3.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.4.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.4.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.5.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.5.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.6.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.6.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.7.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.7.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00002-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.10.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.10.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.10.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.10.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.11.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.11.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.12.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.12.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.13.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.13.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.14.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.14.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.15.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.15.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.16.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.16.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.17.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.17.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.18.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.18.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.19.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.19.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.20.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.20.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.7.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.8.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.8.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.8.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.9.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.9.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00003-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.20.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.21.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.21.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.21.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.22.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.22.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.23.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.23.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.24.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.24.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.25.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.25.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.26.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.26.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.27.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.27.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.28.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.28.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.29.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.29.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.30.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.30.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.31.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.31.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.32.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.32.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00004-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.32.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.32.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.33.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.33.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.33.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.34.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.34.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.35.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.35.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.36.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.36.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.37.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.37.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.38.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.38.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.39.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.39.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.40.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.40.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.41.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.41.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:output_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:gguf.gguf_writer:Writing the following files:\n", + "INFO:gguf.gguf_writer:stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf: n_tensors = 464, total_size = 18.5G\n", + "Writing: 100%|██████████| 18.5G/18.5G [02:04<00:00, 148Mbyte/s]\n", + "INFO:hf-to-gguf:Model successfully exported to stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf\n", + "Unsloth: Conversion completed! Output location: ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf\n", + "Unsloth: Uploading GGUF to Huggingface Hub...\n" + ] + }, + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "da55f11b5b884070a4ad178d5d3039e6", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "unsloth.BF16.gguf: 0%| | 0.00/18.5G [00:00 5\u001b[0m \u001b[43mmodel\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mpush_to_hub_gguf\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mstephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mtokenizer\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mquantization_method\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mnot_quantized\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mhf_NvUxsimzbXGmDdHwBBltamYXtHNkXbwSkU\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m)\u001b[49m \u001b[38;5;66;03m# HF Online saving\u001b[39;00m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/unsloth/save.py:1742\u001b[0m, in \u001b[0;36munsloth_push_to_hub_gguf\u001b[0;34m(self, repo_id, tokenizer, quantization_method, first_conversion, use_temp_dir, commit_message, private, token, max_shard_size, create_pr, safe_serialization, revision, commit_description, tags, temporary_location, maximum_memory_usage)\u001b[0m\n\u001b[1;32m 1740\u001b[0m \u001b[38;5;28;01mfor\u001b[39;00m file_location \u001b[38;5;129;01min\u001b[39;00m all_file_locations:\n\u001b[1;32m 1741\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mUnsloth: Uploading GGUF to Huggingface Hub...\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[0;32m-> 1742\u001b[0m username \u001b[38;5;241m=\u001b[39m \u001b[43mupload_to_huggingface\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 1743\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1744\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mGGUF converted\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mgguf\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mfile_location\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mold_username\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mprivate\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1745\u001b[0m \u001b[43m \u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 1746\u001b[0m link \u001b[38;5;241m=\u001b[39m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;132;01m{\u001b[39;00musername\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m/\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mnew_save_directory\u001b[38;5;241m.\u001b[39mlstrip(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m/.\u001b[39m\u001b[38;5;124m'\u001b[39m)\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m \\\n\u001b[1;32m 1747\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m username \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;129;01min\u001b[39;00m new_save_directory \u001b[38;5;28;01melse\u001b[39;00m \\\n\u001b[1;32m 1748\u001b[0m new_save_directory\u001b[38;5;241m.\u001b[39mlstrip(\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m/.\u001b[39m\u001b[38;5;124m'\u001b[39m)\n\u001b[1;32m 1750\u001b[0m \u001b[38;5;28mprint\u001b[39m(\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mSaved GGUF to https://huggingface.co/\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mlink\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m\"\u001b[39m)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/unsloth/save.py:1339\u001b[0m, in \u001b[0;36mupload_to_huggingface\u001b[0;34m(model, save_directory, token, method, extra, file_location, old_username, private)\u001b[0m\n\u001b[1;32m 1336\u001b[0m uploaded_location \u001b[38;5;241m=\u001b[39m file_location\n\u001b[1;32m 1337\u001b[0m \u001b[38;5;28;01mpass\u001b[39;00m\n\u001b[0;32m-> 1339\u001b[0m \u001b[43mhf_api\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mupload_file\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 1340\u001b[0m \u001b[43m \u001b[49m\u001b[43mpath_or_fileobj\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[43mfile_location\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1341\u001b[0m \u001b[43m \u001b[49m\u001b[43mpath_in_repo\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[43muploaded_location\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1342\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[43msave_directory\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1343\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mmodel\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1344\u001b[0m \u001b[43m \u001b[49m\u001b[43mcommit_message\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43m(Trained with Unsloth)\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\n\u001b[1;32m 1345\u001b[0m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 1347\u001b[0m \u001b[38;5;66;03m# We also upload a config.json file\u001b[39;00m\n\u001b[1;32m 1348\u001b[0m \u001b[38;5;28;01mimport\u001b[39;00m \u001b[38;5;21;01mjson\u001b[39;00m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:114\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 111\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 112\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 114\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:1286\u001b[0m, in \u001b[0;36mfuture_compatible.._inner\u001b[0;34m(self, *args, **kwargs)\u001b[0m\n\u001b[1;32m 1283\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mrun_as_future(fn, \u001b[38;5;28mself\u001b[39m, \u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\u001b[1;32m 1285\u001b[0m \u001b[38;5;66;03m# Otherwise, call the function normally\u001b[39;00m\n\u001b[0;32m-> 1286\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:4374\u001b[0m, in \u001b[0;36mHfApi.upload_file\u001b[0;34m(self, path_or_fileobj, path_in_repo, repo_id, token, repo_type, revision, commit_message, commit_description, create_pr, parent_commit, run_as_future)\u001b[0m\n\u001b[1;32m 4366\u001b[0m commit_message \u001b[38;5;241m=\u001b[39m (\n\u001b[1;32m 4367\u001b[0m commit_message \u001b[38;5;28;01mif\u001b[39;00m commit_message \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m \u001b[38;5;28;01melse\u001b[39;00m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mUpload \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mpath_in_repo\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m with huggingface_hub\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 4368\u001b[0m )\n\u001b[1;32m 4369\u001b[0m operation \u001b[38;5;241m=\u001b[39m CommitOperationAdd(\n\u001b[1;32m 4370\u001b[0m path_or_fileobj\u001b[38;5;241m=\u001b[39mpath_or_fileobj,\n\u001b[1;32m 4371\u001b[0m path_in_repo\u001b[38;5;241m=\u001b[39mpath_in_repo,\n\u001b[1;32m 4372\u001b[0m )\n\u001b[0;32m-> 4374\u001b[0m commit_info \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mcreate_commit\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 4375\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4376\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_type\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4377\u001b[0m \u001b[43m \u001b[49m\u001b[43moperations\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m[\u001b[49m\u001b[43moperation\u001b[49m\u001b[43m]\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4378\u001b[0m \u001b[43m \u001b[49m\u001b[43mcommit_message\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcommit_message\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4379\u001b[0m \u001b[43m \u001b[49m\u001b[43mcommit_description\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcommit_description\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4380\u001b[0m \u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtoken\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4381\u001b[0m \u001b[43m \u001b[49m\u001b[43mrevision\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrevision\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4382\u001b[0m \u001b[43m \u001b[49m\u001b[43mcreate_pr\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcreate_pr\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4383\u001b[0m \u001b[43m \u001b[49m\u001b[43mparent_commit\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mparent_commit\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4384\u001b[0m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 4386\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m commit_info\u001b[38;5;241m.\u001b[39mpr_url \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[1;32m 4387\u001b[0m revision \u001b[38;5;241m=\u001b[39m quote(_parse_revision_from_pr_url(commit_info\u001b[38;5;241m.\u001b[39mpr_url), safe\u001b[38;5;241m=\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:114\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 111\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 112\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 114\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:1286\u001b[0m, in \u001b[0;36mfuture_compatible.._inner\u001b[0;34m(self, *args, **kwargs)\u001b[0m\n\u001b[1;32m 1283\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mrun_as_future(fn, \u001b[38;5;28mself\u001b[39m, \u001b[38;5;241m*\u001b[39margs, \u001b[38;5;241m*\u001b[39m\u001b[38;5;241m*\u001b[39mkwargs)\n\u001b[1;32m 1285\u001b[0m \u001b[38;5;66;03m# Otherwise, call the function normally\u001b[39;00m\n\u001b[0;32m-> 1286\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:3677\u001b[0m, in \u001b[0;36mHfApi.create_commit\u001b[0;34m(self, repo_id, operations, commit_message, commit_description, token, repo_type, revision, create_pr, num_threads, parent_commit, run_as_future)\u001b[0m\n\u001b[1;32m 3674\u001b[0m \u001b[38;5;66;03m# If updating twice the same file or update then delete a file in a single commit\u001b[39;00m\n\u001b[1;32m 3675\u001b[0m _warn_on_overwriting_operations(operations)\n\u001b[0;32m-> 3677\u001b[0m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mpreupload_lfs_files\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 3678\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3679\u001b[0m \u001b[43m \u001b[49m\u001b[43madditions\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43madditions\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3680\u001b[0m \u001b[43m \u001b[49m\u001b[43mtoken\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtoken\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3681\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_type\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3682\u001b[0m \u001b[43m \u001b[49m\u001b[43mrevision\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43munquoted_revision\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;66;43;03m# first-class methods take unquoted revision\u001b[39;49;00m\n\u001b[1;32m 3683\u001b[0m \u001b[43m \u001b[49m\u001b[43mcreate_pr\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mcreate_pr\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3684\u001b[0m \u001b[43m \u001b[49m\u001b[43mnum_threads\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mnum_threads\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 3685\u001b[0m \u001b[43m \u001b[49m\u001b[43mfree_memory\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mFalse\u001b[39;49;00m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;66;43;03m# do not remove `CommitOperationAdd.path_or_fileobj` on LFS files for \"normal\" users\u001b[39;49;00m\n\u001b[1;32m 3686\u001b[0m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 3687\u001b[0m files_to_copy \u001b[38;5;241m=\u001b[39m _fetch_files_to_copy(\n\u001b[1;32m 3688\u001b[0m copies\u001b[38;5;241m=\u001b[39mcopies,\n\u001b[1;32m 3689\u001b[0m repo_type\u001b[38;5;241m=\u001b[39mrepo_type,\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 3693\u001b[0m endpoint\u001b[38;5;241m=\u001b[39m\u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mendpoint,\n\u001b[1;32m 3694\u001b[0m )\n\u001b[1;32m 3695\u001b[0m commit_payload \u001b[38;5;241m=\u001b[39m _prepare_commit_payload(\n\u001b[1;32m 3696\u001b[0m operations\u001b[38;5;241m=\u001b[39moperations,\n\u001b[1;32m 3697\u001b[0m files_to_copy\u001b[38;5;241m=\u001b[39mfiles_to_copy,\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 3700\u001b[0m parent_commit\u001b[38;5;241m=\u001b[39mparent_commit,\n\u001b[1;32m 3701\u001b[0m )\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/hf_api.py:4184\u001b[0m, in \u001b[0;36mHfApi.preupload_lfs_files\u001b[0;34m(self, repo_id, additions, token, repo_type, revision, create_pr, num_threads, free_memory, gitignore_content)\u001b[0m\n\u001b[1;32m 4178\u001b[0m logger\u001b[38;5;241m.\u001b[39minfo(\n\u001b[1;32m 4179\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mSkipped upload for \u001b[39m\u001b[38;5;132;01m{\u001b[39;00m\u001b[38;5;28mlen\u001b[39m(new_lfs_additions)\u001b[38;5;250m \u001b[39m\u001b[38;5;241m-\u001b[39m\u001b[38;5;250m \u001b[39m\u001b[38;5;28mlen\u001b[39m(new_lfs_additions_to_upload)\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m LFS file(s) \u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 4180\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m(ignored by gitignore file).\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 4181\u001b[0m )\n\u001b[1;32m 4183\u001b[0m \u001b[38;5;66;03m# Upload new LFS files\u001b[39;00m\n\u001b[0;32m-> 4184\u001b[0m \u001b[43m_upload_lfs_files\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 4185\u001b[0m \u001b[43m \u001b[49m\u001b[43madditions\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mnew_lfs_additions_to_upload\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4186\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_type\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_type\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4187\u001b[0m \u001b[43m \u001b[49m\u001b[43mrepo_id\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrepo_id\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4188\u001b[0m \u001b[43m \u001b[49m\u001b[43mheaders\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mheaders\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4189\u001b[0m \u001b[43m \u001b[49m\u001b[43mendpoint\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mendpoint\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4190\u001b[0m \u001b[43m \u001b[49m\u001b[43mnum_threads\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mnum_threads\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 4191\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;66;43;03m# If `create_pr`, we don't want to check user permission on the revision as users with read permission\u001b[39;49;00m\n\u001b[1;32m 4192\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;66;43;03m# should still be able to create PRs even if they don't have write permission on the target branch of the\u001b[39;49;00m\n\u001b[1;32m 4193\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;66;43;03m# PR (i.e. `revision`).\u001b[39;49;00m\n\u001b[1;32m 4194\u001b[0m \u001b[43m \u001b[49m\u001b[43mrevision\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrevision\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;28;43;01mif\u001b[39;49;00m\u001b[43m \u001b[49m\u001b[38;5;129;43;01mnot\u001b[39;49;00m\u001b[43m \u001b[49m\u001b[43mcreate_pr\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;28;43;01melse\u001b[39;49;00m\u001b[43m \u001b[49m\u001b[38;5;28;43;01mNone\u001b[39;49;00m\u001b[43m,\u001b[49m\n\u001b[1;32m 4195\u001b[0m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 4196\u001b[0m \u001b[38;5;28;01mfor\u001b[39;00m addition \u001b[38;5;129;01min\u001b[39;00m new_lfs_additions_to_upload:\n\u001b[1;32m 4197\u001b[0m addition\u001b[38;5;241m.\u001b[39m_is_uploaded \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mTrue\u001b[39;00m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py:114\u001b[0m, in \u001b[0;36mvalidate_hf_hub_args.._inner_fn\u001b[0;34m(*args, **kwargs)\u001b[0m\n\u001b[1;32m 111\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m check_use_auth_token:\n\u001b[1;32m 112\u001b[0m kwargs \u001b[38;5;241m=\u001b[39m smoothly_deprecate_use_auth_token(fn_name\u001b[38;5;241m=\u001b[39mfn\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__name__\u001b[39m, has_token\u001b[38;5;241m=\u001b[39mhas_token, kwargs\u001b[38;5;241m=\u001b[39mkwargs)\n\u001b[0;32m--> 114\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[43mfn\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/_commit_api.py:413\u001b[0m, in \u001b[0;36m_upload_lfs_files\u001b[0;34m(additions, repo_type, repo_id, headers, endpoint, num_threads, revision)\u001b[0m\n\u001b[1;32m 411\u001b[0m \u001b[38;5;28;01melif\u001b[39;00m \u001b[38;5;28mlen\u001b[39m(filtered_actions) \u001b[38;5;241m==\u001b[39m \u001b[38;5;241m1\u001b[39m:\n\u001b[1;32m 412\u001b[0m logger\u001b[38;5;241m.\u001b[39mdebug(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mUploading 1 LFS file to the Hub\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[0;32m--> 413\u001b[0m \u001b[43m_wrapped_lfs_upload\u001b[49m\u001b[43m(\u001b[49m\u001b[43mfiltered_actions\u001b[49m\u001b[43m[\u001b[49m\u001b[38;5;241;43m0\u001b[39;49m\u001b[43m]\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 414\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[1;32m 415\u001b[0m logger\u001b[38;5;241m.\u001b[39mdebug(\n\u001b[1;32m 416\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mUploading \u001b[39m\u001b[38;5;132;01m{\u001b[39;00m\u001b[38;5;28mlen\u001b[39m(filtered_actions)\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m LFS files to the Hub using up to \u001b[39m\u001b[38;5;132;01m{\u001b[39;00mnum_threads\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m threads concurrently\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 417\u001b[0m )\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/_commit_api.py:403\u001b[0m, in \u001b[0;36m_upload_lfs_files.._wrapped_lfs_upload\u001b[0;34m(batch_action)\u001b[0m\n\u001b[1;32m 401\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[1;32m 402\u001b[0m operation \u001b[38;5;241m=\u001b[39m oid2addop[batch_action[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124moid\u001b[39m\u001b[38;5;124m\"\u001b[39m]]\n\u001b[0;32m--> 403\u001b[0m \u001b[43mlfs_upload\u001b[49m\u001b[43m(\u001b[49m\u001b[43moperation\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43moperation\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mlfs_batch_action\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mbatch_action\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mheaders\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mheaders\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mendpoint\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mendpoint\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 404\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m \u001b[38;5;167;01mException\u001b[39;00m \u001b[38;5;28;01mas\u001b[39;00m exc:\n\u001b[1;32m 405\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mRuntimeError\u001b[39;00m(\u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mError while uploading \u001b[39m\u001b[38;5;124m'\u001b[39m\u001b[38;5;132;01m{\u001b[39;00moperation\u001b[38;5;241m.\u001b[39mpath_in_repo\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m to the Hub.\u001b[39m\u001b[38;5;124m\"\u001b[39m) \u001b[38;5;28;01mfrom\u001b[39;00m \u001b[38;5;21;01mexc\u001b[39;00m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/lfs.py:243\u001b[0m, in \u001b[0;36mlfs_upload\u001b[0;34m(operation, lfs_batch_action, token, headers, endpoint)\u001b[0m\n\u001b[1;32m 239\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m (\u001b[38;5;167;01mValueError\u001b[39;00m, \u001b[38;5;167;01mTypeError\u001b[39;00m):\n\u001b[1;32m 240\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mValueError\u001b[39;00m(\n\u001b[1;32m 241\u001b[0m \u001b[38;5;124mf\u001b[39m\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mMalformed response from LFS batch endpoint: `chunk_size` should be an integer. Got \u001b[39m\u001b[38;5;124m'\u001b[39m\u001b[38;5;132;01m{\u001b[39;00mchunk_size\u001b[38;5;132;01m}\u001b[39;00m\u001b[38;5;124m'\u001b[39m\u001b[38;5;124m.\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 242\u001b[0m )\n\u001b[0;32m--> 243\u001b[0m \u001b[43m_upload_multi_part\u001b[49m\u001b[43m(\u001b[49m\u001b[43moperation\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43moperation\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mheader\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mheader\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mchunk_size\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mchunk_size\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mupload_url\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mupload_url\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 244\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[1;32m 245\u001b[0m _upload_single_part(operation\u001b[38;5;241m=\u001b[39moperation, upload_url\u001b[38;5;241m=\u001b[39mupload_url)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/lfs.py:341\u001b[0m, in \u001b[0;36m_upload_multi_part\u001b[0;34m(operation, header, chunk_size, upload_url)\u001b[0m\n\u001b[1;32m 332\u001b[0m warnings\u001b[38;5;241m.\u001b[39mwarn(\n\u001b[1;32m 333\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mhf_transfer is enabled but does not support uploading from bytes or BinaryIO, falling back to regular\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 334\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124m upload\u001b[39m\u001b[38;5;124m\"\u001b[39m\n\u001b[1;32m 335\u001b[0m )\n\u001b[1;32m 336\u001b[0m use_hf_transfer \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mFalse\u001b[39;00m\n\u001b[1;32m 338\u001b[0m response_headers \u001b[38;5;241m=\u001b[39m (\n\u001b[1;32m 339\u001b[0m _upload_parts_hf_transfer(operation\u001b[38;5;241m=\u001b[39moperation, sorted_parts_urls\u001b[38;5;241m=\u001b[39msorted_parts_urls, chunk_size\u001b[38;5;241m=\u001b[39mchunk_size)\n\u001b[1;32m 340\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m use_hf_transfer\n\u001b[0;32m--> 341\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m \u001b[43m_upload_parts_iteratively\u001b[49m\u001b[43m(\u001b[49m\u001b[43moperation\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43moperation\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43msorted_parts_urls\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43msorted_parts_urls\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mchunk_size\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mchunk_size\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 342\u001b[0m )\n\u001b[1;32m 344\u001b[0m \u001b[38;5;66;03m# 3. Send completion request\u001b[39;00m\n\u001b[1;32m 345\u001b[0m completion_res \u001b[38;5;241m=\u001b[39m get_session()\u001b[38;5;241m.\u001b[39mpost(\n\u001b[1;32m 346\u001b[0m upload_url,\n\u001b[1;32m 347\u001b[0m json\u001b[38;5;241m=\u001b[39m_get_completion_payload(response_headers, operation\u001b[38;5;241m.\u001b[39mupload_info\u001b[38;5;241m.\u001b[39msha256\u001b[38;5;241m.\u001b[39mhex()),\n\u001b[1;32m 348\u001b[0m headers\u001b[38;5;241m=\u001b[39mLFS_HEADERS,\n\u001b[1;32m 349\u001b[0m )\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/lfs.py:398\u001b[0m, in \u001b[0;36m_upload_parts_iteratively\u001b[0;34m(operation, sorted_parts_urls, chunk_size)\u001b[0m\n\u001b[1;32m 391\u001b[0m \u001b[38;5;28;01mfor\u001b[39;00m part_idx, part_upload_url \u001b[38;5;129;01min\u001b[39;00m \u001b[38;5;28menumerate\u001b[39m(sorted_parts_urls):\n\u001b[1;32m 392\u001b[0m \u001b[38;5;28;01mwith\u001b[39;00m SliceFileObj(\n\u001b[1;32m 393\u001b[0m fileobj,\n\u001b[1;32m 394\u001b[0m seek_from\u001b[38;5;241m=\u001b[39mchunk_size \u001b[38;5;241m*\u001b[39m part_idx,\n\u001b[1;32m 395\u001b[0m read_limit\u001b[38;5;241m=\u001b[39mchunk_size,\n\u001b[1;32m 396\u001b[0m ) \u001b[38;5;28;01mas\u001b[39;00m fileobj_slice:\n\u001b[1;32m 397\u001b[0m \u001b[38;5;66;03m# S3 might raise a transient 500 error -> let's retry if that happens\u001b[39;00m\n\u001b[0;32m--> 398\u001b[0m part_upload_res \u001b[38;5;241m=\u001b[39m \u001b[43mhttp_backoff\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 399\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[38;5;124;43mPUT\u001b[39;49m\u001b[38;5;124;43m\"\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mpart_upload_url\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mdata\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mfileobj_slice\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mretry_on_status_codes\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43m(\u001b[49m\u001b[38;5;241;43m500\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m502\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m503\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m504\u001b[39;49m\u001b[43m)\u001b[49m\n\u001b[1;32m 400\u001b[0m \u001b[43m \u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 401\u001b[0m hf_raise_for_status(part_upload_res)\n\u001b[1;32m 402\u001b[0m headers\u001b[38;5;241m.\u001b[39mappend(part_upload_res\u001b[38;5;241m.\u001b[39mheaders)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_http.py:280\u001b[0m, in \u001b[0;36mhttp_backoff\u001b[0;34m(method, url, max_retries, base_wait_time, max_wait_time, retry_on_exceptions, retry_on_status_codes, **kwargs)\u001b[0m\n\u001b[1;32m 277\u001b[0m kwargs[\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mdata\u001b[39m\u001b[38;5;124m\"\u001b[39m]\u001b[38;5;241m.\u001b[39mseek(io_obj_initial_pos)\n\u001b[1;32m 279\u001b[0m \u001b[38;5;66;03m# Perform request and return if status_code is not in the retry list.\u001b[39;00m\n\u001b[0;32m--> 280\u001b[0m response \u001b[38;5;241m=\u001b[39m \u001b[43msession\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mrequest\u001b[49m\u001b[43m(\u001b[49m\u001b[43mmethod\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mmethod\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43murl\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43murl\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 281\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m response\u001b[38;5;241m.\u001b[39mstatus_code \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;129;01min\u001b[39;00m retry_on_status_codes:\n\u001b[1;32m 282\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m response\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/requests/sessions.py:589\u001b[0m, in \u001b[0;36mSession.request\u001b[0;34m(self, method, url, params, data, headers, cookies, files, auth, timeout, allow_redirects, proxies, hooks, stream, verify, cert, json)\u001b[0m\n\u001b[1;32m 584\u001b[0m send_kwargs \u001b[38;5;241m=\u001b[39m {\n\u001b[1;32m 585\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mtimeout\u001b[39m\u001b[38;5;124m\"\u001b[39m: timeout,\n\u001b[1;32m 586\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mallow_redirects\u001b[39m\u001b[38;5;124m\"\u001b[39m: allow_redirects,\n\u001b[1;32m 587\u001b[0m }\n\u001b[1;32m 588\u001b[0m send_kwargs\u001b[38;5;241m.\u001b[39mupdate(settings)\n\u001b[0;32m--> 589\u001b[0m resp \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43msend\u001b[49m\u001b[43m(\u001b[49m\u001b[43mprep\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43msend_kwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 591\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m resp\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/requests/sessions.py:703\u001b[0m, in \u001b[0;36mSession.send\u001b[0;34m(self, request, **kwargs)\u001b[0m\n\u001b[1;32m 700\u001b[0m start \u001b[38;5;241m=\u001b[39m preferred_clock()\n\u001b[1;32m 702\u001b[0m \u001b[38;5;66;03m# Send the request\u001b[39;00m\n\u001b[0;32m--> 703\u001b[0m r \u001b[38;5;241m=\u001b[39m \u001b[43madapter\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43msend\u001b[49m\u001b[43m(\u001b[49m\u001b[43mrequest\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 705\u001b[0m \u001b[38;5;66;03m# Total elapsed time of the request (approximately)\u001b[39;00m\n\u001b[1;32m 706\u001b[0m elapsed \u001b[38;5;241m=\u001b[39m preferred_clock() \u001b[38;5;241m-\u001b[39m start\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_http.py:66\u001b[0m, in \u001b[0;36mUniqueRequestIdAdapter.send\u001b[0;34m(self, request, *args, **kwargs)\u001b[0m\n\u001b[1;32m 64\u001b[0m \u001b[38;5;250m\u001b[39m\u001b[38;5;124;03m\"\"\"Catch any RequestException to append request id to the error message for debugging.\"\"\"\u001b[39;00m\n\u001b[1;32m 65\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m---> 66\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43msuper\u001b[39;49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43msend\u001b[49m\u001b[43m(\u001b[49m\u001b[43mrequest\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43margs\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mkwargs\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 67\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m requests\u001b[38;5;241m.\u001b[39mRequestException \u001b[38;5;28;01mas\u001b[39;00m e:\n\u001b[1;32m 68\u001b[0m request_id \u001b[38;5;241m=\u001b[39m request\u001b[38;5;241m.\u001b[39mheaders\u001b[38;5;241m.\u001b[39mget(X_AMZN_TRACE_ID)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/requests/adapters.py:667\u001b[0m, in \u001b[0;36mHTTPAdapter.send\u001b[0;34m(self, request, stream, timeout, verify, cert, proxies)\u001b[0m\n\u001b[1;32m 664\u001b[0m timeout \u001b[38;5;241m=\u001b[39m TimeoutSauce(connect\u001b[38;5;241m=\u001b[39mtimeout, read\u001b[38;5;241m=\u001b[39mtimeout)\n\u001b[1;32m 666\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m--> 667\u001b[0m resp \u001b[38;5;241m=\u001b[39m \u001b[43mconn\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43murlopen\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 668\u001b[0m \u001b[43m \u001b[49m\u001b[43mmethod\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrequest\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mmethod\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 669\u001b[0m \u001b[43m \u001b[49m\u001b[43murl\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43murl\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 670\u001b[0m \u001b[43m \u001b[49m\u001b[43mbody\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrequest\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mbody\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 671\u001b[0m \u001b[43m \u001b[49m\u001b[43mheaders\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mrequest\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mheaders\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 672\u001b[0m \u001b[43m \u001b[49m\u001b[43mredirect\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mFalse\u001b[39;49;00m\u001b[43m,\u001b[49m\n\u001b[1;32m 673\u001b[0m \u001b[43m \u001b[49m\u001b[43massert_same_host\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mFalse\u001b[39;49;00m\u001b[43m,\u001b[49m\n\u001b[1;32m 674\u001b[0m \u001b[43m \u001b[49m\u001b[43mpreload_content\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mFalse\u001b[39;49;00m\u001b[43m,\u001b[49m\n\u001b[1;32m 675\u001b[0m \u001b[43m \u001b[49m\u001b[43mdecode_content\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43;01mFalse\u001b[39;49;00m\u001b[43m,\u001b[49m\n\u001b[1;32m 676\u001b[0m \u001b[43m \u001b[49m\u001b[43mretries\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mmax_retries\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 677\u001b[0m \u001b[43m \u001b[49m\u001b[43mtimeout\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtimeout\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 678\u001b[0m \u001b[43m \u001b[49m\u001b[43mchunked\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mchunked\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 679\u001b[0m \u001b[43m \u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 681\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m (ProtocolError, \u001b[38;5;167;01mOSError\u001b[39;00m) \u001b[38;5;28;01mas\u001b[39;00m err:\n\u001b[1;32m 682\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mConnectionError\u001b[39;00m(err, request\u001b[38;5;241m=\u001b[39mrequest)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py:793\u001b[0m, in \u001b[0;36mHTTPConnectionPool.urlopen\u001b[0;34m(self, method, url, body, headers, retries, redirect, assert_same_host, timeout, pool_timeout, release_conn, chunked, body_pos, preload_content, decode_content, **response_kw)\u001b[0m\n\u001b[1;32m 790\u001b[0m response_conn \u001b[38;5;241m=\u001b[39m conn \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m release_conn \u001b[38;5;28;01melse\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m\n\u001b[1;32m 792\u001b[0m \u001b[38;5;66;03m# Make the request on the HTTPConnection object\u001b[39;00m\n\u001b[0;32m--> 793\u001b[0m response \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_make_request\u001b[49m\u001b[43m(\u001b[49m\n\u001b[1;32m 794\u001b[0m \u001b[43m \u001b[49m\u001b[43mconn\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 795\u001b[0m \u001b[43m \u001b[49m\u001b[43mmethod\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 796\u001b[0m \u001b[43m \u001b[49m\u001b[43murl\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 797\u001b[0m \u001b[43m \u001b[49m\u001b[43mtimeout\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mtimeout_obj\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 798\u001b[0m \u001b[43m \u001b[49m\u001b[43mbody\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mbody\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 799\u001b[0m \u001b[43m \u001b[49m\u001b[43mheaders\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mheaders\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 800\u001b[0m \u001b[43m \u001b[49m\u001b[43mchunked\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mchunked\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 801\u001b[0m \u001b[43m \u001b[49m\u001b[43mretries\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mretries\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 802\u001b[0m \u001b[43m \u001b[49m\u001b[43mresponse_conn\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mresponse_conn\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 803\u001b[0m \u001b[43m \u001b[49m\u001b[43mpreload_content\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mpreload_content\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 804\u001b[0m \u001b[43m \u001b[49m\u001b[43mdecode_content\u001b[49m\u001b[38;5;241;43m=\u001b[39;49m\u001b[43mdecode_content\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 805\u001b[0m \u001b[43m \u001b[49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[38;5;241;43m*\u001b[39;49m\u001b[43mresponse_kw\u001b[49m\u001b[43m,\u001b[49m\n\u001b[1;32m 806\u001b[0m \u001b[43m\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 808\u001b[0m \u001b[38;5;66;03m# Everything went great!\u001b[39;00m\n\u001b[1;32m 809\u001b[0m clean_exit \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mTrue\u001b[39;00m\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/urllib3/connectionpool.py:537\u001b[0m, in \u001b[0;36mHTTPConnectionPool._make_request\u001b[0;34m(self, conn, method, url, body, headers, retries, timeout, chunked, response_conn, preload_content, decode_content, enforce_content_length)\u001b[0m\n\u001b[1;32m 535\u001b[0m \u001b[38;5;66;03m# Receive the response from the server\u001b[39;00m\n\u001b[1;32m 536\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m--> 537\u001b[0m response \u001b[38;5;241m=\u001b[39m \u001b[43mconn\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mgetresponse\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 538\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m (BaseSSLError, \u001b[38;5;167;01mOSError\u001b[39;00m) \u001b[38;5;28;01mas\u001b[39;00m e:\n\u001b[1;32m 539\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_raise_timeout(err\u001b[38;5;241m=\u001b[39me, url\u001b[38;5;241m=\u001b[39murl, timeout_value\u001b[38;5;241m=\u001b[39mread_timeout)\n", + "File \u001b[0;32m/usr/local/lib/python3.10/dist-packages/urllib3/connection.py:466\u001b[0m, in \u001b[0;36mHTTPConnection.getresponse\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 463\u001b[0m \u001b[38;5;28;01mfrom\u001b[39;00m \u001b[38;5;21;01m.\u001b[39;00m\u001b[38;5;21;01mresponse\u001b[39;00m \u001b[38;5;28;01mimport\u001b[39;00m HTTPResponse\n\u001b[1;32m 465\u001b[0m \u001b[38;5;66;03m# Get the response from http.client.HTTPConnection\u001b[39;00m\n\u001b[0;32m--> 466\u001b[0m httplib_response \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43msuper\u001b[39;49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mgetresponse\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 468\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[1;32m 469\u001b[0m assert_header_parsing(httplib_response\u001b[38;5;241m.\u001b[39mmsg)\n", + "File \u001b[0;32m/usr/lib/python3.10/http/client.py:1375\u001b[0m, in \u001b[0;36mHTTPConnection.getresponse\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 1373\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[1;32m 1374\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m-> 1375\u001b[0m \u001b[43mresponse\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mbegin\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 1376\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m \u001b[38;5;167;01mConnectionError\u001b[39;00m:\n\u001b[1;32m 1377\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39mclose()\n", + "File \u001b[0;32m/usr/lib/python3.10/http/client.py:318\u001b[0m, in \u001b[0;36mHTTPResponse.begin\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 316\u001b[0m \u001b[38;5;66;03m# read until we get a non-100 response\u001b[39;00m\n\u001b[1;32m 317\u001b[0m \u001b[38;5;28;01mwhile\u001b[39;00m \u001b[38;5;28;01mTrue\u001b[39;00m:\n\u001b[0;32m--> 318\u001b[0m version, status, reason \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_read_status\u001b[49m\u001b[43m(\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 319\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m status \u001b[38;5;241m!=\u001b[39m CONTINUE:\n\u001b[1;32m 320\u001b[0m \u001b[38;5;28;01mbreak\u001b[39;00m\n", + "File \u001b[0;32m/usr/lib/python3.10/http/client.py:279\u001b[0m, in \u001b[0;36mHTTPResponse._read_status\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 278\u001b[0m \u001b[38;5;28;01mdef\u001b[39;00m \u001b[38;5;21m_read_status\u001b[39m(\u001b[38;5;28mself\u001b[39m):\n\u001b[0;32m--> 279\u001b[0m line \u001b[38;5;241m=\u001b[39m \u001b[38;5;28mstr\u001b[39m(\u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mfp\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mreadline\u001b[49m\u001b[43m(\u001b[49m\u001b[43m_MAXLINE\u001b[49m\u001b[43m \u001b[49m\u001b[38;5;241;43m+\u001b[39;49m\u001b[43m \u001b[49m\u001b[38;5;241;43m1\u001b[39;49m\u001b[43m)\u001b[49m, \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124miso-8859-1\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n\u001b[1;32m 280\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m \u001b[38;5;28mlen\u001b[39m(line) \u001b[38;5;241m>\u001b[39m _MAXLINE:\n\u001b[1;32m 281\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m LineTooLong(\u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mstatus line\u001b[39m\u001b[38;5;124m\"\u001b[39m)\n", + "File \u001b[0;32m/usr/lib/python3.10/socket.py:705\u001b[0m, in \u001b[0;36mSocketIO.readinto\u001b[0;34m(self, b)\u001b[0m\n\u001b[1;32m 703\u001b[0m \u001b[38;5;28;01mwhile\u001b[39;00m \u001b[38;5;28;01mTrue\u001b[39;00m:\n\u001b[1;32m 704\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[0;32m--> 705\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_sock\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mrecv_into\u001b[49m\u001b[43m(\u001b[49m\u001b[43mb\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 706\u001b[0m \u001b[38;5;28;01mexcept\u001b[39;00m timeout:\n\u001b[1;32m 707\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_timeout_occurred \u001b[38;5;241m=\u001b[39m \u001b[38;5;28;01mTrue\u001b[39;00m\n", + "File \u001b[0;32m/usr/lib/python3.10/ssl.py:1303\u001b[0m, in \u001b[0;36mSSLSocket.recv_into\u001b[0;34m(self, buffer, nbytes, flags)\u001b[0m\n\u001b[1;32m 1299\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m flags \u001b[38;5;241m!=\u001b[39m \u001b[38;5;241m0\u001b[39m:\n\u001b[1;32m 1300\u001b[0m \u001b[38;5;28;01mraise\u001b[39;00m \u001b[38;5;167;01mValueError\u001b[39;00m(\n\u001b[1;32m 1301\u001b[0m \u001b[38;5;124m\"\u001b[39m\u001b[38;5;124mnon-zero flags not allowed in calls to recv_into() on \u001b[39m\u001b[38;5;132;01m%s\u001b[39;00m\u001b[38;5;124m\"\u001b[39m \u001b[38;5;241m%\u001b[39m\n\u001b[1;32m 1302\u001b[0m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m\u001b[38;5;18m__class__\u001b[39m)\n\u001b[0;32m-> 1303\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mread\u001b[49m\u001b[43m(\u001b[49m\u001b[43mnbytes\u001b[49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mbuffer\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 1304\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[1;32m 1305\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28msuper\u001b[39m()\u001b[38;5;241m.\u001b[39mrecv_into(buffer, nbytes, flags)\n", + "File \u001b[0;32m/usr/lib/python3.10/ssl.py:1159\u001b[0m, in \u001b[0;36mSSLSocket.read\u001b[0;34m(self, len, buffer)\u001b[0m\n\u001b[1;32m 1157\u001b[0m \u001b[38;5;28;01mtry\u001b[39;00m:\n\u001b[1;32m 1158\u001b[0m \u001b[38;5;28;01mif\u001b[39;00m buffer \u001b[38;5;129;01mis\u001b[39;00m \u001b[38;5;129;01mnot\u001b[39;00m \u001b[38;5;28;01mNone\u001b[39;00m:\n\u001b[0;32m-> 1159\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28;43mself\u001b[39;49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43m_sslobj\u001b[49m\u001b[38;5;241;43m.\u001b[39;49m\u001b[43mread\u001b[49m\u001b[43m(\u001b[49m\u001b[38;5;28;43mlen\u001b[39;49m\u001b[43m,\u001b[49m\u001b[43m \u001b[49m\u001b[43mbuffer\u001b[49m\u001b[43m)\u001b[49m\n\u001b[1;32m 1160\u001b[0m \u001b[38;5;28;01melse\u001b[39;00m:\n\u001b[1;32m 1161\u001b[0m \u001b[38;5;28;01mreturn\u001b[39;00m \u001b[38;5;28mself\u001b[39m\u001b[38;5;241m.\u001b[39m_sslobj\u001b[38;5;241m.\u001b[39mread(\u001b[38;5;28mlen\u001b[39m)\n", + "\u001b[0;31mKeyboardInterrupt\u001b[0m: " + ] + } + ], + "source": [ + "#@title Saving, loading finetuned models\n", + "\n", + "# model.save_pretrained(\"dolphin-llama3-zh-cn-uncensored\") # Local saving\n", + "\n", + "model.push_to_hub_gguf(\"stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored\", tokenizer, quantization_method = \"not_quantized\", token = \"hf_xxxxxxxxxxxxxxxxxxxxxxxx\") # HF Online saving\n" + ] + }, + { + "cell_type": "code", + "execution_count": 18, + "metadata": {}, + "outputs": [ + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "1b37cfffc8944316815af937c8176673", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "README.md: 0%| | 0.00/2.54k [00:00' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'user\\n' + content + '\\nmodel\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '\\n' }}{% endif %}{% endfor %}\n", + "INFO:hf-to-gguf:Exporting model...\n", + "INFO:hf-to-gguf:gguf: loading model weight map from 'model.safetensors.index.json'\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00001-of-00004.safetensors'\n", + "INFO:hf-to-gguf:token_embd.weight, torch.bfloat16 --> Q8_0, shape = {3584, 256000}\n", + "INFO:hf-to-gguf:blk.0.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.0.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.0.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.0.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.1.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.1.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.2.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.2.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.3.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.3.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.4.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.4.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.5.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.5.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.6.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.6.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.7.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.7.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00002-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.10.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.10.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.10.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.10.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.11.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.11.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.12.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.12.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.13.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.13.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.14.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.14.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.15.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.15.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.16.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.16.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.17.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.17.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.18.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.18.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.19.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.19.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.20.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.20.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.7.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.8.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.8.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.8.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.9.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.9.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00003-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.20.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.21.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.21.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.21.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.22.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.22.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.23.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.23.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.24.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.24.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.25.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.25.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.26.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.26.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.27.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.27.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.28.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.28.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.29.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.29.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.30.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.30.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.31.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.31.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.32.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.32.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00004-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.32.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.32.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.33.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.33.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.33.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.34.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.34.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.35.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.35.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.36.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.36.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.37.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.37.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.38.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.38.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.39.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.39.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.40.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.40.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_down.weight, torch.bfloat16 --> Q8_0, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_gate.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.ffn_up.weight, torch.bfloat16 --> Q8_0, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.attn_k.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_output.weight, torch.bfloat16 --> Q8_0, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.41.attn_q.weight, torch.bfloat16 --> Q8_0, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.41.attn_v.weight, torch.bfloat16 --> Q8_0, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:output_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:gguf.gguf_writer:Writing the following files:\n", + "INFO:gguf.gguf_writer:stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.Q8_0.gguf: n_tensors = 464, total_size = 9.8G\n", + "Writing: 100%|██████████| 9.82G/9.82G [02:06<00:00, 77.4Mbyte/s]\n", + "INFO:hf-to-gguf:Model successfully exported to stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.Q8_0.gguf\n", + "Unsloth: Conversion completed! Output location: ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.Q8_0.gguf\n", + "Unsloth: Uploading GGUF to Huggingface Hub...\n" + ] + }, + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "91c304f3ef974f57a03f86314c68349c", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "unsloth.Q8_0.gguf: 0%| | 0.00/9.83G [00:00' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'user\\n' + content + '\\nmodel\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '\\n' }}{% endif %}{% endfor %}\n", + "INFO:hf-to-gguf:Exporting model...\n", + "INFO:hf-to-gguf:gguf: loading model weight map from 'model.safetensors.index.json'\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00001-of-00004.safetensors'\n", + "INFO:hf-to-gguf:token_embd.weight, torch.bfloat16 --> F16, shape = {3584, 256000}\n", + "INFO:hf-to-gguf:blk.0.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.0.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.0.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.0.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.1.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.1.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.2.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.2.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.3.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.3.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.4.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.4.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.5.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.5.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.6.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.6.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.7.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.7.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00002-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.10.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.10.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.10.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.10.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.11.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.11.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.12.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.12.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.13.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.13.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.14.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.14.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.15.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.15.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.16.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.16.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.17.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.17.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.18.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.18.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.19.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.19.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.20.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.20.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.7.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.8.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.8.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.8.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.9.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.9.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00003-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.20.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.21.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.21.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.21.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.22.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.22.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.23.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.23.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.24.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.24.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.25.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.25.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.26.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.26.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.27.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.27.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.28.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.28.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.29.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.29.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.30.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.30.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.31.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.31.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.32.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.32.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00004-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.32.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.32.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.33.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.33.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.33.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.34.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.34.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.35.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.35.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.36.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.36.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.37.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.37.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.38.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.38.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.39.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.39.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.40.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.40.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_down.weight, torch.bfloat16 --> F16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_gate.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.ffn_up.weight, torch.bfloat16 --> F16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.attn_k.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_output.weight, torch.bfloat16 --> F16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.41.attn_q.weight, torch.bfloat16 --> F16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.41.attn_v.weight, torch.bfloat16 --> F16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:output_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:gguf.gguf_writer:Writing the following files:\n", + "INFO:gguf.gguf_writer:stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.F16.gguf: n_tensors = 464, total_size = 18.5G\n", + "Writing: 100%|██████████| 18.5G/18.5G [01:29<00:00, 207Mbyte/s]\n", + "INFO:hf-to-gguf:Model successfully exported to stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.F16.gguf\n", + "Unsloth: Conversion completed! Output location: ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.F16.gguf\n", + "Unsloth: Uploading GGUF to Huggingface Hub...\n" + ] + }, + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "ea37f340b19e4a5589494c31b6ec8912", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "unsloth.F16.gguf: 0%| | 0.00/18.5G [00:00' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'user\\n' + content + '\\nmodel\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '\\n' }}{% endif %}{% endfor %}\n", + "INFO:hf-to-gguf:Exporting model...\n", + "INFO:hf-to-gguf:gguf: loading model weight map from 'model.safetensors.index.json'\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00001-of-00004.safetensors'\n", + "INFO:hf-to-gguf:token_embd.weight, torch.bfloat16 --> BF16, shape = {3584, 256000}\n", + "INFO:hf-to-gguf:blk.0.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.0.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.0.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.0.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.0.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.0.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.1.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.1.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.1.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.1.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.1.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.2.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.2.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.2.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.2.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.2.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.3.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.3.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.3.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.3.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.3.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.4.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.4.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.4.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.4.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.4.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.5.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.5.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.5.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.5.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.5.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.6.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.6.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.6.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.6.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.6.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.7.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.7.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.7.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00002-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.10.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.10.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.10.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.10.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.10.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.10.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.11.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.11.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.11.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.11.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.11.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.12.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.12.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.12.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.12.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.12.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.13.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.13.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.13.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.13.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.13.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.14.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.14.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.14.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.14.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.14.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.15.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.15.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.15.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.15.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.15.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.16.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.16.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.16.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.16.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.16.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.17.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.17.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.17.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.17.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.17.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.18.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.18.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.18.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.18.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.18.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.19.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.19.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.19.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.19.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.19.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.20.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.20.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.20.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.7.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.7.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.7.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.8.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.8.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.8.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.8.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.8.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.9.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.9.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.9.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.9.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.9.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00003-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.20.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.20.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.20.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.21.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.21.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.21.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.21.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.21.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.22.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.22.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.22.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.22.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.22.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.23.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.23.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.23.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.23.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.23.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.24.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.24.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.24.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.24.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.24.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.25.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.25.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.25.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.25.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.25.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.26.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.26.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.26.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.26.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.26.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.27.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.27.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.27.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.27.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.27.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.28.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.28.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.28.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.28.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.28.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.29.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.29.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.29.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.29.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.29.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.30.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.30.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.30.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.30.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.30.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.31.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.31.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.31.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.31.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.31.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.32.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.32.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.32.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.32.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:gguf: loading model part 'model-00004-of-00004.safetensors'\n", + "INFO:hf-to-gguf:blk.32.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.32.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.32.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.33.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.33.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.33.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.33.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.33.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.34.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.34.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.34.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.34.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.34.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.35.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.35.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.35.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.35.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.35.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.36.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.36.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.36.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.36.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.36.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.37.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.37.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.37.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.37.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.37.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.38.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.38.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.38.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.38.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.38.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.39.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.39.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.39.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.39.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.39.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.40.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.40.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.40.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.40.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.40.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_down.weight, torch.bfloat16 --> BF16, shape = {14336, 3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_gate.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.ffn_up.weight, torch.bfloat16 --> BF16, shape = {3584, 14336}\n", + "INFO:hf-to-gguf:blk.41.post_attention_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.post_ffw_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.ffn_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:hf-to-gguf:blk.41.attn_k.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:blk.41.attn_output.weight, torch.bfloat16 --> BF16, shape = {4096, 3584}\n", + "INFO:hf-to-gguf:blk.41.attn_q.weight, torch.bfloat16 --> BF16, shape = {3584, 4096}\n", + "INFO:hf-to-gguf:blk.41.attn_v.weight, torch.bfloat16 --> BF16, shape = {3584, 2048}\n", + "INFO:hf-to-gguf:output_norm.weight, torch.bfloat16 --> F32, shape = {3584}\n", + "INFO:gguf.gguf_writer:Writing the following files:\n", + "INFO:gguf.gguf_writer:stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf: n_tensors = 464, total_size = 18.5G\n", + "Writing: 100%|██████████| 18.5G/18.5G [01:57<00:00, 158Mbyte/s]\n", + "INFO:hf-to-gguf:Model successfully exported to stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf\n", + "Unsloth: Conversion completed! Output location: ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf\n", + "Unsloth: [2] Converting GGUF 16bit into q4_k_m. This will take 20 minutes...\n", + "main: build = 3391 (090fca7a)\n", + "main: built with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu\n", + "main: quantizing './stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf' to './stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.Q4_K_M.gguf' as Q4_K_M using 256 threads\n", + "llama_model_loader: loaded meta data with 29 key-value pairs and 464 tensors from ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.BF16.gguf (version GGUF V3 (latest))\n", + "llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.\n", + "llama_model_loader: - kv 0: general.architecture str = gemma2\n", + "llama_model_loader: - kv 1: general.name str = Gemma-2-9B-Chinese-Chat-Uncensored\n", + "llama_model_loader: - kv 2: gemma2.context_length u32 = 8192\n", + "llama_model_loader: - kv 3: gemma2.embedding_length u32 = 3584\n", + "llama_model_loader: - kv 4: gemma2.block_count u32 = 42\n", + "llama_model_loader: - kv 5: gemma2.feed_forward_length u32 = 14336\n", + "llama_model_loader: - kv 6: gemma2.attention.head_count u32 = 16\n", + "llama_model_loader: - kv 7: gemma2.attention.head_count_kv u32 = 8\n", + "llama_model_loader: - kv 8: gemma2.attention.layer_norm_rms_epsilon f32 = 0.000001\n", + "llama_model_loader: - kv 9: gemma2.attention.key_length u32 = 256\n", + "llama_model_loader: - kv 10: gemma2.attention.value_length u32 = 256\n", + "llama_model_loader: - kv 11: general.file_type u32 = 32\n", + "llama_model_loader: - kv 12: gemma2.attn_logit_softcapping f32 = 50.000000\n", + "llama_model_loader: - kv 13: gemma2.final_logit_softcapping f32 = 30.000000\n", + "llama_model_loader: - kv 14: gemma2.attention.sliding_window u32 = 4096\n", + "llama_model_loader: - kv 15: tokenizer.ggml.model str = llama\n", + "llama_model_loader: - kv 16: tokenizer.ggml.pre str = default\n", + "llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,256000] = [\"\", \"\", \"\", \"\", ...\n", + "llama_model_loader: - kv 18: tokenizer.ggml.scores arr[f32,256000] = [-1000.000000, -1000.000000, -1000.00...\n", + "llama_model_loader: - kv 19: tokenizer.ggml.token_type arr[i32,256000] = [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, ...\n", + "llama_model_loader: - kv 20: tokenizer.ggml.bos_token_id u32 = 2\n", + "llama_model_loader: - kv 21: tokenizer.ggml.eos_token_id u32 = 1\n", + "llama_model_loader: - kv 22: tokenizer.ggml.unknown_token_id u32 = 3\n", + "llama_model_loader: - kv 23: tokenizer.ggml.padding_token_id u32 = 0\n", + "llama_model_loader: - kv 24: tokenizer.ggml.add_bos_token bool = true\n", + "llama_model_loader: - kv 25: tokenizer.ggml.add_eos_token bool = false\n", + "llama_model_loader: - kv 26: tokenizer.chat_template str = {% if messages[0]['role'] == 'system'...\n", + "llama_model_loader: - kv 27: tokenizer.ggml.add_space_prefix bool = false\n", + "llama_model_loader: - kv 28: general.quantization_version u32 = 2\n", + "llama_model_loader: - type f32: 169 tensors\n", + "llama_model_loader: - type bf16: 295 tensors\n", + "[ 1/ 464] token_embd.weight - [ 3584, 256000, 1, 1], type = bf16, converting to q6_K .. size = 1750.00 MiB -> 717.77 MiB\n", + "[ 2/ 464] blk.0.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 3/ 464] blk.0.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 4/ 464] blk.0.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 5/ 464] blk.0.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 6/ 464] blk.0.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 7/ 464] blk.0.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 8/ 464] blk.0.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 9/ 464] blk.0.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 10/ 464] blk.0.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 11/ 464] blk.0.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 12/ 464] blk.0.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 13/ 464] blk.1.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 14/ 464] blk.1.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 15/ 464] blk.1.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 16/ 464] blk.1.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 17/ 464] blk.1.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 18/ 464] blk.1.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 19/ 464] blk.1.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 20/ 464] blk.1.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 21/ 464] blk.1.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 22/ 464] blk.1.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 23/ 464] blk.1.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 24/ 464] blk.2.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 25/ 464] blk.2.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 26/ 464] blk.2.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 27/ 464] blk.2.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 28/ 464] blk.2.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 29/ 464] blk.2.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 30/ 464] blk.2.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 31/ 464] blk.2.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 32/ 464] blk.2.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 33/ 464] blk.2.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 34/ 464] blk.2.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 35/ 464] blk.3.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 36/ 464] blk.3.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 37/ 464] blk.3.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 38/ 464] blk.3.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 39/ 464] blk.3.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 40/ 464] blk.3.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 41/ 464] blk.3.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 42/ 464] blk.3.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 43/ 464] blk.3.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 44/ 464] blk.3.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 45/ 464] blk.3.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 46/ 464] blk.4.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 47/ 464] blk.4.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 48/ 464] blk.4.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 49/ 464] blk.4.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 50/ 464] blk.4.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 51/ 464] blk.4.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 52/ 464] blk.4.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 53/ 464] blk.4.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 54/ 464] blk.4.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 55/ 464] blk.4.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 56/ 464] blk.4.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 57/ 464] blk.5.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 58/ 464] blk.5.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 59/ 464] blk.5.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 60/ 464] blk.5.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 61/ 464] blk.5.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 62/ 464] blk.5.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 63/ 464] blk.5.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 64/ 464] blk.5.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 65/ 464] blk.5.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 66/ 464] blk.5.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 67/ 464] blk.5.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 68/ 464] blk.6.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 69/ 464] blk.6.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 70/ 464] blk.6.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 71/ 464] blk.6.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 72/ 464] blk.6.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 73/ 464] blk.6.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 74/ 464] blk.6.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 75/ 464] blk.6.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 76/ 464] blk.6.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 77/ 464] blk.6.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 78/ 464] blk.6.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 79/ 464] blk.7.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 80/ 464] blk.7.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 81/ 464] blk.7.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 82/ 464] blk.7.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 83/ 464] blk.7.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 84/ 464] blk.7.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 85/ 464] blk.10.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 86/ 464] blk.10.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 87/ 464] blk.10.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 88/ 464] blk.10.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 89/ 464] blk.10.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 90/ 464] blk.10.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 91/ 464] blk.10.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 92/ 464] blk.10.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 93/ 464] blk.10.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 94/ 464] blk.10.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 95/ 464] blk.10.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 96/ 464] blk.11.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 97/ 464] blk.11.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 98/ 464] blk.11.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 99/ 464] blk.11.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 100/ 464] blk.11.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 101/ 464] blk.11.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 102/ 464] blk.11.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 103/ 464] blk.11.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 104/ 464] blk.11.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 105/ 464] blk.11.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 106/ 464] blk.11.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 107/ 464] blk.12.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 108/ 464] blk.12.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 109/ 464] blk.12.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 110/ 464] blk.12.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 111/ 464] blk.12.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 112/ 464] blk.12.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 113/ 464] blk.12.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 114/ 464] blk.12.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 115/ 464] blk.12.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 116/ 464] blk.12.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 117/ 464] blk.12.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 118/ 464] blk.13.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 119/ 464] blk.13.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 120/ 464] blk.13.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 121/ 464] blk.13.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 122/ 464] blk.13.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 123/ 464] blk.13.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 124/ 464] blk.13.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 125/ 464] blk.13.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 126/ 464] blk.13.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 127/ 464] blk.13.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 128/ 464] blk.13.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 129/ 464] blk.14.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 130/ 464] blk.14.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 131/ 464] blk.14.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 132/ 464] blk.14.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 133/ 464] blk.14.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 134/ 464] blk.14.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 135/ 464] blk.14.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 136/ 464] blk.14.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 137/ 464] blk.14.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 138/ 464] blk.14.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 139/ 464] blk.14.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 140/ 464] blk.15.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 141/ 464] blk.15.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 142/ 464] blk.15.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 143/ 464] blk.15.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 144/ 464] blk.15.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 145/ 464] blk.15.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 146/ 464] blk.15.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 147/ 464] blk.15.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 148/ 464] blk.15.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 149/ 464] blk.15.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 150/ 464] blk.15.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 151/ 464] blk.16.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 152/ 464] blk.16.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 153/ 464] blk.16.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 154/ 464] blk.16.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 155/ 464] blk.16.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 156/ 464] blk.16.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 157/ 464] blk.16.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 158/ 464] blk.16.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 159/ 464] blk.16.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 160/ 464] blk.16.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 161/ 464] blk.16.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 162/ 464] blk.17.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 163/ 464] blk.17.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 164/ 464] blk.17.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 165/ 464] blk.17.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 166/ 464] blk.17.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 167/ 464] blk.17.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 168/ 464] blk.17.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 169/ 464] blk.17.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 170/ 464] blk.17.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 171/ 464] blk.17.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 172/ 464] blk.17.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 173/ 464] blk.18.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 174/ 464] blk.18.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 175/ 464] blk.18.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 176/ 464] blk.18.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 177/ 464] blk.18.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 178/ 464] blk.18.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 179/ 464] blk.18.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 180/ 464] blk.18.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 181/ 464] blk.18.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 182/ 464] blk.18.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 183/ 464] blk.18.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 184/ 464] blk.19.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 185/ 464] blk.19.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 186/ 464] blk.19.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 187/ 464] blk.19.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 188/ 464] blk.19.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 189/ 464] blk.19.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 190/ 464] blk.19.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 191/ 464] blk.19.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 192/ 464] blk.19.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 193/ 464] blk.19.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 194/ 464] blk.19.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 195/ 464] blk.20.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 196/ 464] blk.20.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 197/ 464] blk.20.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 198/ 464] blk.20.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 199/ 464] blk.7.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 200/ 464] blk.7.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 201/ 464] blk.7.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 202/ 464] blk.7.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 203/ 464] blk.7.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 204/ 464] blk.8.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 205/ 464] blk.8.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 206/ 464] blk.8.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 207/ 464] blk.8.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 208/ 464] blk.8.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 209/ 464] blk.8.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 210/ 464] blk.8.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 211/ 464] blk.8.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 212/ 464] blk.8.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 213/ 464] blk.8.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 214/ 464] blk.8.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 215/ 464] blk.9.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 216/ 464] blk.9.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 217/ 464] blk.9.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 218/ 464] blk.9.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 219/ 464] blk.9.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 220/ 464] blk.9.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 221/ 464] blk.9.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 222/ 464] blk.9.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 223/ 464] blk.9.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 224/ 464] blk.9.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 225/ 464] blk.9.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 226/ 464] blk.20.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 227/ 464] blk.20.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 228/ 464] blk.20.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 229/ 464] blk.20.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 230/ 464] blk.20.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 231/ 464] blk.20.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 232/ 464] blk.20.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 233/ 464] blk.21.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 234/ 464] blk.21.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 235/ 464] blk.21.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 236/ 464] blk.21.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 237/ 464] blk.21.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 238/ 464] blk.21.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 239/ 464] blk.21.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 240/ 464] blk.21.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 241/ 464] blk.21.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 242/ 464] blk.21.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 243/ 464] blk.21.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 244/ 464] blk.22.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 245/ 464] blk.22.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 246/ 464] blk.22.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 247/ 464] blk.22.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 248/ 464] blk.22.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 249/ 464] blk.22.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 250/ 464] blk.22.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 251/ 464] blk.22.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 252/ 464] blk.22.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 253/ 464] blk.22.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 254/ 464] blk.22.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 255/ 464] blk.23.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 256/ 464] blk.23.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 257/ 464] blk.23.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 258/ 464] blk.23.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 259/ 464] blk.23.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 260/ 464] blk.23.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 261/ 464] blk.23.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 262/ 464] blk.23.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 263/ 464] blk.23.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 264/ 464] blk.23.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 265/ 464] blk.23.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 266/ 464] blk.24.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 267/ 464] blk.24.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 268/ 464] blk.24.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 269/ 464] blk.24.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 270/ 464] blk.24.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 271/ 464] blk.24.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 272/ 464] blk.24.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 273/ 464] blk.24.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 274/ 464] blk.24.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 275/ 464] blk.24.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 276/ 464] blk.24.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 277/ 464] blk.25.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 278/ 464] blk.25.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 279/ 464] blk.25.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 280/ 464] blk.25.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 281/ 464] blk.25.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 282/ 464] blk.25.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 283/ 464] blk.25.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 284/ 464] blk.25.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 285/ 464] blk.25.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 286/ 464] blk.25.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 287/ 464] blk.25.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 288/ 464] blk.26.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 289/ 464] blk.26.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 290/ 464] blk.26.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 291/ 464] blk.26.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 292/ 464] blk.26.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 293/ 464] blk.26.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 294/ 464] blk.26.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 295/ 464] blk.26.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 296/ 464] blk.26.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 297/ 464] blk.26.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 298/ 464] blk.26.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 299/ 464] blk.27.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 300/ 464] blk.27.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 301/ 464] blk.27.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 302/ 464] blk.27.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 303/ 464] blk.27.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 304/ 464] blk.27.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 305/ 464] blk.27.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 306/ 464] blk.27.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 307/ 464] blk.27.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 308/ 464] blk.27.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 309/ 464] blk.27.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 310/ 464] blk.28.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 311/ 464] blk.28.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 312/ 464] blk.28.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 313/ 464] blk.28.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 314/ 464] blk.28.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 315/ 464] blk.28.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 316/ 464] blk.28.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 317/ 464] blk.28.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 318/ 464] blk.28.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 319/ 464] blk.28.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 320/ 464] blk.28.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 321/ 464] blk.29.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 322/ 464] blk.29.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 323/ 464] blk.29.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 324/ 464] blk.29.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 325/ 464] blk.29.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 326/ 464] blk.29.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 327/ 464] blk.29.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 328/ 464] blk.29.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 329/ 464] blk.29.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 330/ 464] blk.29.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 331/ 464] blk.29.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 332/ 464] blk.30.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 333/ 464] blk.30.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 334/ 464] blk.30.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 335/ 464] blk.30.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 336/ 464] blk.30.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 337/ 464] blk.30.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 338/ 464] blk.30.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 339/ 464] blk.30.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 340/ 464] blk.30.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 341/ 464] blk.30.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 342/ 464] blk.30.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 343/ 464] blk.31.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 344/ 464] blk.31.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 345/ 464] blk.31.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 346/ 464] blk.31.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 347/ 464] blk.31.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 348/ 464] blk.31.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 349/ 464] blk.31.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 350/ 464] blk.31.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 351/ 464] blk.31.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 352/ 464] blk.31.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 353/ 464] blk.31.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 354/ 464] blk.32.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 355/ 464] blk.32.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 356/ 464] blk.32.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 357/ 464] blk.32.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 358/ 464] blk.32.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 359/ 464] blk.32.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 360/ 464] blk.32.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 361/ 464] blk.32.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 362/ 464] blk.32.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 363/ 464] blk.32.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 364/ 464] blk.32.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 365/ 464] blk.33.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 366/ 464] blk.33.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 367/ 464] blk.33.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 368/ 464] blk.33.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 369/ 464] blk.33.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 370/ 464] blk.33.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 371/ 464] blk.33.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 372/ 464] blk.33.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 373/ 464] blk.33.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 374/ 464] blk.33.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 375/ 464] blk.33.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 376/ 464] blk.34.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 377/ 464] blk.34.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 378/ 464] blk.34.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 379/ 464] blk.34.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 380/ 464] blk.34.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 381/ 464] blk.34.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 382/ 464] blk.34.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 383/ 464] blk.34.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 384/ 464] blk.34.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 385/ 464] blk.34.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 386/ 464] blk.34.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 387/ 464] blk.35.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 388/ 464] blk.35.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 389/ 464] blk.35.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 390/ 464] blk.35.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 391/ 464] blk.35.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 392/ 464] blk.35.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 393/ 464] blk.35.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 394/ 464] blk.35.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 395/ 464] blk.35.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 396/ 464] blk.35.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 397/ 464] blk.35.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 398/ 464] blk.36.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 399/ 464] blk.36.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 400/ 464] blk.36.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 401/ 464] blk.36.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 402/ 464] blk.36.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 403/ 464] blk.36.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 404/ 464] blk.36.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 405/ 464] blk.36.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 406/ 464] blk.36.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 407/ 464] blk.36.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 408/ 464] blk.36.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 409/ 464] blk.37.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 410/ 464] blk.37.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 411/ 464] blk.37.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 412/ 464] blk.37.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 413/ 464] blk.37.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 414/ 464] blk.37.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 415/ 464] blk.37.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 416/ 464] blk.37.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 417/ 464] blk.37.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 418/ 464] blk.37.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 419/ 464] blk.37.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 420/ 464] blk.38.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 421/ 464] blk.38.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 422/ 464] blk.38.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 423/ 464] blk.38.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 424/ 464] blk.38.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 425/ 464] blk.38.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 426/ 464] blk.38.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 427/ 464] blk.38.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 428/ 464] blk.38.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 429/ 464] blk.38.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 430/ 464] blk.38.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 431/ 464] blk.39.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 432/ 464] blk.39.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 433/ 464] blk.39.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 434/ 464] blk.39.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 435/ 464] blk.39.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 436/ 464] blk.39.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 437/ 464] blk.39.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 438/ 464] blk.39.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 439/ 464] blk.39.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 440/ 464] blk.39.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 441/ 464] blk.39.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 442/ 464] blk.40.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 443/ 464] blk.40.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 444/ 464] blk.40.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 445/ 464] blk.40.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 446/ 464] blk.40.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 447/ 464] blk.40.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 448/ 464] blk.40.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 449/ 464] blk.40.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 450/ 464] blk.40.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 451/ 464] blk.40.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 452/ 464] blk.40.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 453/ 464] blk.41.attn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 454/ 464] blk.41.ffn_down.weight - [14336, 3584, 1, 1], type = bf16, converting to q6_K .. size = 98.00 MiB -> 40.20 MiB\n", + "[ 455/ 464] blk.41.ffn_gate.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 456/ 464] blk.41.ffn_up.weight - [ 3584, 14336, 1, 1], type = bf16, converting to q4_K .. size = 98.00 MiB -> 27.56 MiB\n", + "[ 457/ 464] blk.41.post_attention_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 458/ 464] blk.41.post_ffw_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 459/ 464] blk.41.ffn_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "[ 460/ 464] blk.41.attn_k.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q4_K .. size = 14.00 MiB -> 3.94 MiB\n", + "[ 461/ 464] blk.41.attn_output.weight - [ 4096, 3584, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 462/ 464] blk.41.attn_q.weight - [ 3584, 4096, 1, 1], type = bf16, converting to q4_K .. size = 28.00 MiB -> 7.88 MiB\n", + "[ 463/ 464] blk.41.attn_v.weight - [ 3584, 2048, 1, 1], type = bf16, converting to q6_K .. size = 14.00 MiB -> 5.74 MiB\n", + "[ 464/ 464] output_norm.weight - [ 3584, 1, 1, 1], type = f32, size = 0.014 MB\n", + "llama_model_quantize_internal: model size = 17628.31 MB\n", + "llama_model_quantize_internal: quant size = 5488.40 MB\n", + "\n", + "main: quantize time = 69388.16 ms\n", + "main: total time = 69388.16 ms\n", + "Unsloth: Conversion completed! Output location: ./stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored/unsloth.Q4_K_M.gguf\n", + "Unsloth: Uploading GGUF to Huggingface Hub...\n", + "Saved GGUF to https://huggingface.co/stephenlzc/Gemma-2-9B-Chinese-Chat-Uncensored\n", + "Unsloth: Uploading GGUF to Huggingface Hub...\n" + ] + }, + { + "data": { + "application/vnd.jupyter.widget-view+json": { + "model_id": "6121be8e99e64129991fc1ee81d22847", + "version_major": 2, + "version_minor": 0 + }, + "text/plain": [ + "unsloth.Q4_K_M.gguf: 0%| | 0.00/5.76G [00:00