{"cells":[{"cell_type":"code","execution_count":1,"metadata":{"executionInfo":{"elapsed":476,"status":"ok","timestamp":1720679526275,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"uWKRSV6eZsCn"},"outputs":[],"source":["%load_ext autoreload\n","%autoreload 2"]},{"cell_type":"code","execution_count":3,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{"byteLimit":2048000,"rowLimit":10000},"inputWidgets":{},"nuid":"eb33b19f-1206-41ee-84e2-e6258a12eef7","showTitle":false,"title":""},"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":2534,"status":"ok","timestamp":1720679529344,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"xwFh14uiZBrI","outputId":"d767799c-34c2-46a5-f052-378146a55321"},"outputs":[],"source":["from pathlib import Path\n","\n","try:\n"," from google.colab import drive\n","\n"," drive.mount(\"/content/drive\")\n"," workding_dir = \"/content/drive/MyDrive/logical-reasoning/\"\n","except ModuleNotFoundError:\n"," workding_dir = str(Path.cwd().parent)"]},{"cell_type":"code","execution_count":4,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{"byteLimit":2048000,"rowLimit":10000},"inputWidgets":{},"nuid":"6d394937-6c99-4a7c-9d32-7600a280032f","showTitle":false,"title":""},"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":5,"status":"ok","timestamp":1720679529345,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"G5pNu3zgZBrL","outputId":"160a554f-fb08-4aa0-bc00-0422fb7c1fac"},"outputs":[{"name":"stdout","output_type":"stream","text":["workding dir: /home/inflaton/code/logical-reasoning\n"]}],"source":["import os\n","import sys\n","from pathlib import Path\n","\n","os.chdir(workding_dir)\n","sys.path.append(workding_dir)\n","print(\"workding dir:\", workding_dir)"]},{"cell_type":"code","execution_count":5,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{"byteLimit":2048000,"rowLimit":10000},"inputWidgets":{},"nuid":"9f67ec60-2f24-411c-84eb-0dd664b44775","showTitle":false,"title":""},"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":3,"status":"ok","timestamp":1720679529345,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"hPCC-6m7ZBrM","outputId":"c7aa2c96-5e99-440a-c148-201d79465ff9"},"outputs":[{"name":"stdout","output_type":"stream","text":["loading env vars from: /home/inflaton/code/logical-reasoning/.env\n"]},{"data":{"text/plain":["True"]},"execution_count":5,"metadata":{},"output_type":"execute_result"}],"source":["from dotenv import find_dotenv, load_dotenv\n","\n","found_dotenv = find_dotenv(\".env\")\n","\n","if len(found_dotenv) == 0:\n"," found_dotenv = find_dotenv(\".env.example\")\n","print(f\"loading env vars from: {found_dotenv}\")\n","load_dotenv(found_dotenv, override=True)"]},{"cell_type":"code","execution_count":6,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{"byteLimit":2048000,"rowLimit":10000},"inputWidgets":{},"nuid":"f1597656-8042-4878-9d3b-9ebfb8dd86dc","showTitle":false,"title":""},"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":3,"status":"ok","timestamp":1720679529345,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"1M3IraVtZBrM","outputId":"29ab35f6-2970-4ade-d85d-3174acf8cda0"},"outputs":[{"name":"stdout","output_type":"stream","text":["internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter False datasets/mgtv results/mgtv-results_nv4090.csv\n"]}],"source":["import os\n","\n","model_name = os.getenv(\"MODEL_NAME\")\n","adapter_name_or_path = os.getenv(\"ADAPTER_NAME_OR_PATH\")\n","load_in_4bit = os.getenv(\"LOAD_IN_4BIT\") == \"true\"\n","data_path = os.getenv(\"LOGICAL_REASONING_DATA_PATH\")\n","results_path = os.getenv(\"LOGICAL_REASONING_RESULTS_PATH\")\n","use_english_datasets = os.getenv(\"USE_ENGLISH_DATASETS\") == \"true\"\n","\n","print(model_name, adapter_name_or_path, load_in_4bit, data_path, results_path)"]},{"cell_type":"code","execution_count":7,"metadata":{"application/vnd.databricks.v1+cell":{"cellMetadata":{"byteLimit":2048000,"rowLimit":10000},"inputWidgets":{},"nuid":"b2a43943-9324-4839-9a47-cfa72de2244b","showTitle":false,"title":""},"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":564,"status":"ok","timestamp":1720679529907,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"UgMvt6dIZBrM","outputId":"ce37581c-fd26-46c2-ad87-d933d99f68f7"},"outputs":[{"name":"stdout","output_type":"stream","text":["Python 3.11.9\n","\u001b[33mWARNING: Package(s) not found: torchaudio, torchvision\u001b[0m\u001b[33m\n","\u001b[0mName: transformers\n","Version: 4.41.2\n","Summary: State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow\n","Home-page: https://github.com/huggingface/transformers\n","Author: The Hugging Face team (past and future) with the help of all our contributors (https://github.com/huggingface/transformers/graphs/contributors)\n","Author-email: transformers@huggingface.co\n","License: Apache 2.0 License\n","Location: /home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages\n","Requires: filelock, huggingface-hub, numpy, packaging, pyyaml, regex, requests, safetensors, tokenizers, tqdm\n","Required-by: llamafactory, peft, trl\n","---\n","Name: torch\n","Version: 2.3.1\n","Summary: Tensors and Dynamic neural networks in Python with strong GPU acceleration\n","Home-page: https://pytorch.org/\n","Author: PyTorch Team\n","Author-email: packages@pytorch.org\n","License: BSD-3\n","Location: /home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages\n","Requires: filelock, fsspec, jinja2, networkx, nvidia-cublas-cu12, nvidia-cuda-cupti-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-runtime-cu12, nvidia-cudnn-cu12, nvidia-cufft-cu12, nvidia-curand-cu12, nvidia-cusolver-cu12, nvidia-cusparse-cu12, nvidia-nccl-cu12, nvidia-nvtx-cu12, sympy, triton, typing-extensions\n","Required-by: accelerate, peft, trl\n","CPU times: user 16.9 ms, sys: 0 ns, total: 16.9 ms\n","Wall time: 1.86 s\n"]}],"source":["%%time\n","!python --version\n","!pip show transformers torch torchvision torchaudio"]},{"cell_type":"code","execution_count":10,"metadata":{"colab":{"base_uri":"https://localhost:8080/"},"executionInfo":{"elapsed":1685,"status":"ok","timestamp":1720679531591,"user":{"displayName":"HUANG DONGHAO _","userId":"00977795705617022768"},"user_tz":-480},"id":"ZuS_FsLyZBrN","outputId":"2cba0105-c505-4395-afbd-2f2fee6581d0"},"outputs":[{"name":"stdout","output_type":"stream","text":["CUDA is available, we have found 1 GPU(s)\n","NVIDIA GeForce RTX 4090\n","CUDA version: 12.1\n"]}],"source":["from llm_toolkit.llm_utils import *\n","from llm_toolkit.logical_reasoning_utils import *\n","\n","device = check_gpu()"]},{"cell_type":"code","execution_count":15,"metadata":{},"outputs":[{"name":"stdout","output_type":"stream","text":["Current Directory:\n","/home/inflaton/code/logical-reasoning\n","Mon Jul 22 21:10:43 2024 \n","+---------------------------------------------------------------------------------------+\n","| NVIDIA-SMI 535.54.04 Driver Version: 536.25 CUDA Version: 12.2 |\n","|-----------------------------------------+----------------------+----------------------+\n","| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |\n","| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |\n","| | | MIG M. |\n","|=========================================+======================+======================|\n","| 0 NVIDIA GeForce RTX 4090 On | 00000000:01:00.0 Off | Off |\n","| 0% 46C P8 13W / 450W | 1450MiB / 24564MiB | 1% Default |\n","| | | N/A |\n","+-----------------------------------------+----------------------+----------------------+\n"," \n","+---------------------------------------------------------------------------------------+\n","| Processes: |\n","| GPU GI CI PID Type Process name GPU Memory |\n","| ID ID Usage |\n","|=======================================================================================|\n","| No running processes found |\n","+---------------------------------------------------------------------------------------+\n","Linux Gen-AI 5.15.133.1-microsoft-standard-WSL2 #1 SMP Thu Oct 5 21:02:42 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux\n","PRETTY_NAME=\"Ubuntu 22.04.2 LTS\"\n","NAME=\"Ubuntu\"\n","VERSION_ID=\"22.04\"\n","VERSION=\"22.04.2 LTS (Jammy Jellyfish)\"\n","VERSION_CODENAME=jammy\n","ID=ubuntu\n","ID_LIKE=debian\n","HOME_URL=\"https://www.ubuntu.com/\"\n","SUPPORT_URL=\"https://help.ubuntu.com/\"\n","BUG_REPORT_URL=\"https://bugs.launchpad.net/ubuntu/\"\n","PRIVACY_POLICY_URL=\"https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\"\n","UBUNTU_CODENAME=jammy\n","Architecture: x86_64\n"," CPU op-mode(s): 32-bit, 64-bit\n"," Address sizes: 39 bits physical, 48 bits virtual\n"," Byte Order: Little Endian\n","CPU(s): 32\n"," On-line CPU(s) list: 0-31\n","Vendor ID: GenuineIntel\n"," Model name: 13th Gen Intel(R) Core(TM) i9-13900KF\n"," CPU family: 6\n"," Model: 183\n"," Thread(s) per core: 2\n"," Core(s) per socket: 16\n"," Socket(s): 1\n"," Stepping: 1\n"," BogoMIPS: 5990.40\n"," Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mc\n"," a cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscal\n"," l nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopo\n"," logy tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx s\n"," sse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_dea\n"," dline_timer aes xsave avx f16c rdrand hypervisor lahf_l\n"," m abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced \n"," tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi\n"," 1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushop\n"," t clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni u\n"," mip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64\n"," b fsrm md_clear serialize flush_l1d arch_capabilities\n","Virtualization features: \n"," Virtualization: VT-x\n"," Hypervisor vendor: Microsoft\n"," Virtualization type: full\n","Caches (sum of all): \n"," L1d: 768 KiB (16 instances)\n"," L1i: 512 KiB (16 instances)\n"," L2: 32 MiB (16 instances)\n"," L3: 36 MiB (1 instance)\n","Vulnerabilities: \n"," Gather data sampling: Not affected\n"," Itlb multihit: Not affected\n"," L1tf: Not affected\n"," Mds: Not affected\n"," Meltdown: Not affected\n"," Mmio stale data: Not affected\n"," Retbleed: Mitigation; Enhanced IBRS\n"," Spec rstack overflow: Not affected\n"," Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl\n"," and seccomp\n"," Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer\n"," sanitization\n"," Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB fillin\n"," g, PBRSB-eIBRS SW sequence\n"," Srbds: Not affected\n"," Tsx async abort: Not affected\n","MemTotal: 49330020 kB\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter False datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","tokenizer_config.json: 100%|███████████████| 2.51k/2.51k [00:00<00:00, 38.0MB/s]\n","tokenization_internlm2_fast.py: 100%|██████| 7.80k/7.80k [00:00<00:00, 77.2MB/s]\n","tokenization_internlm2.py: 100%|███████████| 8.81k/8.81k [00:00<00:00, 89.2MB/s]\n","tokenizer.model: 100%|█████████████████████| 1.48M/1.48M [00:01<00:00, 1.02MB/s]\n","special_tokens_map.json: 100%|█████████████████| 713/713 [00:00<00:00, 12.8MB/s]\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:10:55,053 >> loading file ./tokenizer.model from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/./tokenizer.model\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:10:55,053 >> loading file added_tokens.json from cache at None\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:10:55,053 >> loading file special_tokens_map.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/special_tokens_map.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:10:55,053 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/tokenizer_config.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:10:55,053 >> loading file tokenizer.json from cache at None\n","07/22/2024 21:10:55 - INFO - llamafactory.data.template - Add <|im_end|> to stop words.\n","/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n"," warnings.warn(\n","config.json: 100%|█████████████████████████████| 895/895 [00:00<00:00, 15.1MB/s]\n","[INFO|configuration_utils.py:733] 2024-07-22 21:10:55,959 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","configuration_internlm2.py: 100%|███████████| 8.84k/8.84k [00:00<00:00, 104MB/s]\n","[INFO|configuration_utils.py:733] 2024-07-22 21:10:56,833 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","[INFO|configuration_utils.py:796] 2024-07-22 21:10:56,833 >> Model config InternLM2Config {\n"," \"_name_or_path\": \"internlm/internlm2_5-7b-chat-1m\",\n"," \"architectures\": [\n"," \"InternLM2ForCausalLM\"\n"," ],\n"," \"attn_implementation\": \"eager\",\n"," \"auto_map\": {\n"," \"AutoConfig\": \"internlm/internlm2_5-7b-chat-1m--configuration_internlm2.InternLM2Config\",\n"," \"AutoModel\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\",\n"," \"AutoModelForCausalLM\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\"\n"," },\n"," \"bias\": false,\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": 2,\n"," \"hidden_act\": \"silu\",\n"," \"hidden_size\": 4096,\n"," \"initializer_range\": 0.02,\n"," \"intermediate_size\": 14336,\n"," \"max_position_embeddings\": 262144,\n"," \"model_type\": \"internlm2\",\n"," \"num_attention_heads\": 32,\n"," \"num_hidden_layers\": 32,\n"," \"num_key_value_heads\": 8,\n"," \"pad_token_id\": 2,\n"," \"pretraining_tp\": 1,\n"," \"rms_norm_eps\": 1e-05,\n"," \"rope_scaling\": {\n"," \"factor\": 2.5,\n"," \"type\": \"dynamic\"\n"," },\n"," \"rope_theta\": 50000000,\n"," \"tie_word_embeddings\": false,\n"," \"torch_dtype\": \"bfloat16\",\n"," \"transformers_version\": \"4.41.2\",\n"," \"use_cache\": true,\n"," \"vocab_size\": 92544\n","}\n","\n","07/22/2024 21:10:56 - INFO - llamafactory.model.patcher - Using KV cache for faster generation.\n","modeling_internlm2.py: 100%|████████████████| 80.7k/80.7k [00:00<00:00, 263kB/s]\n","model.safetensors.index.json: 100%|████████| 18.2k/18.2k [00:00<00:00, 3.49MB/s]\n","[INFO|modeling_utils.py:3474] 2024-07-22 21:10:59,134 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/model.safetensors.index.json\n","Downloading shards: 0%| | 0/8 [00:00, ?it/s]\n","model-00001-of-00008.safetensors: 0%| | 0.00/1.95G [00:00, ?B/s]\u001b[A\n","model-00001-of-00008.safetensors: 1%| | 10.5M/1.95G [00:01<05:25, 5.95MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 1%| | 21.0M/1.95G [00:02<03:46, 8.51MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 2%| | 31.5M/1.95G [00:03<03:22, 9.45MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 2%| | 41.9M/1.95G [00:04<03:39, 8.68MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 3%| | 52.4M/1.95G [00:05<03:22, 9.39MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 3%|▏ | 62.9M/1.95G [00:06<03:06, 10.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 4%|▏ | 73.4M/1.95G [00:07<02:56, 10.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 4%|▏ | 83.9M/1.95G [00:08<02:52, 10.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 5%|▏ | 94.4M/1.95G [00:09<02:47, 11.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 5%|▎ | 105M/1.95G [00:10<02:44, 11.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 6%|▎ | 115M/1.95G [00:11<02:38, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 6%|▎ | 126M/1.95G [00:12<02:36, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 7%|▎ | 136M/1.95G [00:13<02:39, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 8%|▍ | 147M/1.95G [00:13<02:37, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 8%|▍ | 157M/1.95G [00:14<02:34, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 9%|▍ | 168M/1.95G [00:15<02:35, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 9%|▍ | 178M/1.95G [00:16<02:30, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 10%|▍ | 189M/1.95G [00:17<02:28, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 10%|▌ | 199M/1.95G [00:18<02:54, 10.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 11%|▌ | 210M/1.95G [00:19<02:46, 10.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 11%|▌ | 220M/1.95G [00:20<02:38, 10.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 12%|▌ | 231M/1.95G [00:21<02:33, 11.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 12%|▌ | 241M/1.95G [00:22<02:33, 11.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 13%|▋ | 252M/1.95G [00:23<02:27, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 13%|▋ | 262M/1.95G [00:24<02:24, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 14%|▋ | 273M/1.95G [00:25<02:22, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 15%|▋ | 283M/1.95G [00:26<02:24, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 15%|▊ | 294M/1.95G [00:26<02:21, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 16%|▊ | 304M/1.95G [00:27<02:20, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 16%|▊ | 315M/1.95G [00:28<02:21, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 17%|▊ | 325M/1.95G [00:29<02:19, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 17%|▊ | 336M/1.95G [00:30<02:17, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 18%|▉ | 346M/1.95G [00:31<02:17, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 18%|▉ | 357M/1.95G [00:32<02:14, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 19%|▉ | 367M/1.95G [00:33<02:12, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 19%|▉ | 377M/1.95G [00:34<02:14, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 20%|▉ | 388M/1.95G [00:34<02:12, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 20%|█ | 398M/1.95G [00:35<02:11, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 21%|█ | 409M/1.95G [00:36<02:13, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 22%|█ | 419M/1.95G [00:37<02:11, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 22%|█ | 430M/1.95G [00:39<02:29, 10.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 23%|█▏ | 440M/1.95G [00:39<02:21, 10.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 23%|█▏ | 451M/1.95G [00:40<02:16, 11.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 24%|█▏ | 461M/1.95G [00:41<02:12, 11.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 24%|█▏ | 472M/1.95G [00:42<02:09, 11.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 25%|█▏ | 482M/1.95G [00:43<02:07, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 25%|█▎ | 493M/1.95G [00:44<02:07, 11.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 26%|█▎ | 503M/1.95G [00:45<02:04, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 26%|█▎ | 514M/1.95G [00:46<02:03, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 27%|█▎ | 524M/1.95G [00:47<02:01, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 27%|█▎ | 535M/1.95G [00:47<02:00, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 28%|█▍ | 545M/1.95G [00:48<01:58, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 29%|█▍ | 556M/1.95G [00:49<01:56, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 29%|█▍ | 566M/1.95G [00:50<01:56, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 30%|█▍ | 577M/1.95G [00:51<01:54, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 30%|█▌ | 587M/1.95G [00:52<01:54, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 31%|█▌ | 598M/1.95G [00:53<01:53, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 31%|█▌ | 608M/1.95G [00:54<02:12, 10.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 32%|█▌ | 619M/1.95G [00:55<02:05, 10.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 32%|█▌ | 629M/1.95G [00:56<01:59, 11.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 33%|█▋ | 640M/1.95G [00:57<01:56, 11.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 33%|█▋ | 650M/1.95G [00:58<01:53, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 34%|█▋ | 661M/1.95G [00:58<01:50, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 34%|█▋ | 671M/1.95G [00:59<01:49, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 35%|█▋ | 682M/1.95G [01:00<01:47, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 36%|█▊ | 692M/1.95G [01:03<03:12, 6.53MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 36%|█▊ | 703M/1.95G [01:04<02:45, 7.53MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 37%|█▊ | 713M/1.95G [01:05<02:27, 8.40MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 37%|█▊ | 724M/1.95G [01:07<02:29, 8.22MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 38%|█▉ | 734M/1.95G [01:08<02:14, 9.05MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 38%|█▉ | 744M/1.95G [01:08<02:03, 9.79MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 39%|█▉ | 755M/1.95G [01:09<01:55, 10.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 39%|█▉ | 765M/1.95G [01:10<01:49, 10.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 40%|█▉ | 776M/1.95G [01:11<01:46, 11.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 40%|██ | 786M/1.95G [01:12<01:42, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 41%|██ | 797M/1.95G [01:13<01:39, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 41%|██ | 807M/1.95G [01:14<01:37, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 42%|██ | 818M/1.95G [01:15<01:35, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 42%|██ | 828M/1.95G [01:15<01:33, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 43%|██▏ | 839M/1.95G [01:16<01:32, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 44%|██▏ | 849M/1.95G [01:17<01:31, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 44%|██▏ | 860M/1.95G [01:18<01:30, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 45%|██▏ | 870M/1.95G [01:19<01:29, 12.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 45%|██▎ | 881M/1.95G [01:20<01:28, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 46%|██▎ | 891M/1.95G [01:21<01:27, 12.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 46%|██▎ | 902M/1.95G [01:21<01:27, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 47%|██▎ | 912M/1.95G [01:22<01:27, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 47%|██▎ | 923M/1.95G [01:23<01:25, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 48%|██▍ | 933M/1.95G [01:24<01:24, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 48%|██▍ | 944M/1.95G [01:25<01:24, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 49%|██▍ | 954M/1.95G [01:26<01:23, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 49%|██▍ | 965M/1.95G [01:27<01:35, 10.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 50%|██▌ | 975M/1.95G [01:28<01:31, 10.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 51%|██▌ | 986M/1.95G [01:29<01:26, 11.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 51%|██▌ | 996M/1.95G [01:30<01:24, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 52%|██ | 1.01G/1.95G [01:31<01:21, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 52%|██ | 1.02G/1.95G [01:32<01:19, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 53%|██ | 1.03G/1.95G [01:32<01:17, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 53%|██▏ | 1.04G/1.95G [01:33<01:16, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 54%|██▏ | 1.05G/1.95G [01:34<01:15, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 54%|██▏ | 1.06G/1.95G [01:35<01:14, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 55%|██▏ | 1.07G/1.95G [01:36<01:13, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 55%|██▏ | 1.08G/1.95G [01:37<01:12, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 56%|██▏ | 1.09G/1.95G [01:38<01:12, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 56%|██▎ | 1.10G/1.95G [01:39<01:10, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 57%|██▎ | 1.11G/1.95G [01:39<01:10, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 58%|██▎ | 1.12G/1.95G [01:40<01:08, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 58%|██▎ | 1.13G/1.95G [01:41<01:07, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 59%|██▎ | 1.14G/1.95G [01:42<01:07, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 59%|██▎ | 1.15G/1.95G [01:43<01:07, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 60%|██▍ | 1.16G/1.95G [01:44<01:05, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 60%|██▍ | 1.17G/1.95G [01:45<01:04, 12.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 61%|██▍ | 1.18G/1.95G [01:46<01:04, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 61%|██▍ | 1.20G/1.95G [01:46<01:03, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 62%|██▍ | 1.21G/1.95G [01:47<01:02, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 62%|██▍ | 1.22G/1.95G [01:49<01:12, 10.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 63%|██▌ | 1.23G/1.95G [01:50<01:08, 10.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 63%|██▌ | 1.24G/1.95G [01:50<01:04, 11.0MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 64%|██▌ | 1.25G/1.95G [01:51<01:01, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 65%|██▌ | 1.26G/1.95G [01:52<01:01, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 65%|██▌ | 1.27G/1.95G [01:53<00:58, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 66%|██▋ | 1.28G/1.95G [01:54<00:57, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 66%|██▋ | 1.29G/1.95G [01:55<00:57, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 67%|██▋ | 1.30G/1.95G [01:56<00:55, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 67%|██▋ | 1.31G/1.95G [01:57<00:54, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 68%|██▋ | 1.32G/1.95G [01:58<00:52, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 68%|██▋ | 1.33G/1.95G [01:58<00:52, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 69%|██▊ | 1.34G/1.95G [01:59<00:51, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 69%|██▊ | 1.35G/1.95G [02:00<00:50, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 70%|██▊ | 1.36G/1.95G [02:01<00:50, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 70%|██▊ | 1.37G/1.95G [02:03<01:11, 8.05MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 71%|██▊ | 1.38G/1.95G [02:04<01:03, 8.92MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 72%|██▊ | 1.39G/1.95G [02:05<00:57, 9.58MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 72%|██▉ | 1.41G/1.95G [02:06<00:53, 10.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 73%|██▉ | 1.42G/1.95G [02:07<00:50, 10.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 73%|██▉ | 1.43G/1.95G [02:08<00:48, 10.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 74%|██▉ | 1.44G/1.95G [02:09<00:46, 11.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 74%|██▉ | 1.45G/1.95G [02:10<00:44, 11.2MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 75%|██▉ | 1.46G/1.95G [02:11<00:42, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 75%|███ | 1.47G/1.95G [02:11<00:41, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 76%|███ | 1.48G/1.95G [02:12<00:40, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 76%|███ | 1.49G/1.95G [02:14<00:46, 9.84MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 77%|███ | 1.50G/1.95G [02:15<00:43, 10.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 77%|███ | 1.51G/1.95G [02:16<00:40, 10.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 78%|███ | 1.52G/1.95G [02:16<00:38, 11.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 79%|███▏| 1.53G/1.95G [02:17<00:36, 11.4MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 79%|███▏| 1.54G/1.95G [02:18<00:35, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 80%|███▏| 1.55G/1.95G [02:19<00:34, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 80%|███▏| 1.56G/1.95G [02:20<00:34, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 81%|███▏| 1.57G/1.95G [02:21<00:32, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 81%|███▏| 1.58G/1.95G [02:22<00:31, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 82%|███▎| 1.59G/1.95G [02:23<00:30, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 82%|███▎| 1.60G/1.95G [02:24<00:29, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 83%|███▎| 1.61G/1.95G [02:24<00:28, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 83%|███▎| 1.63G/1.95G [02:25<00:27, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 84%|███▎| 1.64G/1.95G [02:26<00:26, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 84%|███▍| 1.65G/1.95G [02:27<00:26, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 85%|███▍| 1.66G/1.95G [02:28<00:25, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 86%|███▍| 1.67G/1.95G [02:29<00:23, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 86%|███▍| 1.68G/1.95G [02:30<00:22, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 87%|███▍| 1.69G/1.95G [02:31<00:22, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 87%|███▍| 1.70G/1.95G [02:32<00:21, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 88%|███▌| 1.71G/1.95G [02:33<00:20, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 88%|███▌| 1.72G/1.95G [02:33<00:19, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 89%|███▌| 1.73G/1.95G [02:34<00:18, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 89%|███▌| 1.74G/1.95G [02:35<00:17, 11.8MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 90%|███▌| 1.75G/1.95G [02:36<00:17, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 90%|███▌| 1.76G/1.95G [02:37<00:16, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 91%|███▋| 1.77G/1.95G [02:38<00:15, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 91%|███▋| 1.78G/1.95G [02:39<00:14, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 92%|███▋| 1.79G/1.95G [02:40<00:13, 11.6MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 93%|███▋| 1.80G/1.95G [02:41<00:12, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 93%|███▋| 1.81G/1.95G [02:42<00:13, 10.1MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 94%|███▋| 1.82G/1.95G [02:43<00:11, 10.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 94%|███▊| 1.84G/1.95G [02:44<00:10, 10.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 95%|███▊| 1.85G/1.95G [02:45<00:09, 10.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 95%|███▊| 1.86G/1.95G [02:46<00:08, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 96%|███▊| 1.87G/1.95G [02:46<00:07, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 96%|███▊| 1.88G/1.95G [02:47<00:06, 11.3MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 97%|███▊| 1.89G/1.95G [02:48<00:05, 11.5MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 97%|███▉| 1.90G/1.95G [02:49<00:04, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 98%|███▉| 1.91G/1.95G [02:50<00:03, 11.9MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 98%|███▉| 1.92G/1.95G [02:51<00:02, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 99%|███▉| 1.93G/1.95G [02:52<00:01, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 100%|███▉| 1.94G/1.95G [02:53<00:00, 11.7MB/s]\u001b[A\n","model-00001-of-00008.safetensors: 100%|████| 1.95G/1.95G [02:54<00:00, 11.2MB/s]\u001b[A\n","Downloading shards: 12%|███ | 1/8 [02:54<20:21, 174.55s/it]\n","model-00002-of-00008.safetensors: 0%| | 0.00/1.95G [00:00, ?B/s]\u001b[A\n","model-00002-of-00008.safetensors: 1%| | 10.5M/1.95G [00:00<02:42, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 1%| | 21.0M/1.95G [00:01<02:43, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 2%| | 31.5M/1.95G [00:02<02:39, 12.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 2%| | 41.9M/1.95G [00:03<02:39, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 3%| | 52.4M/1.95G [00:04<02:45, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 3%|▏ | 62.9M/1.95G [00:05<02:42, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 4%|▏ | 73.4M/1.95G [00:06<02:39, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 4%|▏ | 83.9M/1.95G [00:07<02:38, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 5%|▏ | 94.4M/1.95G [00:08<02:37, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 5%|▎ | 105M/1.95G [00:09<03:07, 9.83MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 6%|▎ | 115M/1.95G [00:10<02:57, 10.3MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 6%|▎ | 126M/1.95G [00:11<03:17, 9.21MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 7%|▎ | 136M/1.95G [00:12<03:02, 9.92MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 8%|▍ | 147M/1.95G [00:13<02:51, 10.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 8%|▍ | 157M/1.95G [00:14<02:45, 10.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 9%|▍ | 168M/1.95G [00:15<02:39, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 9%|▍ | 178M/1.95G [00:16<02:35, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 10%|▍ | 189M/1.95G [00:17<02:35, 11.3MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 10%|▌ | 199M/1.95G [00:17<02:30, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 11%|▌ | 210M/1.95G [00:18<02:29, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 11%|▌ | 220M/1.95G [00:19<02:28, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 12%|▌ | 231M/1.95G [00:20<02:25, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 12%|▌ | 241M/1.95G [00:21<02:24, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 13%|▋ | 252M/1.95G [00:22<02:45, 10.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 13%|▋ | 262M/1.95G [00:23<02:36, 10.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 14%|▋ | 273M/1.95G [00:24<02:34, 10.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 15%|▋ | 283M/1.95G [00:25<02:30, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 15%|▊ | 294M/1.95G [00:26<02:25, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 16%|▊ | 304M/1.95G [00:27<02:23, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 16%|▊ | 315M/1.95G [00:28<02:20, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 17%|▊ | 325M/1.95G [00:29<02:17, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 17%|▊ | 336M/1.95G [00:29<02:16, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 18%|▉ | 346M/1.95G [00:30<02:16, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 18%|▉ | 357M/1.95G [00:31<02:15, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 19%|▉ | 367M/1.95G [00:32<02:14, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 19%|▉ | 377M/1.95G [00:33<02:15, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 20%|▉ | 388M/1.95G [00:34<02:12, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 20%|█ | 398M/1.95G [00:35<02:10, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 21%|█ | 409M/1.95G [00:36<02:09, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 22%|█ | 419M/1.95G [00:37<02:09, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 22%|█ | 430M/1.95G [00:37<02:06, 12.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 23%|█▏ | 440M/1.95G [00:38<02:06, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 23%|█▏ | 451M/1.95G [00:39<02:07, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 24%|█▏ | 461M/1.95G [00:40<02:05, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 24%|█▏ | 472M/1.95G [00:42<02:25, 10.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 25%|█▏ | 482M/1.95G [00:42<02:17, 10.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 25%|█▎ | 493M/1.95G [00:43<02:12, 11.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 26%|█▎ | 503M/1.95G [00:44<02:08, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 26%|█▎ | 514M/1.95G [00:45<02:05, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 27%|█▎ | 524M/1.95G [00:46<02:06, 11.3MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 27%|█▎ | 535M/1.95G [00:47<02:02, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 28%|█▍ | 545M/1.95G [00:48<01:59, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 29%|█▍ | 556M/1.95G [00:49<01:57, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 29%|█▍ | 566M/1.95G [00:49<01:57, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 30%|█▍ | 577M/1.95G [00:50<01:55, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 30%|█▌ | 587M/1.95G [00:51<01:53, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 31%|█▌ | 598M/1.95G [00:53<02:14, 10.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 31%|█▌ | 608M/1.95G [00:54<02:07, 10.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 32%|█▌ | 619M/1.95G [00:54<02:00, 11.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 32%|█▌ | 629M/1.95G [00:55<01:57, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 33%|█▋ | 640M/1.95G [00:56<01:53, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 33%|█▋ | 650M/1.95G [00:57<01:52, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 34%|█▋ | 661M/1.95G [00:58<01:49, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 34%|█▋ | 671M/1.95G [00:59<01:48, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 35%|█▊ | 682M/1.95G [01:00<01:48, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 36%|█▊ | 692M/1.95G [01:01<01:46, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 36%|█▊ | 703M/1.95G [01:01<01:44, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 37%|█▊ | 713M/1.95G [01:02<01:43, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 37%|█▊ | 724M/1.95G [01:04<01:59, 10.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 38%|█▉ | 734M/1.95G [01:05<01:53, 10.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 38%|█▉ | 744M/1.95G [01:05<01:48, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 39%|█▉ | 755M/1.95G [01:09<03:08, 6.30MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 39%|█▉ | 765M/1.95G [01:10<02:43, 7.21MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 40%|█▉ | 776M/1.95G [01:11<02:23, 8.13MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 40%|██ | 786M/1.95G [01:11<02:08, 9.01MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 41%|██ | 797M/1.95G [01:12<01:58, 9.70MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 41%|██ | 807M/1.95G [01:13<01:54, 9.95MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 42%|██ | 818M/1.95G [01:14<01:46, 10.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 43%|██▏ | 828M/1.95G [01:15<01:42, 11.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 43%|██▏ | 839M/1.95G [01:16<01:38, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 44%|██▏ | 849M/1.95G [01:17<01:37, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 44%|██▏ | 860M/1.95G [01:18<01:34, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 45%|██▏ | 870M/1.95G [01:19<01:32, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 45%|██▎ | 881M/1.95G [01:20<01:30, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 46%|██▎ | 891M/1.95G [01:20<01:28, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 46%|██▎ | 902M/1.95G [01:21<01:27, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 47%|██▎ | 912M/1.95G [01:22<01:28, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 47%|██▎ | 923M/1.95G [01:23<01:26, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 48%|██▍ | 933M/1.95G [01:24<01:25, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 48%|██▍ | 944M/1.95G [01:25<01:24, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 49%|██▍ | 954M/1.95G [01:26<01:23, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 50%|██▍ | 965M/1.95G [01:27<01:35, 10.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 50%|██▌ | 975M/1.95G [01:28<01:30, 10.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 51%|██▌ | 986M/1.95G [01:29<01:27, 11.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 51%|██▌ | 996M/1.95G [01:30<01:24, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 52%|██ | 1.01G/1.95G [01:31<01:24, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 52%|██ | 1.02G/1.95G [01:32<01:21, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 53%|██ | 1.03G/1.95G [01:32<01:19, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 53%|██▏ | 1.04G/1.95G [01:33<01:18, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 54%|██▏ | 1.05G/1.95G [01:34<01:16, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 54%|██▏ | 1.06G/1.95G [01:35<01:15, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 55%|██▏ | 1.07G/1.95G [01:36<01:14, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 55%|██▏ | 1.08G/1.95G [01:37<01:13, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 56%|██▏ | 1.09G/1.95G [01:38<01:12, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 57%|██▎ | 1.10G/1.95G [01:39<01:13, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 57%|██▎ | 1.11G/1.95G [01:40<01:11, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 58%|██▎ | 1.12G/1.95G [01:40<01:10, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 58%|██▎ | 1.13G/1.95G [01:41<01:11, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 59%|██▎ | 1.14G/1.95G [01:42<01:09, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 59%|██▎ | 1.15G/1.95G [01:43<01:08, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 60%|██▍ | 1.16G/1.95G [01:44<01:06, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 60%|██▍ | 1.17G/1.95G [01:45<01:05, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 61%|██▍ | 1.18G/1.95G [01:46<01:06, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 61%|██▍ | 1.20G/1.95G [01:47<01:04, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 62%|██▍ | 1.21G/1.95G [01:48<01:02, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 62%|██▍ | 1.22G/1.95G [01:49<01:02, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 63%|██▌ | 1.23G/1.95G [01:49<01:01, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 64%|██▌ | 1.24G/1.95G [01:50<00:59, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 64%|██▌ | 1.25G/1.95G [01:51<00:59, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 65%|██▌ | 1.26G/1.95G [01:52<00:57, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 65%|██▌ | 1.27G/1.95G [01:53<00:57, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 66%|██▋ | 1.28G/1.95G [01:54<00:56, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 66%|██▋ | 1.29G/1.95G [01:55<00:55, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 67%|██▋ | 1.30G/1.95G [01:56<01:02, 10.3MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 67%|██▋ | 1.31G/1.95G [01:57<00:59, 10.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 68%|██▋ | 1.32G/1.95G [01:58<00:56, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 68%|██▋ | 1.33G/1.95G [01:59<00:54, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 69%|██▊ | 1.34G/1.95G [02:00<00:52, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 70%|██▊ | 1.35G/1.95G [02:00<00:50, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 70%|██▊ | 1.36G/1.95G [02:01<00:49, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 71%|██▊ | 1.37G/1.95G [02:02<00:48, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 71%|██▊ | 1.38G/1.95G [02:03<00:47, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 72%|██▊ | 1.39G/1.95G [02:04<00:46, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 72%|██▉ | 1.41G/1.95G [02:05<00:45, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 73%|██▉ | 1.42G/1.95G [02:06<00:44, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 73%|██▉ | 1.43G/1.95G [02:07<00:43, 12.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 74%|██▉ | 1.44G/1.95G [02:07<00:42, 12.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 74%|██▉ | 1.45G/1.95G [02:08<00:41, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 75%|██▉ | 1.46G/1.95G [02:09<00:40, 12.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 75%|███ | 1.47G/1.95G [02:11<00:58, 8.13MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 76%|███ | 1.48G/1.95G [02:12<00:53, 8.69MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 77%|███ | 1.49G/1.95G [02:13<00:48, 9.49MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 77%|███ | 1.50G/1.95G [02:14<00:43, 10.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 78%|███ | 1.51G/1.95G [02:15<00:41, 10.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 78%|███ | 1.52G/1.95G [02:16<00:38, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 79%|███▏| 1.53G/1.95G [02:17<00:36, 11.3MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 79%|███▏| 1.54G/1.95G [02:18<00:35, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 80%|███▏| 1.55G/1.95G [02:19<00:34, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 80%|███▏| 1.56G/1.95G [02:19<00:32, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 81%|███▏| 1.57G/1.95G [02:21<00:37, 9.87MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 81%|███▎| 1.58G/1.95G [02:22<00:35, 10.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 82%|███▎| 1.59G/1.95G [02:23<00:32, 10.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 82%|███▎| 1.60G/1.95G [02:24<00:30, 11.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 83%|███▎| 1.61G/1.95G [02:24<00:29, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 84%|███▎| 1.63G/1.95G [02:25<00:27, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 84%|███▎| 1.64G/1.95G [02:26<00:27, 11.4MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 85%|███▍| 1.65G/1.95G [02:27<00:25, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 85%|███▍| 1.66G/1.95G [02:28<00:24, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 86%|███▍| 1.67G/1.95G [02:29<00:23, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 86%|███▍| 1.68G/1.95G [02:30<00:22, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 87%|███▍| 1.69G/1.95G [02:31<00:21, 11.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 87%|███▍| 1.70G/1.95G [02:32<00:20, 12.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 88%|███▌| 1.71G/1.95G [02:32<00:20, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 88%|███▌| 1.72G/1.95G [02:34<00:22, 10.0MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 89%|███▌| 1.73G/1.95G [02:35<00:20, 10.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 89%|███▌| 1.74G/1.95G [02:36<00:18, 10.9MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 90%|███▌| 1.75G/1.95G [02:37<00:17, 11.2MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 91%|███▌| 1.76G/1.95G [02:37<00:16, 11.5MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 91%|███▋| 1.77G/1.95G [02:38<00:15, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 92%|███▋| 1.78G/1.95G [02:39<00:14, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 92%|███▋| 1.79G/1.95G [02:40<00:12, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 93%|███▋| 1.80G/1.95G [02:41<00:12, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 93%|███▋| 1.81G/1.95G [02:42<00:11, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 94%|███▋| 1.82G/1.95G [02:43<00:10, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 94%|███▊| 1.84G/1.95G [02:44<00:09, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 95%|███▊| 1.85G/1.95G [02:44<00:08, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 95%|███▊| 1.86G/1.95G [02:45<00:07, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 96%|███▊| 1.87G/1.95G [02:46<00:06, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 96%|███▊| 1.88G/1.95G [02:47<00:05, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 97%|███▉| 1.89G/1.95G [02:48<00:05, 11.6MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 98%|███▉| 1.90G/1.95G [02:49<00:04, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 98%|███▉| 1.91G/1.95G [02:50<00:03, 11.8MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 99%|███▉| 1.92G/1.95G [02:51<00:02, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 99%|███▉| 1.93G/1.95G [02:52<00:01, 11.7MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 100%|███▉| 1.94G/1.95G [02:53<00:00, 10.1MB/s]\u001b[A\n","model-00002-of-00008.safetensors: 100%|████| 1.95G/1.95G [02:54<00:00, 11.2MB/s]\u001b[A\n","Downloading shards: 25%|██████ | 2/8 [05:49<17:27, 174.66s/it]\n","model-00003-of-00008.safetensors: 0%| | 0.00/1.98G [00:00, ?B/s]\u001b[A\n","model-00003-of-00008.safetensors: 1%| | 10.5M/1.98G [00:00<02:42, 12.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 1%| | 21.0M/1.98G [00:01<02:44, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 2%| | 31.5M/1.98G [00:02<02:40, 12.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 2%| | 41.9M/1.98G [00:03<02:42, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 3%| | 52.4M/1.98G [00:04<02:41, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 3%|▏ | 62.9M/1.98G [00:05<02:41, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 4%|▏ | 73.4M/1.98G [00:06<02:40, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 4%|▏ | 83.9M/1.98G [00:07<02:39, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 5%|▏ | 94.4M/1.98G [00:07<02:39, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 5%|▎ | 105M/1.98G [00:09<03:02, 10.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 6%|▎ | 115M/1.98G [00:10<02:54, 10.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 6%|▎ | 126M/1.98G [00:11<02:47, 11.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 7%|▎ | 136M/1.98G [00:11<02:46, 11.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 7%|▎ | 147M/1.98G [00:12<02:42, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 8%|▍ | 157M/1.98G [00:13<02:40, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 8%|▍ | 168M/1.98G [00:14<02:37, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 9%|▍ | 178M/1.98G [00:15<02:35, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 10%|▍ | 189M/1.98G [00:17<03:17, 9.06MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 10%|▌ | 199M/1.98G [00:18<03:15, 9.12MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 11%|▌ | 210M/1.98G [00:19<02:59, 9.87MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 11%|▌ | 220M/1.98G [00:20<02:50, 10.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 12%|▌ | 231M/1.98G [00:21<02:43, 10.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 12%|▌ | 241M/1.98G [00:21<02:37, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 13%|▋ | 252M/1.98G [00:22<02:33, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 13%|▋ | 262M/1.98G [00:23<02:32, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 14%|▋ | 273M/1.98G [00:24<02:29, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 14%|▋ | 283M/1.98G [00:26<02:51, 9.86MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 15%|▋ | 294M/1.98G [00:26<02:41, 10.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 15%|▊ | 304M/1.98G [00:27<02:34, 10.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 16%|▊ | 315M/1.98G [00:28<02:29, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 16%|▊ | 325M/1.98G [00:29<02:24, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 17%|▊ | 336M/1.98G [00:30<02:21, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 17%|▊ | 346M/1.98G [00:31<02:19, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 18%|▉ | 357M/1.98G [00:32<02:19, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 19%|▉ | 367M/1.98G [00:33<02:16, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 19%|▉ | 377M/1.98G [00:33<02:15, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 20%|▉ | 388M/1.98G [00:34<02:13, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 20%|█ | 398M/1.98G [00:35<02:11, 12.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 21%|█ | 409M/1.98G [00:36<02:10, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 21%|█ | 419M/1.98G [00:37<02:33, 10.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 22%|█ | 430M/1.98G [00:38<02:25, 10.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 22%|█ | 440M/1.98G [00:39<02:20, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 23%|█▏ | 451M/1.98G [00:40<02:15, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 23%|█▏ | 461M/1.98G [00:41<02:14, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 24%|█▏ | 472M/1.98G [00:42<02:11, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 24%|█▏ | 482M/1.98G [00:43<02:08, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 25%|█▏ | 493M/1.98G [00:44<02:06, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 25%|█▎ | 503M/1.98G [00:45<02:06, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 26%|█▎ | 514M/1.98G [00:45<02:04, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 26%|█▎ | 524M/1.98G [00:46<02:05, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 27%|█▎ | 535M/1.98G [00:47<02:02, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 28%|█▍ | 545M/1.98G [00:48<02:01, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 28%|█▍ | 556M/1.98G [00:49<02:01, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 29%|█▍ | 566M/1.98G [00:50<02:00, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 29%|█▍ | 577M/1.98G [00:51<01:59, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 30%|█▍ | 587M/1.98G [00:52<02:00, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 30%|█▌ | 598M/1.98G [00:53<01:58, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 31%|█▌ | 608M/1.98G [00:53<01:57, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 31%|█▌ | 619M/1.98G [00:54<01:58, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 32%|█▌ | 629M/1.98G [00:55<01:55, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 32%|█▌ | 640M/1.98G [00:56<01:54, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 33%|█▋ | 650M/1.98G [00:57<01:55, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 33%|█▋ | 661M/1.98G [00:58<01:53, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 34%|█▋ | 671M/1.98G [00:59<02:10, 10.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 34%|█▋ | 682M/1.98G [01:00<02:03, 10.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 35%|█▋ | 692M/1.98G [01:01<02:00, 10.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 35%|█▊ | 703M/1.98G [01:02<01:55, 11.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 36%|█▊ | 713M/1.98G [01:03<01:51, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 37%|█▊ | 724M/1.98G [01:04<01:49, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 37%|█▊ | 734M/1.98G [01:05<01:47, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 38%|█▉ | 744M/1.98G [01:06<01:45, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 38%|█▉ | 755M/1.98G [01:07<01:46, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 39%|█▉ | 765M/1.98G [01:07<01:44, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 39%|█▉ | 776M/1.98G [01:08<01:42, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 40%|█▉ | 786M/1.98G [01:09<01:44, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 40%|██ | 797M/1.98G [01:11<01:57, 10.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 41%|██ | 807M/1.98G [01:14<03:02, 6.41MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 41%|██ | 818M/1.98G [01:14<02:35, 7.48MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 42%|██ | 828M/1.98G [01:15<02:16, 8.45MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 42%|██ | 839M/1.98G [01:16<02:03, 9.25MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 43%|██▏ | 849M/1.98G [01:19<02:49, 6.69MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 43%|██▏ | 860M/1.98G [01:20<02:28, 7.54MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 44%|██▏ | 870M/1.98G [01:21<02:26, 7.58MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 44%|██▏ | 881M/1.98G [01:22<02:09, 8.50MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 45%|██▎ | 891M/1.98G [01:23<01:56, 9.34MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 46%|██▎ | 902M/1.98G [01:24<01:48, 9.97MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 46%|██▎ | 912M/1.98G [01:25<01:40, 10.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 47%|██▎ | 923M/1.98G [01:26<01:35, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 47%|██▎ | 933M/1.98G [01:26<01:32, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 48%|██▍ | 944M/1.98G [01:27<01:30, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 48%|██▍ | 954M/1.98G [01:28<01:27, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 49%|██▍ | 965M/1.98G [01:29<01:26, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 49%|██▍ | 975M/1.98G [01:30<01:25, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 50%|██▍ | 986M/1.98G [01:31<01:23, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 50%|██▌ | 996M/1.98G [01:32<01:22, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 51%|██ | 1.01G/1.98G [01:33<01:22, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 51%|██ | 1.02G/1.98G [01:33<01:21, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 52%|██ | 1.03G/1.98G [01:34<01:19, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 52%|██ | 1.04G/1.98G [01:35<01:18, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 53%|██ | 1.05G/1.98G [01:36<01:18, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 53%|██▏ | 1.06G/1.98G [01:37<01:16, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 54%|██▏ | 1.07G/1.98G [01:38<01:16, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 55%|██▏ | 1.08G/1.98G [01:39<01:14, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 55%|██▏ | 1.09G/1.98G [01:39<01:13, 12.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 56%|██▏ | 1.10G/1.98G [01:40<01:12, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 56%|██▏ | 1.11G/1.98G [01:41<01:11, 12.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 57%|██▎ | 1.12G/1.98G [01:43<01:23, 10.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 57%|██▎ | 1.13G/1.98G [01:44<01:19, 10.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 58%|██▎ | 1.14G/1.98G [01:44<01:15, 11.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 58%|██▎ | 1.15G/1.98G [01:45<01:12, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 59%|██▎ | 1.16G/1.98G [01:46<01:12, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 59%|██▎ | 1.17G/1.98G [01:47<01:09, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 60%|██▍ | 1.18G/1.98G [01:48<01:08, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 60%|██▍ | 1.20G/1.98G [01:49<01:07, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 61%|██▍ | 1.21G/1.98G [01:50<01:05, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 61%|██▍ | 1.22G/1.98G [01:51<01:03, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 62%|██▍ | 1.23G/1.98G [01:51<01:05, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 62%|██▍ | 1.24G/1.98G [01:52<01:02, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 63%|██▌ | 1.25G/1.98G [01:53<01:01, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 64%|██▌ | 1.26G/1.98G [01:54<01:00, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 64%|██▌ | 1.27G/1.98G [01:55<01:00, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 65%|██▌ | 1.28G/1.98G [01:56<00:58, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 65%|██▌ | 1.29G/1.98G [01:57<00:57, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 66%|██▋ | 1.30G/1.98G [01:58<00:58, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 66%|██▋ | 1.31G/1.98G [01:59<00:57, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 67%|██▋ | 1.32G/1.98G [01:59<00:55, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 67%|██▋ | 1.33G/1.98G [02:00<00:54, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 68%|██▋ | 1.34G/1.98G [02:01<00:53, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 68%|██▋ | 1.35G/1.98G [02:02<00:52, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 69%|██▊ | 1.36G/1.98G [02:03<01:00, 10.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 69%|██▊ | 1.37G/1.98G [02:04<00:57, 10.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 70%|██▊ | 1.38G/1.98G [02:05<00:54, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 70%|██▊ | 1.39G/1.98G [02:06<00:51, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 71%|██▊ | 1.41G/1.98G [02:07<00:50, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 72%|██▊ | 1.42G/1.98G [02:08<00:48, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 72%|██▉ | 1.43G/1.98G [02:09<00:47, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 73%|██▉ | 1.44G/1.98G [02:10<00:46, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 73%|██▉ | 1.45G/1.98G [02:11<00:46, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 74%|██▉ | 1.46G/1.98G [02:11<00:44, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 74%|██▉ | 1.47G/1.98G [02:12<00:44, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 75%|██▉ | 1.48G/1.98G [02:13<00:42, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 75%|███ | 1.49G/1.98G [02:14<00:41, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 76%|███ | 1.50G/1.98G [02:15<00:41, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 76%|███ | 1.51G/1.98G [02:16<00:47, 9.98MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 77%|███ | 1.52G/1.98G [02:17<00:43, 10.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 77%|███ | 1.53G/1.98G [02:20<01:05, 6.88MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 78%|███ | 1.54G/1.98G [02:21<00:55, 7.86MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 78%|███▏| 1.55G/1.98G [02:22<00:49, 8.72MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 79%|███▏| 1.56G/1.98G [02:23<00:43, 9.55MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 79%|███▏| 1.57G/1.98G [02:24<00:40, 10.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 80%|███▏| 1.58G/1.98G [02:25<00:37, 10.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 81%|███▏| 1.59G/1.98G [02:25<00:35, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 81%|███▏| 1.60G/1.98G [02:26<00:34, 10.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 82%|███▎| 1.61G/1.98G [02:27<00:32, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 82%|███▎| 1.63G/1.98G [02:28<00:31, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 83%|███▎| 1.64G/1.98G [02:29<00:30, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 83%|███▎| 1.65G/1.98G [02:30<00:28, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 84%|███▎| 1.66G/1.98G [02:31<00:27, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 84%|███▎| 1.67G/1.98G [02:32<00:27, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 85%|███▍| 1.68G/1.98G [02:33<00:25, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 85%|███▍| 1.69G/1.98G [02:33<00:24, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 86%|███▍| 1.70G/1.98G [02:34<00:23, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 86%|███▍| 1.71G/1.98G [02:35<00:22, 11.8MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 87%|███▍| 1.72G/1.98G [02:36<00:21, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 87%|███▍| 1.73G/1.98G [02:37<00:20, 12.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 88%|███▌| 1.74G/1.98G [02:38<00:20, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 88%|███▌| 1.75G/1.98G [02:39<00:22, 10.1MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 89%|███▌| 1.76G/1.98G [02:40<00:20, 10.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 90%|███▌| 1.77G/1.98G [02:41<00:18, 11.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 90%|███▌| 1.78G/1.98G [02:42<00:17, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 91%|███▌| 1.79G/1.98G [02:43<00:16, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 91%|███▋| 1.80G/1.98G [02:44<00:15, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 92%|███▋| 1.81G/1.98G [02:45<00:14, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 92%|███▋| 1.82G/1.98G [02:46<00:13, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 93%|███▋| 1.84G/1.98G [02:46<00:12, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 93%|███▋| 1.85G/1.98G [02:47<00:11, 11.4MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 94%|███▋| 1.86G/1.98G [02:48<00:10, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 94%|███▊| 1.87G/1.98G [02:49<00:09, 11.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 95%|███▊| 1.88G/1.98G [02:50<00:08, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 95%|███▊| 1.89G/1.98G [02:51<00:07, 11.7MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 96%|███▊| 1.90G/1.98G [02:52<00:08, 10.0MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 96%|███▊| 1.91G/1.98G [02:53<00:06, 10.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 97%|███▉| 1.92G/1.98G [02:54<00:05, 10.9MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 97%|███▉| 1.93G/1.98G [02:55<00:04, 11.2MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 98%|███▉| 1.94G/1.98G [02:56<00:03, 11.5MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 99%|███▉| 1.95G/1.98G [02:57<00:02, 11.3MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 99%|███▉| 1.96G/1.98G [02:58<00:01, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 100%|███▉| 1.97G/1.98G [02:58<00:00, 11.6MB/s]\u001b[A\n","model-00003-of-00008.safetensors: 100%|████| 1.98G/1.98G [02:59<00:00, 11.0MB/s]\u001b[A\n","Downloading shards: 38%|█████████ | 3/8 [08:50<14:47, 177.50s/it]\n","model-00004-of-00008.safetensors: 0%| | 0.00/1.95G [00:00, ?B/s]\u001b[A\n","model-00004-of-00008.safetensors: 1%| | 10.5M/1.95G [00:00<02:39, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 1%| | 21.0M/1.95G [00:01<02:40, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 2%| | 31.5M/1.95G [00:03<03:18, 9.67MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 2%| | 41.9M/1.95G [00:03<03:02, 10.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 3%| | 52.4M/1.95G [00:04<02:58, 10.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 3%|▏ | 62.9M/1.95G [00:05<02:49, 11.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 4%|▏ | 73.4M/1.95G [00:06<02:43, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 4%|▏ | 83.9M/1.95G [00:07<02:40, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 5%|▏ | 94.4M/1.95G [00:08<02:38, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 5%|▎ | 105M/1.95G [00:09<02:37, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 6%|▎ | 115M/1.95G [00:10<02:35, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 6%|▎ | 126M/1.95G [00:11<02:36, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 7%|▎ | 136M/1.95G [00:11<02:33, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 8%|▍ | 147M/1.95G [00:12<02:31, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 8%|▍ | 157M/1.95G [00:13<02:31, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 9%|▍ | 168M/1.95G [00:14<02:30, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 9%|▍ | 178M/1.95G [00:15<02:29, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 10%|▍ | 189M/1.95G [00:16<02:27, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 10%|▌ | 199M/1.95G [00:17<02:55, 9.94MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 11%|▌ | 210M/1.95G [00:19<03:13, 8.97MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 11%|▌ | 220M/1.95G [00:20<03:08, 9.16MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 12%|▌ | 231M/1.95G [00:21<02:55, 9.76MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 12%|▌ | 241M/1.95G [00:22<02:44, 10.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 13%|▋ | 252M/1.95G [00:22<02:35, 10.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 13%|▋ | 262M/1.95G [00:23<02:30, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 14%|▋ | 273M/1.95G [00:24<02:27, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 15%|▋ | 283M/1.95G [00:25<02:23, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 15%|▊ | 294M/1.95G [00:26<02:21, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 16%|▊ | 304M/1.95G [00:27<02:23, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 16%|▊ | 315M/1.95G [00:28<02:20, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 17%|▊ | 325M/1.95G [00:29<02:17, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 17%|▊ | 336M/1.95G [00:30<02:15, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 18%|▉ | 346M/1.95G [00:30<02:15, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 18%|▉ | 357M/1.95G [00:31<02:14, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 19%|▉ | 367M/1.95G [00:32<02:12, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 19%|▉ | 377M/1.95G [00:33<02:13, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 20%|▉ | 388M/1.95G [00:34<02:11, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 20%|█ | 398M/1.95G [00:35<02:09, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 21%|█ | 409M/1.95G [00:36<02:10, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 22%|█ | 419M/1.95G [00:37<02:09, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 22%|█ | 430M/1.95G [00:38<02:07, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 23%|█▏ | 440M/1.95G [00:39<02:28, 10.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 23%|█▏ | 451M/1.95G [00:40<02:21, 10.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 24%|█▏ | 461M/1.95G [00:41<02:15, 10.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 24%|█▏ | 472M/1.95G [00:42<02:11, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 25%|█▏ | 482M/1.95G [00:43<02:10, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 25%|█▎ | 493M/1.95G [00:43<02:06, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 26%|█▎ | 503M/1.95G [00:44<02:03, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 26%|█▎ | 514M/1.95G [00:45<02:02, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 27%|█▎ | 524M/1.95G [00:46<02:01, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 27%|█▎ | 535M/1.95G [00:47<01:58, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 28%|█▍ | 545M/1.95G [00:48<01:58, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 29%|█▍ | 556M/1.95G [00:49<01:58, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 29%|█▍ | 566M/1.95G [00:50<01:55, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 30%|█▍ | 577M/1.95G [00:50<01:54, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 30%|█▌ | 587M/1.95G [00:52<02:12, 10.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 31%|█▌ | 598M/1.95G [00:53<02:08, 10.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 31%|█▌ | 608M/1.95G [00:54<02:02, 11.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 32%|█▌ | 619M/1.95G [00:54<01:57, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 32%|█▌ | 629M/1.95G [00:55<01:55, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 33%|█▋ | 640M/1.95G [00:56<01:53, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 33%|█▋ | 650M/1.95G [00:57<01:50, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 34%|█▋ | 661M/1.95G [00:58<01:48, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 34%|█▋ | 671M/1.95G [00:59<01:48, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 35%|█▊ | 682M/1.95G [01:00<01:46, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 36%|█▊ | 692M/1.95G [01:01<01:46, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 36%|█▊ | 703M/1.95G [01:01<01:45, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 37%|█▊ | 713M/1.95G [01:02<01:44, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 37%|█▊ | 724M/1.95G [01:03<01:41, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 38%|█▉ | 734M/1.95G [01:04<01:41, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 38%|█▉ | 744M/1.95G [01:05<01:40, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 39%|█▉ | 755M/1.95G [01:06<01:56, 10.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 39%|█▉ | 765M/1.95G [01:07<01:50, 10.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 40%|█▉ | 776M/1.95G [01:08<01:46, 11.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 40%|██ | 786M/1.95G [01:09<01:43, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 41%|██ | 797M/1.95G [01:10<01:40, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 41%|██ | 807M/1.95G [01:11<01:37, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 42%|██ | 818M/1.95G [01:12<01:36, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 43%|██▏ | 828M/1.95G [01:15<02:53, 6.46MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 43%|██▏ | 839M/1.95G [01:16<02:28, 7.43MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 44%|██▏ | 849M/1.95G [01:17<02:11, 8.31MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 44%|██▏ | 860M/1.95G [01:18<01:58, 9.17MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 45%|██▏ | 870M/1.95G [01:19<01:49, 9.83MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 45%|██▎ | 881M/1.95G [01:21<02:28, 7.15MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 46%|██▎ | 891M/1.95G [01:22<02:11, 8.00MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 46%|██▎ | 902M/1.95G [01:23<01:57, 8.91MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 47%|██▎ | 912M/1.95G [01:24<01:47, 9.63MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 47%|██▎ | 923M/1.95G [01:24<01:39, 10.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 48%|██▍ | 933M/1.95G [01:25<01:33, 10.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 48%|██▍ | 944M/1.95G [01:26<01:30, 11.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 49%|██▍ | 954M/1.95G [01:27<01:28, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 50%|██▍ | 965M/1.95G [01:28<01:39, 9.86MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 50%|██▌ | 975M/1.95G [01:29<01:34, 10.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 51%|██▌ | 986M/1.95G [01:30<01:30, 10.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 51%|██▌ | 996M/1.95G [01:31<01:26, 11.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 52%|██ | 1.01G/1.95G [01:32<01:23, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 52%|██ | 1.02G/1.95G [01:33<01:21, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 53%|██ | 1.03G/1.95G [01:34<01:19, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 53%|██▏ | 1.04G/1.95G [01:35<01:19, 11.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 54%|██▏ | 1.05G/1.95G [01:36<01:17, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 54%|██▏ | 1.06G/1.95G [01:36<01:15, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 55%|██▏ | 1.07G/1.95G [01:37<01:15, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 55%|██▏ | 1.08G/1.95G [01:38<01:13, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 56%|██▏ | 1.09G/1.95G [01:39<01:13, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 57%|██▎ | 1.10G/1.95G [01:40<01:12, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 57%|██▎ | 1.11G/1.95G [01:41<01:10, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 58%|██▎ | 1.12G/1.95G [01:42<01:10, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 58%|██▎ | 1.13G/1.95G [01:43<01:08, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 59%|██▎ | 1.14G/1.95G [01:44<01:07, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 59%|██▎ | 1.15G/1.95G [01:44<01:06, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 60%|██▍ | 1.16G/1.95G [01:45<01:05, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 60%|██▍ | 1.17G/1.95G [01:46<01:04, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 61%|██▍ | 1.18G/1.95G [01:47<01:03, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 61%|██▍ | 1.20G/1.95G [01:48<01:02, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 62%|██▍ | 1.21G/1.95G [01:49<01:02, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 62%|██▍ | 1.22G/1.95G [01:50<01:01, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 63%|██▌ | 1.23G/1.95G [01:51<01:00, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 64%|██▌ | 1.24G/1.95G [01:51<00:59, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 64%|██▌ | 1.25G/1.95G [01:52<00:58, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 65%|██▌ | 1.26G/1.95G [01:53<00:57, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 65%|██▌ | 1.27G/1.95G [01:54<00:56, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 66%|██▋ | 1.28G/1.95G [01:55<00:55, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 66%|██▋ | 1.29G/1.95G [01:56<00:54, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 67%|██▋ | 1.30G/1.95G [01:57<00:53, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 67%|██▋ | 1.31G/1.95G [01:58<00:52, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 68%|██▋ | 1.32G/1.95G [01:59<01:01, 10.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 68%|██▋ | 1.33G/1.95G [02:00<00:57, 10.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 69%|██▊ | 1.34G/1.95G [02:01<00:54, 11.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 70%|██▊ | 1.35G/1.95G [02:02<00:52, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 70%|██▊ | 1.36G/1.95G [02:02<00:50, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 71%|██▊ | 1.37G/1.95G [02:03<00:49, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 71%|██▊ | 1.38G/1.95G [02:04<00:48, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 72%|██▊ | 1.39G/1.95G [02:05<00:46, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 72%|██▉ | 1.41G/1.95G [02:06<00:45, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 73%|██▉ | 1.42G/1.95G [02:07<00:44, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 73%|██▉ | 1.43G/1.95G [02:08<00:43, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 74%|██▉ | 1.44G/1.95G [02:09<00:42, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 74%|██▉ | 1.45G/1.95G [02:09<00:41, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 75%|██▉ | 1.46G/1.95G [02:10<00:40, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 75%|███ | 1.47G/1.95G [02:11<00:39, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 76%|███ | 1.48G/1.95G [02:12<00:38, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 77%|███ | 1.49G/1.95G [02:13<00:37, 12.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 77%|███ | 1.50G/1.95G [02:14<00:37, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 78%|███ | 1.51G/1.95G [02:15<00:42, 10.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 78%|███ | 1.52G/1.95G [02:16<00:39, 10.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 79%|███▏| 1.53G/1.95G [02:17<00:37, 11.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 79%|███▏| 1.54G/1.95G [02:18<00:35, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 80%|███▏| 1.55G/1.95G [02:19<00:34, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 80%|███▏| 1.56G/1.95G [02:20<00:33, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 81%|███▏| 1.57G/1.95G [02:22<00:44, 8.43MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 81%|███▎| 1.58G/1.95G [02:23<00:39, 9.17MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 82%|███▎| 1.59G/1.95G [02:23<00:35, 9.93MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 82%|███▎| 1.60G/1.95G [02:24<00:33, 10.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 83%|███▎| 1.61G/1.95G [02:25<00:30, 10.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 84%|███▎| 1.63G/1.95G [02:26<00:28, 11.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 84%|███▎| 1.64G/1.95G [02:27<00:27, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 85%|███▍| 1.65G/1.95G [02:28<00:25, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 85%|███▍| 1.66G/1.95G [02:29<00:29, 9.92MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 86%|███▍| 1.67G/1.95G [02:30<00:26, 10.4MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 86%|███▍| 1.68G/1.95G [02:31<00:24, 10.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 87%|███▍| 1.69G/1.95G [02:32<00:23, 11.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 87%|███▍| 1.70G/1.95G [02:33<00:21, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 88%|███▌| 1.71G/1.95G [02:34<00:20, 11.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 88%|███▌| 1.72G/1.95G [02:35<00:19, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 89%|███▌| 1.73G/1.95G [02:35<00:18, 11.6MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 89%|███▌| 1.74G/1.95G [02:36<00:17, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 90%|███▌| 1.75G/1.95G [02:37<00:16, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 91%|███▌| 1.76G/1.95G [02:38<00:15, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 91%|███▋| 1.77G/1.95G [02:39<00:14, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 92%|███▋| 1.78G/1.95G [02:40<00:13, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 92%|███▋| 1.79G/1.95G [02:41<00:12, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 93%|███▋| 1.80G/1.95G [02:42<00:12, 11.7MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 93%|███▋| 1.81G/1.95G [02:42<00:11, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 94%|███▋| 1.82G/1.95G [02:43<00:10, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 94%|███▊| 1.84G/1.95G [02:44<00:09, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 95%|███▊| 1.85G/1.95G [02:45<00:08, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 95%|███▊| 1.86G/1.95G [02:46<00:07, 12.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 96%|███▊| 1.87G/1.95G [02:47<00:06, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 96%|███▊| 1.88G/1.95G [02:48<00:05, 11.8MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 97%|███▉| 1.89G/1.95G [02:49<00:04, 11.9MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 98%|███▉| 1.90G/1.95G [02:50<00:04, 10.1MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 98%|███▉| 1.91G/1.95G [02:51<00:03, 10.5MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 99%|███▉| 1.92G/1.95G [02:52<00:02, 11.0MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 99%|███▉| 1.93G/1.95G [02:53<00:01, 11.2MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 100%|███▉| 1.94G/1.95G [02:54<00:00, 11.3MB/s]\u001b[A\n","model-00004-of-00008.safetensors: 100%|████| 1.95G/1.95G [02:54<00:00, 11.1MB/s]\u001b[A\n","Downloading shards: 50%|████████████ | 4/8 [11:45<11:46, 176.73s/it]\n","model-00005-of-00008.safetensors: 0%| | 0.00/1.98G [00:00, ?B/s]\u001b[A\n","model-00005-of-00008.safetensors: 1%| | 10.5M/1.98G [00:00<02:42, 12.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 1%| | 21.0M/1.98G [00:01<02:44, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 2%| | 31.5M/1.98G [00:02<02:42, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 2%| | 41.9M/1.98G [00:03<02:41, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 3%| | 52.4M/1.98G [00:04<02:41, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 3%|▏ | 62.9M/1.98G [00:05<02:43, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 4%|▏ | 73.4M/1.98G [00:06<03:09, 10.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 4%|▏ | 83.9M/1.98G [00:07<02:58, 10.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 5%|▏ | 94.4M/1.98G [00:08<02:50, 11.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 5%|▎ | 105M/1.98G [00:09<02:46, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 6%|▎ | 115M/1.98G [00:10<02:44, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 6%|▎ | 126M/1.98G [00:11<02:40, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 7%|▎ | 136M/1.98G [00:11<02:37, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 7%|▎ | 147M/1.98G [00:12<02:36, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 8%|▍ | 157M/1.98G [00:13<02:35, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 8%|▍ | 168M/1.98G [00:14<02:33, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 9%|▍ | 178M/1.98G [00:15<02:31, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 10%|▍ | 189M/1.98G [00:16<02:31, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 10%|▌ | 199M/1.98G [00:17<02:29, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 11%|▌ | 210M/1.98G [00:18<02:27, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 11%|▌ | 220M/1.98G [00:18<02:27, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 12%|▌ | 231M/1.98G [00:19<02:28, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 12%|▌ | 241M/1.98G [00:20<02:26, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 13%|▋ | 252M/1.98G [00:21<02:24, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 13%|▋ | 262M/1.98G [00:22<02:24, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 14%|▋ | 273M/1.98G [00:23<02:46, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 14%|▋ | 283M/1.98G [00:24<02:38, 10.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 15%|▋ | 294M/1.98G [00:25<02:32, 11.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 15%|▊ | 304M/1.98G [00:27<03:27, 8.08MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 16%|▊ | 315M/1.98G [00:28<03:15, 8.53MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 16%|▊ | 325M/1.98G [00:29<02:56, 9.35MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 17%|▊ | 336M/1.98G [00:30<02:44, 10.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 17%|▊ | 346M/1.98G [00:31<02:35, 10.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 18%|▉ | 357M/1.98G [00:32<02:28, 10.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 19%|▉ | 367M/1.98G [00:33<02:22, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 19%|▉ | 377M/1.98G [00:34<02:19, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 20%|▉ | 388M/1.98G [00:34<02:17, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 20%|█ | 398M/1.98G [00:35<02:14, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 21%|█ | 409M/1.98G [00:36<02:13, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 21%|█ | 419M/1.98G [00:37<02:12, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 22%|█ | 430M/1.98G [00:38<02:11, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 22%|█ | 440M/1.98G [00:39<02:09, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 23%|█▏ | 451M/1.98G [00:40<02:07, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 23%|█▏ | 461M/1.98G [00:41<02:07, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 24%|█▏ | 472M/1.98G [00:42<02:28, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 24%|█▏ | 482M/1.98G [00:43<02:19, 10.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 25%|█▏ | 493M/1.98G [00:44<02:13, 11.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 25%|█▎ | 503M/1.98G [00:45<02:10, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 26%|█▎ | 514M/1.98G [00:45<02:07, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 26%|█▎ | 524M/1.98G [00:46<02:04, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 27%|█▎ | 535M/1.98G [00:47<02:02, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 28%|█▍ | 545M/1.98G [00:48<02:00, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 28%|█▍ | 556M/1.98G [00:49<01:59, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 29%|█▍ | 566M/1.98G [00:50<01:57, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 29%|█▍ | 577M/1.98G [00:51<01:59, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 30%|█▍ | 587M/1.98G [00:52<01:57, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 30%|█▌ | 598M/1.98G [00:52<01:56, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 31%|█▌ | 608M/1.98G [00:53<01:55, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 31%|█▌ | 619M/1.98G [00:54<01:54, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 32%|█▌ | 629M/1.98G [00:55<01:53, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 32%|█▌ | 640M/1.98G [00:56<01:51, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 33%|█▋ | 650M/1.98G [00:57<01:50, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 33%|█▋ | 661M/1.98G [00:58<01:50, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 34%|█▋ | 671M/1.98G [00:59<02:07, 10.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 34%|█▋ | 682M/1.98G [01:00<02:00, 10.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 35%|█▋ | 692M/1.98G [01:01<01:56, 11.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 35%|█▊ | 703M/1.98G [01:02<01:53, 11.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 36%|█▊ | 713M/1.98G [01:03<01:49, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 37%|█▊ | 724M/1.98G [01:03<01:46, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 37%|█▊ | 734M/1.98G [01:04<01:45, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 38%|█▉ | 744M/1.98G [01:05<01:44, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 38%|█▉ | 755M/1.98G [01:06<01:42, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 39%|█▉ | 765M/1.98G [01:07<01:41, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 39%|█▉ | 776M/1.98G [01:08<01:41, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 40%|█▉ | 786M/1.98G [01:09<01:43, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 40%|██ | 797M/1.98G [01:10<01:42, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 41%|██ | 807M/1.98G [01:11<01:39, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 41%|██ | 818M/1.98G [01:11<01:37, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 42%|██ | 828M/1.98G [01:12<01:37, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 42%|██ | 839M/1.98G [01:13<01:36, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 43%|██▏ | 849M/1.98G [01:14<01:35, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 43%|██▏ | 860M/1.98G [01:15<01:33, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 44%|██▏ | 870M/1.98G [01:16<01:32, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 44%|██▏ | 881M/1.98G [01:17<01:31, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 45%|██▎ | 891M/1.98G [01:22<03:54, 4.64MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 46%|██▎ | 902M/1.98G [01:23<03:15, 5.52MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 46%|██▎ | 912M/1.98G [01:24<02:41, 6.62MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 47%|██▎ | 923M/1.98G [01:25<02:33, 6.89MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 47%|██▎ | 933M/1.98G [01:26<02:13, 7.86MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 48%|██▍ | 944M/1.98G [01:28<02:33, 6.75MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 48%|██▍ | 954M/1.98G [01:29<02:15, 7.59MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 49%|██▍ | 965M/1.98G [01:30<01:58, 8.58MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 49%|██▍ | 975M/1.98G [01:31<01:48, 9.28MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 50%|██▍ | 986M/1.98G [01:32<01:39, 10.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 50%|██▌ | 996M/1.98G [01:33<01:32, 10.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 51%|██ | 1.01G/1.98G [01:34<01:28, 11.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 51%|██ | 1.02G/1.98G [01:35<01:25, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 52%|██ | 1.03G/1.98G [01:35<01:23, 11.4MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 52%|██ | 1.04G/1.98G [01:36<01:22, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 53%|██ | 1.05G/1.98G [01:37<01:20, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 53%|██▏ | 1.06G/1.98G [01:38<01:19, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 54%|██▏ | 1.07G/1.98G [01:39<01:18, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 55%|██▏ | 1.08G/1.98G [01:40<01:16, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 55%|██▏ | 1.09G/1.98G [01:41<01:15, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 56%|██▏ | 1.10G/1.98G [01:42<01:15, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 56%|██▏ | 1.11G/1.98G [01:43<01:13, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 57%|██▎ | 1.12G/1.98G [01:44<01:12, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 57%|██▎ | 1.13G/1.98G [01:45<01:23, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 58%|██▎ | 1.14G/1.98G [01:46<01:18, 10.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 58%|██▎ | 1.15G/1.98G [01:47<01:15, 10.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 59%|██▎ | 1.16G/1.98G [01:48<01:12, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 59%|██▎ | 1.17G/1.98G [01:48<01:10, 11.4MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 60%|██▍ | 1.18G/1.98G [01:49<01:08, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 60%|██▍ | 1.20G/1.98G [01:50<01:07, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 61%|██▍ | 1.21G/1.98G [01:51<01:06, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 61%|██▍ | 1.22G/1.98G [01:52<01:05, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 62%|██▍ | 1.23G/1.98G [01:53<01:03, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 62%|██▍ | 1.24G/1.98G [01:54<01:02, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 63%|██▌ | 1.25G/1.98G [01:55<01:00, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 64%|██▌ | 1.26G/1.98G [01:55<01:00, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 64%|██▌ | 1.27G/1.98G [01:56<00:59, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 65%|██▌ | 1.28G/1.98G [01:57<00:58, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 65%|██▌ | 1.29G/1.98G [01:58<00:57, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 66%|██▋ | 1.30G/1.98G [01:59<00:57, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 66%|██▋ | 1.31G/1.98G [02:00<00:56, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 67%|██▋ | 1.32G/1.98G [02:01<00:55, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 67%|██▋ | 1.33G/1.98G [02:02<00:54, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 68%|██▋ | 1.34G/1.98G [02:03<00:54, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 68%|██▋ | 1.35G/1.98G [02:03<00:52, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 69%|██▊ | 1.36G/1.98G [02:05<01:00, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 69%|██▊ | 1.37G/1.98G [02:06<00:57, 10.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 70%|██▊ | 1.38G/1.98G [02:07<00:55, 10.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 70%|██▊ | 1.39G/1.98G [02:07<00:52, 11.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 71%|██▊ | 1.41G/1.98G [02:08<00:50, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 72%|██▊ | 1.42G/1.98G [02:09<00:49, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 72%|██▉ | 1.43G/1.98G [02:10<00:47, 11.6MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 73%|██▉ | 1.44G/1.98G [02:11<00:47, 11.4MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 73%|██▉ | 1.45G/1.98G [02:12<00:46, 11.4MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 74%|██▉ | 1.46G/1.98G [02:13<00:45, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 74%|██▉ | 1.47G/1.98G [02:14<00:43, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 75%|██▉ | 1.48G/1.98G [02:15<00:42, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 75%|███ | 1.49G/1.98G [02:16<00:41, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 76%|███ | 1.50G/1.98G [02:16<00:40, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 76%|███ | 1.51G/1.98G [02:17<00:39, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 77%|███ | 1.52G/1.98G [02:20<01:01, 7.52MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 77%|███ | 1.53G/1.98G [02:22<01:09, 6.42MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 78%|███ | 1.54G/1.98G [02:23<01:00, 7.20MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 78%|███▏| 1.55G/1.98G [02:24<00:53, 8.03MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 79%|███▏| 1.56G/1.98G [02:25<00:46, 8.90MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 79%|███▏| 1.57G/1.98G [02:26<00:42, 9.66MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 80%|███▏| 1.58G/1.98G [02:27<00:39, 10.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 81%|███▏| 1.59G/1.98G [02:28<00:36, 10.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 81%|███▏| 1.60G/1.98G [02:30<00:49, 7.57MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 82%|███▎| 1.61G/1.98G [02:31<00:43, 8.44MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 82%|███▎| 1.63G/1.98G [02:32<00:38, 9.28MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 83%|███▎| 1.64G/1.98G [02:33<00:35, 9.65MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 83%|███▎| 1.65G/1.98G [02:34<00:32, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 84%|███▎| 1.66G/1.98G [02:34<00:30, 10.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 84%|███▎| 1.67G/1.98G [02:35<00:28, 11.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 85%|███▍| 1.68G/1.98G [02:36<00:26, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 85%|███▍| 1.69G/1.98G [02:37<00:25, 11.4MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 86%|███▍| 1.70G/1.98G [02:38<00:24, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 86%|███▍| 1.71G/1.98G [02:39<00:23, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 87%|███▍| 1.72G/1.98G [02:40<00:21, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 87%|███▍| 1.73G/1.98G [02:41<00:21, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 88%|███▌| 1.74G/1.98G [02:41<00:20, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 88%|███▌| 1.75G/1.98G [02:42<00:19, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 89%|███▌| 1.76G/1.98G [02:43<00:18, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 90%|███▌| 1.77G/1.98G [02:44<00:17, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 90%|███▌| 1.78G/1.98G [02:45<00:16, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 91%|███▌| 1.79G/1.98G [02:46<00:15, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 91%|███▋| 1.80G/1.98G [02:47<00:14, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 92%|███▋| 1.81G/1.98G [02:48<00:13, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 92%|███▋| 1.82G/1.98G [02:49<00:13, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 93%|███▋| 1.84G/1.98G [02:49<00:12, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 93%|███▋| 1.85G/1.98G [02:50<00:11, 11.8MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 94%|███▋| 1.86G/1.98G [02:51<00:10, 11.9MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 94%|███▊| 1.87G/1.98G [02:52<00:09, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 95%|███▊| 1.88G/1.98G [02:53<00:08, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 95%|███▊| 1.89G/1.98G [02:54<00:07, 12.0MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 96%|███▊| 1.90G/1.98G [02:55<00:06, 12.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 96%|███▊| 1.91G/1.98G [02:56<00:06, 10.2MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 97%|███▉| 1.92G/1.98G [02:57<00:05, 10.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 97%|███▉| 1.93G/1.98G [02:58<00:04, 11.1MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 98%|███▉| 1.94G/1.98G [02:59<00:03, 11.3MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 99%|███▉| 1.95G/1.98G [02:59<00:02, 11.5MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 99%|███▉| 1.96G/1.98G [03:00<00:01, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 100%|███▉| 1.97G/1.98G [03:01<00:00, 11.7MB/s]\u001b[A\n","model-00005-of-00008.safetensors: 100%|████| 1.98G/1.98G [03:02<00:00, 10.9MB/s]\u001b[A\n","Downloading shards: 62%|███████████████ | 5/8 [14:48<08:57, 179.01s/it]\n","model-00006-of-00008.safetensors: 0%| | 0.00/1.95G [00:00, ?B/s]\u001b[A\n","model-00006-of-00008.safetensors: 1%| | 10.5M/1.95G [00:00<02:38, 12.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 1%| | 21.0M/1.95G [00:01<02:41, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 2%| | 31.5M/1.95G [00:02<02:39, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 2%| | 41.9M/1.95G [00:03<02:39, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 3%| | 52.4M/1.95G [00:04<02:39, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 3%|▏ | 62.9M/1.95G [00:05<02:41, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 4%|▏ | 73.4M/1.95G [00:06<02:38, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 4%|▏ | 83.9M/1.95G [00:07<02:36, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 5%|▏ | 94.4M/1.95G [00:07<02:37, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 5%|▎ | 105M/1.95G [00:08<02:36, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 6%|▎ | 115M/1.95G [00:09<02:34, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 6%|▎ | 126M/1.95G [00:10<02:33, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 7%|▎ | 136M/1.95G [00:11<02:33, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 8%|▍ | 147M/1.95G [00:12<02:33, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 8%|▍ | 157M/1.95G [00:13<02:29, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 9%|▍ | 168M/1.95G [00:14<02:55, 10.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 9%|▍ | 178M/1.95G [00:15<02:46, 10.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 10%|▍ | 189M/1.95G [00:16<02:40, 11.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 10%|▌ | 199M/1.95G [00:17<02:34, 11.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 11%|▌ | 210M/1.95G [00:18<02:32, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 11%|▌ | 220M/1.95G [00:19<02:34, 11.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 12%|▌ | 231M/1.95G [00:20<02:30, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 12%|▌ | 241M/1.95G [00:20<02:26, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 13%|▋ | 252M/1.95G [00:21<02:25, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 13%|▋ | 262M/1.95G [00:22<02:22, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 14%|▋ | 273M/1.95G [00:23<02:22, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 15%|▋ | 283M/1.95G [00:24<02:20, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 15%|▊ | 294M/1.95G [00:25<02:20, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 16%|▊ | 304M/1.95G [00:26<02:19, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 16%|▊ | 315M/1.95G [00:28<03:39, 7.43MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 17%|▊ | 325M/1.95G [00:29<03:13, 8.40MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 17%|▊ | 336M/1.95G [00:30<02:55, 9.18MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 18%|▉ | 346M/1.95G [00:31<02:42, 9.85MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 18%|▉ | 357M/1.95G [00:32<02:33, 10.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 19%|▉ | 367M/1.95G [00:33<02:52, 9.15MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 19%|▉ | 377M/1.95G [00:34<02:39, 9.85MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 20%|▉ | 388M/1.95G [00:35<02:28, 10.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 20%|█ | 398M/1.95G [00:36<02:23, 10.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 21%|█ | 409M/1.95G [00:37<02:17, 11.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 22%|█ | 419M/1.95G [00:38<02:15, 11.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 22%|█ | 430M/1.95G [00:39<02:11, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 23%|█▏ | 440M/1.95G [00:40<02:12, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 23%|█▏ | 451M/1.95G [00:40<02:09, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 24%|█▏ | 461M/1.95G [00:41<02:08, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 24%|█▏ | 472M/1.95G [00:42<02:08, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 25%|█▏ | 482M/1.95G [00:43<02:05, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 25%|█▎ | 493M/1.95G [00:44<02:02, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 26%|█▎ | 503M/1.95G [00:45<02:24, 9.99MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 26%|█▎ | 514M/1.95G [00:46<02:17, 10.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 27%|█▎ | 524M/1.95G [00:47<02:11, 10.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 27%|█▎ | 535M/1.95G [00:48<02:06, 11.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 28%|█▍ | 545M/1.95G [00:49<02:04, 11.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 29%|█▍ | 556M/1.95G [00:50<02:00, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 29%|█▍ | 566M/1.95G [00:51<01:58, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 30%|█▍ | 577M/1.95G [00:52<01:58, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 30%|█▌ | 587M/1.95G [00:53<01:56, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 31%|█▌ | 598M/1.95G [00:53<01:55, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 31%|█▌ | 608M/1.95G [00:54<01:57, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 32%|█▌ | 619M/1.95G [00:55<01:53, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 32%|█▌ | 629M/1.95G [00:56<01:51, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 33%|█▋ | 640M/1.95G [00:57<01:51, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 33%|█▋ | 650M/1.95G [00:58<01:50, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 34%|█▋ | 661M/1.95G [00:59<01:49, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 34%|█▋ | 671M/1.95G [01:00<02:05, 10.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 35%|█▊ | 682M/1.95G [01:01<02:01, 10.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 36%|█▊ | 692M/1.95G [01:02<01:55, 10.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 36%|█▊ | 703M/1.95G [01:03<01:53, 11.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 37%|█▊ | 713M/1.95G [01:04<01:48, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 37%|█▊ | 724M/1.95G [01:05<01:46, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 38%|█▉ | 734M/1.95G [01:06<01:46, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 38%|█▉ | 744M/1.95G [01:06<01:44, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 39%|█▉ | 755M/1.95G [01:07<01:42, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 39%|█▉ | 765M/1.95G [01:08<01:40, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 40%|█▉ | 776M/1.95G [01:09<01:39, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 40%|██ | 786M/1.95G [01:10<01:38, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 41%|██ | 797M/1.95G [01:11<01:37, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 41%|██ | 807M/1.95G [01:12<01:38, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 42%|██ | 818M/1.95G [01:13<01:36, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 43%|██▏ | 828M/1.95G [01:14<01:35, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 43%|██▏ | 839M/1.95G [01:15<01:37, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 44%|██▏ | 849M/1.95G [01:21<04:20, 4.22MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 44%|██▏ | 860M/1.95G [01:22<03:27, 5.24MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 45%|██▏ | 870M/1.95G [01:22<02:50, 6.31MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 45%|██▎ | 881M/1.95G [01:23<02:23, 7.40MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 46%|██▎ | 891M/1.95G [01:24<02:05, 8.39MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 46%|██▎ | 902M/1.95G [01:25<01:53, 9.16MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 47%|██▎ | 912M/1.95G [01:26<01:44, 9.88MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 47%|██▎ | 923M/1.95G [01:27<01:37, 10.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 48%|██▍ | 933M/1.95G [01:28<01:53, 8.93MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 48%|██▍ | 944M/1.95G [01:29<01:43, 9.64MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 49%|██▍ | 954M/1.95G [01:30<01:37, 10.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 50%|██▍ | 965M/1.95G [01:31<01:31, 10.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 50%|██▌ | 975M/1.95G [01:32<01:27, 11.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 51%|██▌ | 986M/1.95G [01:33<01:25, 11.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 51%|██▌ | 996M/1.95G [01:34<01:22, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 52%|██ | 1.01G/1.95G [01:35<01:34, 9.97MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 52%|██ | 1.02G/1.95G [01:36<01:29, 10.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 53%|██ | 1.03G/1.95G [01:37<01:24, 10.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 53%|██▏ | 1.04G/1.95G [01:38<01:21, 11.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 54%|██▏ | 1.05G/1.95G [01:39<01:18, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 54%|██▏ | 1.06G/1.95G [01:39<01:16, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 55%|██▏ | 1.07G/1.95G [01:40<01:14, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 55%|██▏ | 1.08G/1.95G [01:41<01:13, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 56%|██▏ | 1.09G/1.95G [01:42<01:11, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 57%|██▎ | 1.10G/1.95G [01:43<01:10, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 57%|██▎ | 1.11G/1.95G [01:44<01:10, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 58%|██▎ | 1.12G/1.95G [01:45<01:09, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 58%|██▎ | 1.13G/1.95G [01:46<01:08, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 59%|██▎ | 1.14G/1.95G [01:46<01:07, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 59%|██▎ | 1.15G/1.95G [01:47<01:06, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 60%|██▍ | 1.16G/1.95G [01:48<01:05, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 60%|██▍ | 1.17G/1.95G [01:49<01:04, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 61%|██▍ | 1.18G/1.95G [01:50<01:03, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 61%|██▍ | 1.20G/1.95G [01:51<01:03, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 62%|██▍ | 1.21G/1.95G [01:52<01:01, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 62%|██▍ | 1.22G/1.95G [01:53<01:00, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 63%|██▌ | 1.23G/1.95G [01:53<00:59, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 64%|██▌ | 1.24G/1.95G [01:55<01:09, 10.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 64%|██▌ | 1.25G/1.95G [01:56<01:05, 10.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 65%|██▌ | 1.26G/1.95G [01:57<01:03, 10.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 65%|██▌ | 1.27G/1.95G [01:57<01:00, 11.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 66%|██▋ | 1.28G/1.95G [01:58<00:58, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 66%|██▋ | 1.29G/1.95G [01:59<00:56, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 67%|██▋ | 1.30G/1.95G [02:00<00:55, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 67%|██▋ | 1.31G/1.95G [02:01<00:54, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 68%|██▋ | 1.32G/1.95G [02:02<00:52, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 68%|██▋ | 1.33G/1.95G [02:03<00:52, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 69%|██▊ | 1.34G/1.95G [02:04<00:51, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 70%|██▊ | 1.35G/1.95G [02:04<00:49, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 70%|██▊ | 1.36G/1.95G [02:05<00:48, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 71%|██▊ | 1.37G/1.95G [02:06<00:48, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 71%|██▊ | 1.38G/1.95G [02:07<00:47, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 72%|██▊ | 1.39G/1.95G [02:09<00:54, 10.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 72%|██▉ | 1.41G/1.95G [02:09<00:50, 10.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 73%|██▉ | 1.42G/1.95G [02:10<00:48, 11.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 73%|██▉ | 1.43G/1.95G [02:11<00:45, 11.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 74%|██▉ | 1.44G/1.95G [02:12<00:44, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 74%|██▉ | 1.45G/1.95G [02:13<00:43, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 75%|██▉ | 1.46G/1.95G [02:14<00:41, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 75%|███ | 1.47G/1.95G [02:15<00:40, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 76%|███ | 1.48G/1.95G [02:16<00:39, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 77%|███ | 1.49G/1.95G [02:16<00:38, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 77%|███ | 1.50G/1.95G [02:17<00:37, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 78%|███ | 1.51G/1.95G [02:18<00:36, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 78%|███ | 1.52G/1.95G [02:19<00:35, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 79%|███▏| 1.53G/1.95G [02:20<00:34, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 79%|███▏| 1.54G/1.95G [02:21<00:33, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 80%|███▏| 1.55G/1.95G [02:22<00:32, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 80%|███▏| 1.56G/1.95G [02:23<00:32, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 81%|███▏| 1.57G/1.95G [02:23<00:31, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 81%|███▎| 1.58G/1.95G [02:24<00:30, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 82%|███▎| 1.59G/1.95G [02:25<00:29, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 82%|███▎| 1.60G/1.95G [02:26<00:28, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 83%|███▎| 1.61G/1.95G [02:27<00:27, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 84%|███▎| 1.63G/1.95G [02:28<00:26, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 84%|███▎| 1.64G/1.95G [02:29<00:32, 9.55MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 85%|███▍| 1.65G/1.95G [02:30<00:29, 10.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 85%|███▍| 1.66G/1.95G [02:31<00:27, 10.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 86%|███▍| 1.67G/1.95G [02:32<00:25, 11.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 86%|███▍| 1.68G/1.95G [02:33<00:23, 11.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 87%|███▍| 1.69G/1.95G [02:34<00:22, 11.4MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 87%|███▍| 1.70G/1.95G [02:35<00:21, 11.6MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 88%|███▌| 1.71G/1.95G [02:36<00:20, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 88%|███▌| 1.72G/1.95G [02:36<00:19, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 89%|███▌| 1.73G/1.95G [02:37<00:18, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 89%|███▌| 1.74G/1.95G [02:38<00:17, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 90%|███▌| 1.75G/1.95G [02:39<00:16, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 91%|███▌| 1.76G/1.95G [02:40<00:15, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 91%|███▋| 1.77G/1.95G [02:41<00:14, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 92%|███▋| 1.78G/1.95G [02:42<00:13, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 92%|███▋| 1.79G/1.95G [02:43<00:12, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 93%|███▋| 1.80G/1.95G [02:43<00:11, 12.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 93%|███▋| 1.81G/1.95G [02:44<00:10, 12.1MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 94%|███▋| 1.82G/1.95G [02:45<00:10, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 94%|███▊| 1.84G/1.95G [02:47<00:10, 10.2MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 95%|███▊| 1.85G/1.95G [02:47<00:09, 10.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 95%|███▊| 1.86G/1.95G [02:48<00:08, 11.0MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 96%|███▊| 1.87G/1.95G [02:49<00:07, 11.3MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 96%|███▊| 1.88G/1.95G [02:50<00:06, 11.5MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 97%|███▉| 1.89G/1.95G [02:51<00:05, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 98%|███▉| 1.90G/1.95G [02:52<00:04, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 98%|███▉| 1.91G/1.95G [02:53<00:03, 11.7MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 99%|███▉| 1.92G/1.95G [02:54<00:02, 11.8MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 99%|███▉| 1.93G/1.95G [02:54<00:01, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 100%|███▉| 1.94G/1.95G [02:55<00:00, 11.9MB/s]\u001b[A\n","model-00006-of-00008.safetensors: 100%|████| 1.95G/1.95G [02:56<00:00, 11.0MB/s]\u001b[A\n","Downloading shards: 75%|██████████████████ | 6/8 [17:45<05:56, 178.31s/it]\n","model-00007-of-00008.safetensors: 0%| | 0.00/1.98G [00:00, ?B/s]\u001b[A\n","model-00007-of-00008.safetensors: 1%| | 10.5M/1.98G [00:00<02:44, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 1%| | 21.0M/1.98G [00:01<02:45, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 2%| | 31.5M/1.98G [00:02<02:42, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 2%| | 41.9M/1.98G [00:03<02:41, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 3%| | 52.4M/1.98G [00:04<02:41, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 3%|▏ | 62.9M/1.98G [00:05<02:40, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 4%|▏ | 73.4M/1.98G [00:06<02:40, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 4%|▏ | 83.9M/1.98G [00:07<02:38, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 5%|▏ | 94.4M/1.98G [00:07<02:40, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 5%|▎ | 105M/1.98G [00:08<02:37, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 6%|▎ | 115M/1.98G [00:10<03:05, 10.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 6%|▎ | 126M/1.98G [00:11<02:55, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 7%|▎ | 136M/1.98G [00:11<02:49, 10.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 7%|▎ | 147M/1.98G [00:12<02:42, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 8%|▍ | 157M/1.98G [00:13<02:39, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 8%|▍ | 168M/1.98G [00:14<02:39, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 9%|▍ | 178M/1.98G [00:15<02:35, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 10%|▍ | 189M/1.98G [00:16<02:33, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 10%|▌ | 199M/1.98G [00:17<02:32, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 11%|▌ | 210M/1.98G [00:18<02:29, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 11%|▌ | 220M/1.98G [00:19<02:28, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 12%|▌ | 231M/1.98G [00:19<02:27, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 12%|▌ | 241M/1.98G [00:20<02:26, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 13%|▋ | 252M/1.98G [00:21<02:25, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 13%|▋ | 262M/1.98G [00:22<02:23, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 14%|▋ | 273M/1.98G [00:23<02:25, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 14%|▋ | 283M/1.98G [00:24<02:23, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 15%|▋ | 294M/1.98G [00:25<02:46, 10.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 15%|▊ | 304M/1.98G [00:26<02:36, 10.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 16%|▊ | 315M/1.98G [00:27<02:31, 11.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 16%|▊ | 325M/1.98G [00:28<02:27, 11.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 17%|▊ | 336M/1.98G [00:29<02:24, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 17%|▊ | 346M/1.98G [00:30<02:23, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 18%|▉ | 357M/1.98G [00:31<02:19, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 19%|▉ | 367M/1.98G [00:31<02:16, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 19%|▉ | 377M/1.98G [00:35<04:06, 6.50MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 20%|▉ | 388M/1.98G [00:36<03:30, 7.55MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 20%|█ | 398M/1.98G [00:36<03:06, 8.47MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 21%|█ | 409M/1.98G [00:37<02:49, 9.26MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 21%|█ | 419M/1.98G [00:38<02:38, 9.87MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 22%|█ | 430M/1.98G [00:39<02:29, 10.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 22%|█ | 440M/1.98G [00:40<02:23, 10.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 23%|█▏ | 451M/1.98G [00:41<02:17, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 23%|█▏ | 461M/1.98G [00:42<02:15, 11.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 24%|█▏ | 472M/1.98G [00:43<02:31, 9.93MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 24%|█▏ | 482M/1.98G [00:44<02:23, 10.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 25%|█▏ | 493M/1.98G [00:45<02:19, 10.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 25%|█▎ | 503M/1.98G [00:46<02:14, 11.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 26%|█▎ | 514M/1.98G [00:47<02:10, 11.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 26%|█▎ | 524M/1.98G [00:48<02:10, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 27%|█▎ | 535M/1.98G [00:49<02:06, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 28%|█▍ | 545M/1.98G [00:49<02:02, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 28%|█▍ | 556M/1.98G [00:50<02:00, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 29%|█▍ | 566M/1.98G [00:51<01:59, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 29%|█▍ | 577M/1.98G [00:52<01:58, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 30%|█▍ | 587M/1.98G [00:53<01:58, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 30%|█▌ | 598M/1.98G [00:54<01:58, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 31%|█▌ | 608M/1.98G [00:55<01:56, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 31%|█▌ | 619M/1.98G [00:56<01:54, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 32%|█▌ | 629M/1.98G [00:57<01:54, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 32%|█▌ | 640M/1.98G [00:57<01:52, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 33%|█▋ | 650M/1.98G [00:58<01:51, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 33%|█▋ | 661M/1.98G [00:59<01:54, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 34%|█▋ | 671M/1.98G [01:00<01:51, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 34%|█▋ | 682M/1.98G [01:01<01:49, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 35%|█▋ | 692M/1.98G [01:02<02:07, 10.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 35%|█▊ | 703M/1.98G [01:03<02:01, 10.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 36%|█▊ | 713M/1.98G [01:04<01:56, 10.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 37%|█▊ | 724M/1.98G [01:05<01:52, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 37%|█▊ | 734M/1.98G [01:06<01:52, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 38%|█▉ | 744M/1.98G [01:07<01:48, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 38%|█▉ | 755M/1.98G [01:08<01:45, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 39%|█▉ | 765M/1.98G [01:09<01:44, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 39%|█▉ | 776M/1.98G [01:09<01:41, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 40%|█▉ | 786M/1.98G [01:10<01:41, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 40%|██ | 797M/1.98G [01:11<01:43, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 41%|██ | 807M/1.98G [01:12<01:41, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 41%|██ | 818M/1.98G [01:13<01:38, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 42%|██ | 828M/1.98G [01:14<01:37, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 42%|██ | 839M/1.98G [01:15<01:37, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 43%|██▏ | 849M/1.98G [01:16<01:35, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 43%|██▏ | 860M/1.98G [01:17<01:33, 12.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 44%|██▏ | 870M/1.98G [01:18<01:34, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 44%|██▏ | 881M/1.98G [01:23<03:56, 4.65MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 45%|██▎ | 891M/1.98G [01:24<03:11, 5.68MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 46%|██▎ | 902M/1.98G [01:25<02:39, 6.77MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 46%|██▎ | 912M/1.98G [01:26<02:17, 7.76MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 47%|██▎ | 923M/1.98G [01:26<02:01, 8.69MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 47%|██▎ | 933M/1.98G [01:27<01:50, 9.47MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 48%|██▍ | 944M/1.98G [01:28<01:42, 10.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 48%|██▍ | 954M/1.98G [01:29<01:36, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 49%|██▍ | 965M/1.98G [01:30<01:32, 11.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 49%|██▍ | 975M/1.98G [01:31<01:28, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 50%|██▍ | 986M/1.98G [01:32<01:27, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 50%|██▌ | 996M/1.98G [01:33<01:38, 9.98MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 51%|██ | 1.01G/1.98G [01:36<02:38, 6.15MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 51%|██ | 1.02G/1.98G [01:37<02:21, 6.80MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 52%|██ | 1.03G/1.98G [01:38<02:01, 7.82MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 52%|██ | 1.04G/1.98G [01:39<01:48, 8.70MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 53%|██ | 1.05G/1.98G [01:40<01:37, 9.56MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 53%|██▏ | 1.06G/1.98G [01:41<01:30, 10.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 54%|██▏ | 1.07G/1.98G [01:42<01:25, 10.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 55%|██▏ | 1.08G/1.98G [01:43<01:21, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 55%|██▏ | 1.09G/1.98G [01:44<01:18, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 56%|██▏ | 1.10G/1.98G [01:44<01:15, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 56%|██▏ | 1.11G/1.98G [01:46<01:27, 9.87MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 57%|██▎ | 1.12G/1.98G [01:47<01:33, 9.14MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 57%|██▎ | 1.13G/1.98G [01:48<01:25, 9.90MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 58%|██▎ | 1.14G/1.98G [01:49<01:20, 10.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 58%|██▎ | 1.15G/1.98G [01:50<01:17, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 59%|██▎ | 1.16G/1.98G [01:51<01:14, 11.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 59%|██▎ | 1.17G/1.98G [01:52<01:11, 11.2MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 60%|██▍ | 1.18G/1.98G [01:53<01:10, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 60%|██▍ | 1.20G/1.98G [01:53<01:08, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 61%|██▍ | 1.21G/1.98G [01:54<01:06, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 61%|██▍ | 1.22G/1.98G [01:55<01:05, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 62%|██▍ | 1.23G/1.98G [01:56<01:04, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 62%|██▍ | 1.24G/1.98G [01:57<01:02, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 63%|██▌ | 1.25G/1.98G [01:58<01:03, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 64%|██▌ | 1.26G/1.98G [01:59<01:11, 10.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 64%|██▌ | 1.27G/1.98G [02:00<01:07, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 65%|██▌ | 1.28G/1.98G [02:01<01:04, 10.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 65%|██▌ | 1.29G/1.98G [02:02<01:02, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 66%|██▋ | 1.30G/1.98G [02:03<01:00, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 66%|██▋ | 1.31G/1.98G [02:04<00:58, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 67%|██▋ | 1.32G/1.98G [02:05<00:57, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 67%|██▋ | 1.33G/1.98G [02:06<00:56, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 68%|██▋ | 1.34G/1.98G [02:06<00:55, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 68%|██▋ | 1.35G/1.98G [02:07<00:53, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 69%|██▊ | 1.36G/1.98G [02:08<00:52, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 69%|██▊ | 1.37G/1.98G [02:09<00:50, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 70%|██▊ | 1.38G/1.98G [02:10<00:51, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 70%|██▊ | 1.39G/1.98G [02:11<00:50, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 71%|██▊ | 1.41G/1.98G [02:12<00:49, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 72%|██▊ | 1.42G/1.98G [02:13<00:48, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 72%|██▉ | 1.43G/1.98G [02:14<00:47, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 73%|██▉ | 1.44G/1.98G [02:14<00:45, 11.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 73%|██▉ | 1.45G/1.98G [02:16<00:52, 10.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 74%|██▉ | 1.46G/1.98G [02:17<00:49, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 74%|██▉ | 1.47G/1.98G [02:18<00:46, 10.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 75%|██▉ | 1.48G/1.98G [02:19<00:45, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 75%|███ | 1.49G/1.98G [02:20<00:55, 8.84MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 76%|███ | 1.50G/1.98G [02:21<00:49, 9.62MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 76%|███ | 1.51G/1.98G [02:22<00:46, 10.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 77%|███ | 1.52G/1.98G [02:26<01:18, 5.83MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 77%|███ | 1.53G/1.98G [02:27<01:06, 6.72MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 78%|███ | 1.54G/1.98G [02:27<00:56, 7.76MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 78%|███▏| 1.55G/1.98G [02:28<00:49, 8.67MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 79%|███▏| 1.56G/1.98G [02:30<00:50, 8.20MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 79%|███▏| 1.57G/1.98G [02:31<00:45, 9.04MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 80%|███▏| 1.58G/1.98G [02:32<00:40, 9.81MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 81%|███▏| 1.59G/1.98G [02:32<00:37, 10.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 81%|███▏| 1.60G/1.98G [02:33<00:35, 10.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 82%|███▎| 1.61G/1.98G [02:34<00:32, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 82%|███▎| 1.63G/1.98G [02:38<00:56, 6.26MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 83%|███▎| 1.64G/1.98G [02:39<00:47, 7.17MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 83%|███▎| 1.65G/1.98G [02:39<00:41, 8.12MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 84%|███▎| 1.66G/1.98G [02:40<00:35, 9.01MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 84%|███▎| 1.67G/1.98G [02:41<00:32, 9.69MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 85%|███▍| 1.68G/1.98G [02:42<00:29, 10.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 85%|███▍| 1.69G/1.98G [02:43<00:27, 10.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 86%|███▍| 1.70G/1.98G [02:44<00:25, 11.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 86%|███▍| 1.71G/1.98G [02:45<00:23, 11.3MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 87%|███▍| 1.72G/1.98G [02:46<00:22, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 87%|███▍| 1.73G/1.98G [02:46<00:21, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 88%|███▌| 1.74G/1.98G [02:47<00:20, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 88%|███▌| 1.75G/1.98G [02:48<00:19, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 89%|███▌| 1.76G/1.98G [02:49<00:18, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 90%|███▌| 1.77G/1.98G [02:50<00:17, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 90%|███▌| 1.78G/1.98G [02:51<00:17, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 91%|███▌| 1.79G/1.98G [02:52<00:15, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 91%|███▋| 1.80G/1.98G [02:53<00:15, 11.6MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 92%|███▋| 1.81G/1.98G [02:54<00:14, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 92%|███▋| 1.82G/1.98G [02:55<00:13, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 93%|███▋| 1.84G/1.98G [02:55<00:12, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 93%|███▋| 1.85G/1.98G [02:58<00:16, 8.31MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 94%|███▋| 1.86G/1.98G [02:59<00:14, 8.62MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 94%|███▊| 1.87G/1.98G [03:00<00:12, 9.38MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 95%|███▊| 1.88G/1.98G [03:00<00:10, 10.0MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 95%|███▊| 1.89G/1.98G [03:01<00:08, 10.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 96%|███▊| 1.90G/1.98G [03:02<00:07, 10.9MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 96%|███▊| 1.91G/1.98G [03:03<00:06, 11.1MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 97%|███▉| 1.92G/1.98G [03:04<00:05, 11.4MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 97%|███▉| 1.93G/1.98G [03:05<00:04, 11.5MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 98%|███▉| 1.94G/1.98G [03:06<00:03, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 99%|███▉| 1.95G/1.98G [03:07<00:02, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 99%|███▉| 1.96G/1.98G [03:07<00:01, 11.7MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 100%|███▉| 1.97G/1.98G [03:08<00:00, 11.8MB/s]\u001b[A\n","model-00007-of-00008.safetensors: 100%|████| 1.98G/1.98G [03:09<00:00, 10.4MB/s]\u001b[A\n","Downloading shards: 88%|█████████████████████ | 7/8 [20:56<03:02, 182.23s/it]\n","model-00008-of-00008.safetensors: 0%| | 0.00/1.75G [00:00, ?B/s]\u001b[A\n","model-00008-of-00008.safetensors: 1%| | 10.5M/1.75G [00:00<02:24, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 1%| | 21.0M/1.75G [00:01<02:25, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 2%| | 31.5M/1.75G [00:02<02:25, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 2%| | 41.9M/1.75G [00:03<02:26, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 3%| | 52.4M/1.75G [00:04<02:22, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 4%|▏ | 62.9M/1.75G [00:05<02:23, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 4%|▏ | 73.4M/1.75G [00:06<02:22, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 5%|▏ | 83.9M/1.75G [00:07<02:21, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 5%|▏ | 94.4M/1.75G [00:08<02:22, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 6%|▎ | 105M/1.75G [00:08<02:22, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 7%|▎ | 115M/1.75G [00:10<02:43, 9.96MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 7%|▎ | 126M/1.75G [00:11<02:36, 10.4MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 8%|▍ | 136M/1.75G [00:12<02:28, 10.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 8%|▍ | 147M/1.75G [00:13<02:25, 11.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 9%|▍ | 157M/1.75G [00:13<02:23, 11.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 10%|▍ | 168M/1.75G [00:14<02:19, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 10%|▌ | 178M/1.75G [00:15<02:15, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 11%|▌ | 189M/1.75G [00:16<02:15, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 11%|▌ | 199M/1.75G [00:17<02:13, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 12%|▌ | 210M/1.75G [00:18<02:10, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 13%|▋ | 220M/1.75G [00:19<02:08, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 13%|▋ | 231M/1.75G [00:20<02:10, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 14%|▋ | 241M/1.75G [00:21<02:08, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 14%|▋ | 252M/1.75G [00:21<02:06, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 15%|▋ | 262M/1.75G [00:22<02:07, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 16%|▊ | 273M/1.75G [00:23<02:05, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 16%|▊ | 283M/1.75G [00:24<02:04, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 17%|▊ | 294M/1.75G [00:25<02:02, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 17%|▊ | 304M/1.75G [00:28<03:15, 7.39MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 18%|▉ | 315M/1.75G [00:29<02:52, 8.30MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 19%|▉ | 325M/1.75G [00:29<02:35, 9.12MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 19%|▉ | 336M/1.75G [00:30<02:24, 9.79MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 20%|▉ | 346M/1.75G [00:31<02:14, 10.4MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 20%|█ | 357M/1.75G [00:32<02:08, 10.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 21%|█ | 367M/1.75G [00:33<02:06, 10.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 22%|█ | 377M/1.75G [00:34<02:02, 11.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 22%|█ | 388M/1.75G [00:35<01:59, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 23%|█▏ | 398M/1.75G [00:36<01:57, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 23%|█▏ | 409M/1.75G [00:37<01:56, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 24%|█▏ | 419M/1.75G [00:37<01:54, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 25%|█▏ | 430M/1.75G [00:39<02:08, 10.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 25%|█▎ | 440M/1.75G [00:40<02:02, 10.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 26%|█▎ | 451M/1.75G [00:41<01:59, 10.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 26%|█▎ | 461M/1.75G [00:41<01:55, 11.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 27%|█▎ | 472M/1.75G [00:42<01:52, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 28%|█▍ | 482M/1.75G [00:43<01:52, 11.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 28%|█▍ | 493M/1.75G [00:44<01:50, 11.4MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 29%|█▍ | 503M/1.75G [00:45<01:48, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 29%|█▍ | 514M/1.75G [00:46<01:47, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 30%|█▍ | 524M/1.75G [00:47<01:44, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 31%|█▌ | 535M/1.75G [00:48<01:42, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 31%|█▌ | 545M/1.75G [00:49<01:43, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 32%|█▌ | 556M/1.75G [00:50<01:41, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 32%|█▌ | 566M/1.75G [00:50<01:41, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 33%|█▋ | 577M/1.75G [00:51<01:39, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 34%|█▋ | 587M/1.75G [00:52<01:37, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 34%|█▋ | 598M/1.75G [00:53<01:36, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 35%|█▋ | 608M/1.75G [00:54<01:35, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 35%|█▊ | 619M/1.75G [00:55<01:33, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 36%|█▊ | 629M/1.75G [00:56<01:32, 12.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 37%|█▊ | 640M/1.75G [00:57<01:32, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 37%|█▊ | 650M/1.75G [00:57<01:32, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 38%|█▉ | 661M/1.75G [00:58<01:30, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 38%|█▉ | 671M/1.75G [01:00<01:45, 10.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 39%|█▉ | 682M/1.75G [01:01<01:39, 10.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 40%|█▉ | 692M/1.75G [01:01<01:36, 11.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 40%|██ | 703M/1.75G [01:02<01:32, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 41%|██ | 713M/1.75G [01:03<01:29, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 41%|██ | 724M/1.75G [01:04<01:27, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 42%|██ | 734M/1.75G [01:05<01:26, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 43%|██▏ | 744M/1.75G [01:06<01:24, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 43%|██▏ | 755M/1.75G [01:07<01:23, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 44%|██▏ | 765M/1.75G [01:08<01:22, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 44%|██▏ | 776M/1.75G [01:12<03:09, 5.14MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 45%|██▏ | 786M/1.75G [01:13<02:42, 5.91MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 46%|██▎ | 797M/1.75G [01:14<02:16, 6.96MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 46%|██▎ | 807M/1.75G [01:15<02:00, 7.83MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 47%|██▎ | 818M/1.75G [01:16<01:46, 8.74MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 47%|██▎ | 828M/1.75G [01:17<01:36, 9.53MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 48%|██▍ | 839M/1.75G [01:18<01:42, 8.84MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 49%|██▍ | 849M/1.75G [01:19<01:33, 9.62MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 49%|██▍ | 860M/1.75G [01:20<01:27, 10.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 50%|██▍ | 870M/1.75G [01:21<01:22, 10.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 50%|██▌ | 881M/1.75G [01:22<01:18, 11.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 51%|██▌ | 891M/1.75G [01:23<01:16, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 52%|██▌ | 902M/1.75G [01:24<01:14, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 52%|██▌ | 912M/1.75G [01:25<01:12, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 53%|██▋ | 923M/1.75G [01:25<01:10, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 53%|██▋ | 933M/1.75G [01:26<01:09, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 54%|██▋ | 944M/1.75G [01:28<01:25, 9.40MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 55%|██▋ | 954M/1.75G [01:29<01:19, 10.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 55%|██▊ | 965M/1.75G [01:30<01:14, 10.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 56%|██▊ | 975M/1.75G [01:31<01:10, 11.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 56%|██▊ | 986M/1.75G [01:31<01:08, 11.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 57%|██▊ | 996M/1.75G [01:32<01:05, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 58%|██▎ | 1.01G/1.75G [01:33<01:04, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 58%|██▎ | 1.02G/1.75G [01:34<01:02, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 59%|██▎ | 1.03G/1.75G [01:35<01:11, 10.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 59%|██▍ | 1.04G/1.75G [01:36<01:07, 10.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 60%|██▍ | 1.05G/1.75G [01:37<01:03, 10.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 61%|██▍ | 1.06G/1.75G [01:38<01:01, 11.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 61%|██▍ | 1.07G/1.75G [01:39<00:59, 11.4MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 62%|██▍ | 1.08G/1.75G [01:40<00:57, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 62%|██▍ | 1.09G/1.75G [01:41<00:56, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 63%|██▌ | 1.10G/1.75G [01:42<00:55, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 64%|██▌ | 1.11G/1.75G [01:43<00:54, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 64%|██▌ | 1.12G/1.75G [01:43<00:52, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 65%|██▌ | 1.13G/1.75G [01:44<00:52, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 65%|██▌ | 1.14G/1.75G [01:45<00:50, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 66%|██▋ | 1.15G/1.75G [01:46<00:50, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 67%|██▋ | 1.16G/1.75G [01:47<00:49, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 67%|██▋ | 1.17G/1.75G [01:48<00:48, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 68%|██▋ | 1.18G/1.75G [01:49<00:46, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 68%|██▋ | 1.20G/1.75G [01:50<00:46, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 69%|██▊ | 1.21G/1.75G [01:50<00:44, 12.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 70%|██▊ | 1.22G/1.75G [01:51<00:44, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 70%|██▊ | 1.23G/1.75G [01:52<00:43, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 71%|██▊ | 1.24G/1.75G [01:54<00:49, 10.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 71%|██▊ | 1.25G/1.75G [01:54<00:46, 10.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 72%|██▉ | 1.26G/1.75G [01:55<00:44, 11.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 73%|██▉ | 1.27G/1.75G [01:56<00:42, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 73%|██▉ | 1.28G/1.75G [01:57<00:40, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 74%|██▉ | 1.29G/1.75G [01:58<00:39, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 74%|██▉ | 1.30G/1.75G [01:59<00:38, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 75%|██▉ | 1.31G/1.75G [02:00<00:36, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 76%|███ | 1.32G/1.75G [02:01<00:36, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 76%|███ | 1.33G/1.75G [02:01<00:35, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 77%|███ | 1.34G/1.75G [02:02<00:34, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 77%|███ | 1.35G/1.75G [02:04<00:38, 10.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 78%|███ | 1.36G/1.75G [02:05<00:36, 10.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 79%|███▏| 1.37G/1.75G [02:05<00:33, 11.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 79%|███▏| 1.38G/1.75G [02:06<00:32, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 80%|███▏| 1.39G/1.75G [02:07<00:30, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 80%|███▏| 1.41G/1.75G [02:08<00:29, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 81%|███▏| 1.42G/1.75G [02:09<00:28, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 82%|███▎| 1.43G/1.75G [02:10<00:27, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 82%|███▎| 1.44G/1.75G [02:11<00:26, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 83%|███▎| 1.45G/1.75G [02:12<00:25, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 83%|███▎| 1.46G/1.75G [02:12<00:24, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 84%|███▎| 1.47G/1.75G [02:13<00:23, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 85%|███▍| 1.48G/1.75G [02:14<00:22, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 85%|███▍| 1.49G/1.75G [02:15<00:21, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 86%|███▍| 1.50G/1.75G [02:16<00:20, 12.1MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 86%|███▍| 1.51G/1.75G [02:17<00:19, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 87%|███▍| 1.52G/1.75G [02:18<00:22, 10.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 88%|███▌| 1.53G/1.75G [02:19<00:20, 10.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 88%|███▌| 1.54G/1.75G [02:20<00:18, 11.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 89%|███▌| 1.55G/1.75G [02:21<00:17, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 89%|███▌| 1.56G/1.75G [02:22<00:16, 11.4MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 90%|███▌| 1.57G/1.75G [02:23<00:15, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 91%|███▌| 1.58G/1.75G [02:23<00:14, 11.7MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 91%|███▋| 1.59G/1.75G [02:24<00:13, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 92%|███▋| 1.60G/1.75G [02:25<00:12, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 92%|███▋| 1.61G/1.75G [02:26<00:11, 11.9MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 93%|███▋| 1.63G/1.75G [02:27<00:10, 12.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 94%|███▋| 1.64G/1.75G [02:29<00:14, 7.99MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 94%|███▊| 1.65G/1.75G [02:31<00:12, 8.00MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 95%|███▊| 1.66G/1.75G [02:31<00:10, 8.88MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 95%|███▊| 1.67G/1.75G [02:32<00:08, 9.65MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 96%|███▊| 1.68G/1.75G [02:33<00:06, 10.2MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 97%|███▊| 1.69G/1.75G [02:34<00:05, 10.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 97%|███▉| 1.70G/1.75G [02:35<00:04, 11.0MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 98%|███▉| 1.71G/1.75G [02:36<00:03, 11.3MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 98%|███▉| 1.72G/1.75G [02:37<00:02, 11.5MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 99%|███▉| 1.73G/1.75G [02:38<00:01, 11.6MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 100%|███▉| 1.74G/1.75G [02:38<00:00, 11.8MB/s]\u001b[A\n","model-00008-of-00008.safetensors: 100%|████| 1.75G/1.75G [02:39<00:00, 11.0MB/s]\u001b[A\n","Downloading shards: 100%|████████████████████████| 8/8 [23:36<00:00, 177.02s/it]\n","[INFO|modeling_utils.py:1519] 2024-07-22 21:34:35,272 >> Instantiating InternLM2ForCausalLM model under default dtype torch.bfloat16.\n","[INFO|configuration_utils.py:962] 2024-07-22 21:34:35,272 >> Generate config GenerationConfig {\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": 2,\n"," \"pad_token_id\": 2\n","}\n","\n","Loading checkpoint shards: 100%|██████████████████| 8/8 [00:02<00:00, 3.11it/s]\n","[INFO|modeling_utils.py:4280] 2024-07-22 21:34:41,963 >> All model checkpoint weights were used when initializing InternLM2ForCausalLM.\n","\n","[INFO|modeling_utils.py:4288] 2024-07-22 21:34:41,963 >> All the weights of InternLM2ForCausalLM were initialized from the model checkpoint at internlm/internlm2_5-7b-chat-1m.\n","If your task is similar to the task the model of the checkpoint was trained on, you can already use InternLM2ForCausalLM for predictions without further training.\n","generation_config.json: 100%|██████████████████| 123/123 [00:00<00:00, 1.97MB/s]\n","[INFO|configuration_utils.py:917] 2024-07-22 21:34:42,702 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/generation_config.json\n","[INFO|configuration_utils.py:962] 2024-07-22 21:34:42,702 >> Generate config GenerationConfig {\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": [\n"," 2,\n"," 92542\n"," ],\n"," \"pad_token_id\": 2\n","}\n","\n","07/22/2024 21:34:42 - INFO - llamafactory.model.model_utils.attention - Using vanilla attention implementation.\n","adapter_config.json: 100%|█████████████████████| 679/679 [00:00<00:00, 9.82MB/s]\n","adapter_model.safetensors: 100%|███████████| 37.8M/37.8M [00:03<00:00, 11.8MB/s]\n","07/22/2024 21:34:48 - INFO - llamafactory.model.adapter - Merged 1 adapter(s).\n","07/22/2024 21:34:48 - INFO - llamafactory.model.adapter - Loaded adapter(s): inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","07/22/2024 21:34:48 - INFO - llamafactory.model.loader - all params: 7,737,708,544\n","(2) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","14.818 GB of memory reserved.\n","loading train/test data files\n","Generating train split: 25000 examples [00:00, 110406.75 examples/s]\n","Generating test split: 3000 examples [00:00, 201720.35 examples/s]\n","Map: 100%|██████████████████████| 25000/25000 [00:00<00:00, 48814.27 examples/s]\n","Map: 100%|████████████████████████| 3000/3000 [00:00<00:00, 40855.20 examples/s]\n","DatasetDict({\n"," train: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 25000\n"," })\n"," test: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 3000\n"," })\n","})\n","--------------------------------------------------\n","text: 甄加索是自杀吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 海岸之谜\n","--------------------------------------------------\n","puzzle: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","--------------------------------------------------\n","truth: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","--------------------------------------------------\n","text: 死者受伤了吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 甄庄哭声\n","--------------------------------------------------\n","puzzle: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","--------------------------------------------------\n","truth: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","Evaluating model: internlm/internlm2_5-7b-chat-1m\n"," 0%| | 0/3000 [00:00, ?it/s]Batch output: ['不是']\n","100%|███████████████████████████████████████| 3000/3000 [07:36<00:00, 6.57it/s]\n","(3) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","16.713 GB of memory reserved.\n"," text ... internlm/internlm2_5-7b-chat-1m/InternLM_2_5-7b_LoRA-Adapter_torch.float16_lf\n","0 甄加索是自杀吗 ... 不是 \n","\n","[1 rows x 11 columns]\n","{'accuracy': 0.812, 'incorrect_ids': [6, 9, 13, 18, 29, 31, 36, 55, 58, 59, 61, 65, 66, 67, 78, 81, 83, 84, 88, 91, 97, 103, 104, 106, 110, 112, 121, 128, 129, 135, 137, 139, 143, 149, 150, 153, 155, 161, 164, 170, 173, 190, 199, 200, 202, 224, 227, 228, 245, 250, 259, 260, 269, 271, 275, 276, 279, 284, 286, 293, 299, 304, 311, 314, 318, 323, 326, 328, 330, 334, 335, 337, 341, 346, 350, 353, 355, 356, 360, 361, 368, 370, 373, 375, 377, 383, 384, 389, 397, 410, 428, 429, 430, 432, 438, 445, 447, 450, 452, 453, 454, 456, 457, 458, 465, 469, 470, 471, 472, 476, 479, 481, 486, 495, 497, 501, 502, 503, 506, 507, 510, 511, 517, 518, 519, 520, 536, 540, 560, 561, 566, 570, 571, 579, 581, 589, 591, 593, 597, 601, 613, 614, 621, 622, 625, 628, 632, 636, 644, 647, 666, 671, 682, 690, 695, 702, 708, 721, 727, 729, 730, 734, 740, 754, 758, 770, 773, 774, 778, 788, 791, 795, 798, 801, 805, 809, 817, 819, 820, 821, 823, 824, 828, 837, 840, 847, 856, 861, 866, 869, 870, 875, 886, 889, 890, 894, 899, 904, 927, 935, 937, 942, 945, 952, 962, 964, 966, 969, 980, 982, 989, 991, 994, 998, 1001, 1003, 1006, 1011, 1012, 1014, 1015, 1019, 1022, 1036, 1038, 1040, 1043, 1051, 1053, 1076, 1077, 1080, 1087, 1089, 1096, 1120, 1125, 1126, 1129, 1158, 1161, 1166, 1172, 1174, 1178, 1180, 1181, 1185, 1198, 1203, 1212, 1228, 1232, 1239, 1240, 1241, 1251, 1252, 1254, 1258, 1259, 1266, 1276, 1282, 1289, 1298, 1305, 1308, 1311, 1313, 1315, 1317, 1324, 1331, 1339, 1342, 1349, 1353, 1357, 1363, 1364, 1367, 1385, 1387, 1388, 1406, 1409, 1417, 1420, 1422, 1426, 1440, 1451, 1453, 1454, 1459, 1462, 1469, 1476, 1478, 1486, 1490, 1493, 1494, 1496, 1501, 1504, 1512, 1515, 1517, 1518, 1525, 1526, 1544, 1547, 1554, 1560, 1562, 1580, 1581, 1585, 1590, 1593, 1594, 1603, 1604, 1605, 1613, 1622, 1627, 1633, 1636, 1637, 1641, 1645, 1648, 1650, 1654, 1655, 1659, 1662, 1665, 1672, 1673, 1674, 1679, 1686, 1690, 1691, 1695, 1712, 1716, 1718, 1726, 1727, 1734, 1736, 1751, 1755, 1756, 1769, 1780, 1786, 1791, 1796, 1809, 1812, 1827, 1835, 1841, 1858, 1860, 1867, 1869, 1879, 1880, 1888, 1897, 1907, 1914, 1915, 1945, 1953, 1958, 1963, 1964, 1973, 1978, 1981, 1984, 1986, 1989, 1990, 1995, 2017, 2021, 2028, 2029, 2035, 2036, 2049, 2054, 2064, 2067, 2070, 2072, 2076, 2077, 2106, 2109, 2110, 2112, 2119, 2120, 2121, 2126, 2133, 2147, 2159, 2161, 2162, 2164, 2169, 2172, 2174, 2177, 2181, 2185, 2186, 2193, 2194, 2195, 2196, 2197, 2199, 2212, 2223, 2226, 2237, 2240, 2243, 2262, 2274, 2278, 2281, 2285, 2297, 2304, 2311, 2317, 2320, 2324, 2330, 2333, 2339, 2344, 2348, 2354, 2360, 2364, 2366, 2367, 2378, 2389, 2395, 2396, 2400, 2404, 2406, 2409, 2410, 2423, 2429, 2437, 2440, 2442, 2463, 2488, 2512, 2522, 2529, 2530, 2532, 2535, 2538, 2546, 2548, 2549, 2553, 2556, 2559, 2574, 2575, 2577, 2581, 2589, 2590, 2616, 2629, 2632, 2644, 2663, 2667, 2676, 2678, 2714, 2716, 2731, 2736, 2742, 2744, 2745, 2746, 2749, 2754, 2757, 2758, 2762, 2766, 2777, 2788, 2791, 2794, 2797, 2798, 2801, 2806, 2807, 2815, 2816, 2820, 2823, 2837, 2843, 2854, 2857, 2860, 2861, 2867, 2877, 2880, 2882, 2884, 2899, 2905, 2912, 2913, 2916, 2921, 2933, 2937, 2938, 2942, 2944, 2949, 2953, 2955, 2963, 2968, 2977, 2979, 2985, 2995, 2998]}\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter False datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:42:34,236 >> loading file ./tokenizer.model from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/./tokenizer.model\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:42:34,236 >> loading file added_tokens.json from cache at None\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:42:34,236 >> loading file special_tokens_map.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/special_tokens_map.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:42:34,236 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/tokenizer_config.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:42:34,236 >> loading file tokenizer.json from cache at None\n","07/22/2024 21:42:34 - INFO - llamafactory.data.template - Add <|im_end|> to stop words.\n","/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n"," warnings.warn(\n","[INFO|configuration_utils.py:733] 2024-07-22 21:42:34,806 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","[INFO|configuration_utils.py:733] 2024-07-22 21:42:35,328 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","[INFO|configuration_utils.py:796] 2024-07-22 21:42:35,328 >> Model config InternLM2Config {\n"," \"_name_or_path\": \"internlm/internlm2_5-7b-chat-1m\",\n"," \"architectures\": [\n"," \"InternLM2ForCausalLM\"\n"," ],\n"," \"attn_implementation\": \"eager\",\n"," \"auto_map\": {\n"," \"AutoConfig\": \"internlm/internlm2_5-7b-chat-1m--configuration_internlm2.InternLM2Config\",\n"," \"AutoModel\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\",\n"," \"AutoModelForCausalLM\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\"\n"," },\n"," \"bias\": false,\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": 2,\n"," \"hidden_act\": \"silu\",\n"," \"hidden_size\": 4096,\n"," \"initializer_range\": 0.02,\n"," \"intermediate_size\": 14336,\n"," \"max_position_embeddings\": 262144,\n"," \"model_type\": \"internlm2\",\n"," \"num_attention_heads\": 32,\n"," \"num_hidden_layers\": 32,\n"," \"num_key_value_heads\": 8,\n"," \"pad_token_id\": 2,\n"," \"pretraining_tp\": 1,\n"," \"rms_norm_eps\": 1e-05,\n"," \"rope_scaling\": {\n"," \"factor\": 2.5,\n"," \"type\": \"dynamic\"\n"," },\n"," \"rope_theta\": 50000000,\n"," \"tie_word_embeddings\": false,\n"," \"torch_dtype\": \"bfloat16\",\n"," \"transformers_version\": \"4.41.2\",\n"," \"use_cache\": true,\n"," \"vocab_size\": 92544\n","}\n","\n","07/22/2024 21:42:35 - INFO - llamafactory.model.patcher - Using KV cache for faster generation.\n","[INFO|modeling_utils.py:3474] 2024-07-22 21:42:35,599 >> loading weights file model.safetensors from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/model.safetensors.index.json\n","[INFO|modeling_utils.py:1519] 2024-07-22 21:42:35,600 >> Instantiating InternLM2ForCausalLM model under default dtype torch.bfloat16.\n","[INFO|configuration_utils.py:962] 2024-07-22 21:42:35,600 >> Generate config GenerationConfig {\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": 2,\n"," \"pad_token_id\": 2\n","}\n","\n","Loading checkpoint shards: 100%|██████████████████| 8/8 [00:02<00:00, 2.69it/s]\n","[INFO|modeling_utils.py:4280] 2024-07-22 21:42:38,783 >> All model checkpoint weights were used when initializing InternLM2ForCausalLM.\n","\n","[INFO|modeling_utils.py:4288] 2024-07-22 21:42:38,783 >> All the weights of InternLM2ForCausalLM were initialized from the model checkpoint at internlm/internlm2_5-7b-chat-1m.\n","If your task is similar to the task the model of the checkpoint was trained on, you can already use InternLM2ForCausalLM for predictions without further training.\n","[INFO|configuration_utils.py:917] 2024-07-22 21:42:39,069 >> loading configuration file generation_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/generation_config.json\n","[INFO|configuration_utils.py:962] 2024-07-22 21:42:39,069 >> Generate config GenerationConfig {\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": [\n"," 2,\n"," 92542\n"," ],\n"," \"pad_token_id\": 2\n","}\n","\n","07/22/2024 21:42:39 - INFO - llamafactory.model.model_utils.attention - Using vanilla attention implementation.\n","07/22/2024 21:42:40 - INFO - llamafactory.model.adapter - Merged 1 adapter(s).\n","07/22/2024 21:42:40 - INFO - llamafactory.model.adapter - Loaded adapter(s): inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","07/22/2024 21:42:40 - INFO - llamafactory.model.loader - all params: 7,737,708,544\n","(2) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","14.818 GB of memory reserved.\n","loading train/test data files\n","DatasetDict({\n"," train: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 25000\n"," })\n"," test: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 3000\n"," })\n","})\n","--------------------------------------------------\n","text: 甄加索是自杀吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 海岸之谜\n","--------------------------------------------------\n","puzzle: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","--------------------------------------------------\n","truth: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","--------------------------------------------------\n","text: 死者受伤了吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 甄庄哭声\n","--------------------------------------------------\n","puzzle: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","--------------------------------------------------\n","truth: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","Evaluating model: internlm/internlm2_5-7b-chat-1m\n"," 0%| | 0/3000 [00:00, ?it/s]Batch output: ['不是']\n","100%|███████████████████████████████████████| 3000/3000 [07:35<00:00, 6.58it/s]\n","(3) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","16.713 GB of memory reserved.\n"," text ... internlm/internlm2_5-7b-chat-1m/InternLM_2_5-7b_LoRA-Adapter_torch.bfloat16_lf\n","0 甄加索是自杀吗 ... 不是 \n","\n","[1 rows x 12 columns]\n","{'accuracy': 0.812, 'incorrect_ids': [6, 9, 13, 18, 29, 31, 36, 55, 58, 59, 61, 65, 66, 67, 78, 81, 83, 84, 88, 91, 97, 103, 104, 106, 110, 112, 121, 128, 129, 135, 137, 139, 143, 149, 150, 153, 155, 161, 164, 170, 173, 190, 199, 200, 202, 224, 227, 228, 245, 250, 259, 260, 269, 271, 275, 276, 279, 284, 286, 293, 299, 304, 311, 314, 318, 323, 326, 328, 330, 334, 335, 337, 341, 346, 350, 353, 355, 356, 360, 361, 368, 370, 373, 375, 377, 383, 384, 389, 397, 410, 428, 429, 430, 432, 438, 445, 447, 450, 452, 453, 454, 456, 457, 458, 465, 469, 470, 471, 472, 476, 479, 481, 486, 495, 497, 501, 502, 503, 506, 507, 510, 511, 517, 518, 519, 520, 536, 540, 560, 561, 566, 570, 571, 579, 581, 589, 591, 593, 597, 601, 613, 614, 621, 622, 625, 628, 632, 636, 644, 647, 666, 671, 682, 690, 695, 702, 708, 721, 727, 729, 730, 734, 740, 754, 758, 770, 773, 774, 778, 788, 791, 795, 798, 801, 805, 809, 817, 819, 820, 821, 823, 824, 828, 837, 840, 847, 856, 861, 866, 869, 870, 875, 886, 889, 890, 894, 899, 904, 927, 935, 937, 942, 945, 952, 962, 964, 966, 969, 980, 982, 989, 991, 994, 998, 1001, 1003, 1006, 1011, 1012, 1014, 1015, 1019, 1022, 1036, 1038, 1040, 1043, 1051, 1053, 1076, 1077, 1080, 1087, 1089, 1096, 1120, 1125, 1126, 1129, 1158, 1161, 1166, 1172, 1174, 1178, 1180, 1181, 1185, 1198, 1203, 1212, 1228, 1232, 1239, 1240, 1241, 1251, 1252, 1254, 1258, 1259, 1266, 1276, 1282, 1289, 1298, 1305, 1308, 1311, 1313, 1315, 1317, 1324, 1331, 1339, 1342, 1349, 1353, 1357, 1363, 1364, 1367, 1385, 1387, 1388, 1406, 1409, 1417, 1420, 1422, 1426, 1440, 1451, 1453, 1454, 1459, 1462, 1469, 1476, 1478, 1486, 1490, 1493, 1494, 1496, 1501, 1504, 1512, 1515, 1517, 1518, 1525, 1526, 1544, 1547, 1554, 1560, 1562, 1580, 1581, 1585, 1590, 1593, 1594, 1603, 1604, 1605, 1613, 1622, 1627, 1633, 1636, 1637, 1641, 1645, 1648, 1650, 1654, 1655, 1659, 1662, 1665, 1672, 1673, 1674, 1679, 1686, 1690, 1691, 1695, 1712, 1716, 1718, 1726, 1727, 1734, 1736, 1751, 1755, 1756, 1769, 1780, 1786, 1791, 1796, 1809, 1812, 1827, 1835, 1841, 1858, 1860, 1867, 1869, 1879, 1880, 1888, 1897, 1907, 1914, 1915, 1945, 1953, 1958, 1963, 1964, 1973, 1978, 1981, 1984, 1986, 1989, 1990, 1995, 2017, 2021, 2028, 2029, 2035, 2036, 2049, 2054, 2064, 2067, 2070, 2072, 2076, 2077, 2106, 2109, 2110, 2112, 2119, 2120, 2121, 2126, 2133, 2147, 2159, 2161, 2162, 2164, 2169, 2172, 2174, 2177, 2181, 2185, 2186, 2193, 2194, 2195, 2196, 2197, 2199, 2212, 2223, 2226, 2237, 2240, 2243, 2262, 2274, 2278, 2281, 2285, 2297, 2304, 2311, 2317, 2320, 2324, 2330, 2333, 2339, 2344, 2348, 2354, 2360, 2364, 2366, 2367, 2378, 2389, 2395, 2396, 2400, 2404, 2406, 2409, 2410, 2423, 2429, 2437, 2440, 2442, 2463, 2488, 2512, 2522, 2529, 2530, 2532, 2535, 2538, 2546, 2548, 2549, 2553, 2556, 2559, 2574, 2575, 2577, 2581, 2589, 2590, 2616, 2629, 2632, 2644, 2663, 2667, 2676, 2678, 2714, 2716, 2731, 2736, 2742, 2744, 2745, 2746, 2749, 2754, 2757, 2758, 2762, 2766, 2777, 2788, 2791, 2794, 2797, 2798, 2801, 2806, 2807, 2815, 2816, 2820, 2823, 2837, 2843, 2854, 2857, 2860, 2861, 2867, 2877, 2880, 2882, 2884, 2899, 2905, 2912, 2913, 2916, 2921, 2933, 2937, 2938, 2942, 2944, 2949, 2953, 2955, 2963, 2968, 2977, 2979, 2985, 2995, 2998]}\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter True datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:50:25,126 >> loading file ./tokenizer.model from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/./tokenizer.model\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:50:25,126 >> loading file added_tokens.json from cache at None\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:50:25,126 >> loading file special_tokens_map.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/special_tokens_map.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:50:25,126 >> loading file tokenizer_config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/tokenizer_config.json\n","[INFO|tokenization_utils_base.py:2108] 2024-07-22 21:50:25,126 >> loading file tokenizer.json from cache at None\n","07/22/2024 21:50:25 - INFO - llamafactory.data.template - Add <|im_end|> to stop words.\n","/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.\n"," warnings.warn(\n","[INFO|configuration_utils.py:733] 2024-07-22 21:50:25,733 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","[INFO|configuration_utils.py:733] 2024-07-22 21:50:26,241 >> loading configuration file config.json from cache at /home/inflaton/.cache/huggingface/hub/models--internlm--internlm2_5-7b-chat-1m/snapshots/8d1a709a04d71440ef3df6ebbe204672f411c8b6/config.json\n","[INFO|configuration_utils.py:796] 2024-07-22 21:50:26,242 >> Model config InternLM2Config {\n"," \"_name_or_path\": \"internlm/internlm2_5-7b-chat-1m\",\n"," \"architectures\": [\n"," \"InternLM2ForCausalLM\"\n"," ],\n"," \"attn_implementation\": \"eager\",\n"," \"auto_map\": {\n"," \"AutoConfig\": \"internlm/internlm2_5-7b-chat-1m--configuration_internlm2.InternLM2Config\",\n"," \"AutoModel\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\",\n"," \"AutoModelForCausalLM\": \"internlm/internlm2_5-7b-chat-1m--modeling_internlm2.InternLM2ForCausalLM\"\n"," },\n"," \"bias\": false,\n"," \"bos_token_id\": 1,\n"," \"eos_token_id\": 2,\n"," \"hidden_act\": \"silu\",\n"," \"hidden_size\": 4096,\n"," \"initializer_range\": 0.02,\n"," \"intermediate_size\": 14336,\n"," \"max_position_embeddings\": 262144,\n"," \"model_type\": \"internlm2\",\n"," \"num_attention_heads\": 32,\n"," \"num_hidden_layers\": 32,\n"," \"num_key_value_heads\": 8,\n"," \"pad_token_id\": 2,\n"," \"pretraining_tp\": 1,\n"," \"rms_norm_eps\": 1e-05,\n"," \"rope_scaling\": {\n"," \"factor\": 2.5,\n"," \"type\": \"dynamic\"\n"," },\n"," \"rope_theta\": 50000000,\n"," \"tie_word_embeddings\": false,\n"," \"torch_dtype\": \"bfloat16\",\n"," \"transformers_version\": \"4.41.2\",\n"," \"use_cache\": true,\n"," \"vocab_size\": 92544\n","}\n","\n","Traceback (most recent call last):\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 563, in from_name\n"," return next(cls.discover(name=name))\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n","StopIteration\n","\n","During handling of the above exception, another exception occurred:\n","\n","Traceback (most recent call last):\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/versions.py\", line 102, in require_version\n"," got_ver = importlib.metadata.version(pkg)\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 1009, in version\n"," return distribution(distribution_name).version\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 982, in distribution\n"," return Distribution.from_name(distribution_name)\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 565, in from_name\n"," raise PackageNotFoundError(name)\n","importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes\n","\n","During handling of the above exception, another exception occurred:\n","\n","Traceback (most recent call last):\n"," File \"/home/inflaton/code/logical-reasoning/llm_toolkit/eval_logical_reasoning.py\", line 49, in \n"," model, tokenizer = load_model(\n"," ^^^^^^^^^^^\n"," File \"/home/inflaton/code/logical-reasoning/llm_toolkit/llm_utils.py\", line 45, in load_model\n"," chat_model = ChatModel(args)\n"," ^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/chat/chat_model.py\", line 44, in __init__\n"," self.engine: \"BaseEngine\" = HuggingfaceEngine(model_args, data_args, finetuning_args, generating_args)\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/chat/hf_engine.py\", line 58, in __init__\n"," self.model = load_model(\n"," ^^^^^^^^^^^\n"," File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/model/loader.py\", line 132, in load_model\n"," patch_config(config, tokenizer, model_args, init_kwargs, is_trainable)\n"," File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/model/patcher.py\", line 75, in patch_config\n"," configure_quantization(config, tokenizer, model_args, init_kwargs)\n"," File \"/home/inflaton/code/LLaMA-Factory/src/llamafactory/model/model_utils/quantization.py\", line 160, in configure_quantization\n"," require_version(\"bitsandbytes>=0.39.0\", \"To fix: pip install bitsandbytes>=0.39.0\")\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/versions.py\", line 104, in require_version\n"," raise importlib.metadata.PackageNotFoundError(\n","importlib.metadata.PackageNotFoundError: No package metadata was found for The 'bitsandbytes>=0.39.0' distribution was not found and is required by this application. \n","To fix: pip install bitsandbytes>=0.39.0\n","CPU times: user 15.4 s, sys: 5.13 s, total: 20.5 s\n","Wall time: 39min 43s\n"]}],"source":["%%time\n","\n","os.environ[\"USING_LLAMA_FACTORY\"] = \"true\"\n","!./scripts/eval-mgtv-nv4090.sh"]},{"cell_type":"code","execution_count":16,"metadata":{},"outputs":[{"name":"stdout","output_type":"stream","text":["Current Directory:\n","/home/inflaton/code/logical-reasoning\n","Mon Jul 22 21:50:27 2024 \n","+---------------------------------------------------------------------------------------+\n","| NVIDIA-SMI 535.54.04 Driver Version: 536.25 CUDA Version: 12.2 |\n","|-----------------------------------------+----------------------+----------------------+\n","| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |\n","| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |\n","| | | MIG M. |\n","|=========================================+======================+======================|\n","| 0 NVIDIA GeForce RTX 4090 On | 00000000:01:00.0 Off | Off |\n","| 48% 55C P8 19W / 450W | 1460MiB / 24564MiB | 3% Default |\n","| | | N/A |\n","+-----------------------------------------+----------------------+----------------------+\n"," \n","+---------------------------------------------------------------------------------------+\n","| Processes: |\n","| GPU GI CI PID Type Process name GPU Memory |\n","| ID ID Usage |\n","|=======================================================================================|\n","| No running processes found |\n","+---------------------------------------------------------------------------------------+\n","Linux Gen-AI 5.15.133.1-microsoft-standard-WSL2 #1 SMP Thu Oct 5 21:02:42 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux\n","PRETTY_NAME=\"Ubuntu 22.04.2 LTS\"\n","NAME=\"Ubuntu\"\n","VERSION_ID=\"22.04\"\n","VERSION=\"22.04.2 LTS (Jammy Jellyfish)\"\n","VERSION_CODENAME=jammy\n","ID=ubuntu\n","ID_LIKE=debian\n","HOME_URL=\"https://www.ubuntu.com/\"\n","SUPPORT_URL=\"https://help.ubuntu.com/\"\n","BUG_REPORT_URL=\"https://bugs.launchpad.net/ubuntu/\"\n","PRIVACY_POLICY_URL=\"https://www.ubuntu.com/legal/terms-and-policies/privacy-policy\"\n","UBUNTU_CODENAME=jammy\n","Architecture: x86_64\n"," CPU op-mode(s): 32-bit, 64-bit\n"," Address sizes: 39 bits physical, 48 bits virtual\n"," Byte Order: Little Endian\n","CPU(s): 32\n"," On-line CPU(s) list: 0-31\n","Vendor ID: GenuineIntel\n"," Model name: 13th Gen Intel(R) Core(TM) i9-13900KF\n"," CPU family: 6\n"," Model: 183\n"," Thread(s) per core: 2\n"," Core(s) per socket: 16\n"," Socket(s): 1\n"," Stepping: 1\n"," BogoMIPS: 5990.40\n"," Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mc\n"," a cmov pat pse36 clflush mmx fxsr sse sse2 ss ht syscal\n"," l nx pdpe1gb rdtscp lm constant_tsc rep_good nopl xtopo\n"," logy tsc_reliable nonstop_tsc cpuid pni pclmulqdq vmx s\n"," sse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt tsc_dea\n"," dline_timer aes xsave avx f16c rdrand hypervisor lahf_l\n"," m abm 3dnowprefetch ssbd ibrs ibpb stibp ibrs_enhanced \n"," tpr_shadow vnmi ept vpid ept_ad fsgsbase tsc_adjust bmi\n"," 1 avx2 smep bmi2 erms invpcid rdseed adx smap clflushop\n"," t clwb sha_ni xsaveopt xsavec xgetbv1 xsaves avx_vnni u\n"," mip waitpkg gfni vaes vpclmulqdq rdpid movdiri movdir64\n"," b fsrm md_clear serialize flush_l1d arch_capabilities\n","Virtualization features: \n"," Virtualization: VT-x\n"," Hypervisor vendor: Microsoft\n"," Virtualization type: full\n","Caches (sum of all): \n"," L1d: 768 KiB (16 instances)\n"," L1i: 512 KiB (16 instances)\n"," L2: 32 MiB (16 instances)\n"," L3: 36 MiB (1 instance)\n","Vulnerabilities: \n"," Gather data sampling: Not affected\n"," Itlb multihit: Not affected\n"," L1tf: Not affected\n"," Mds: Not affected\n"," Meltdown: Not affected\n"," Mmio stale data: Not affected\n"," Retbleed: Mitigation; Enhanced IBRS\n"," Spec rstack overflow: Not affected\n"," Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl\n"," and seccomp\n"," Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer\n"," sanitization\n"," Spectre v2: Mitigation; Enhanced IBRS, IBPB conditional, RSB fillin\n"," g, PBRSB-eIBRS SW sequence\n"," Srbds: Not affected\n"," Tsx async abort: Not affected\n","MemTotal: 49330020 kB\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter False datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","Loading checkpoint shards: 100%|██████████████████| 8/8 [00:03<00:00, 2.56it/s]\n","(2) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","14.486 GB of memory reserved.\n","loading train/test data files\n","Map: 100%|██████████████████████| 25000/25000 [00:00<00:00, 64711.02 examples/s]\n","Map: 100%|████████████████████████| 3000/3000 [00:00<00:00, 62306.45 examples/s]\n","DatasetDict({\n"," train: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 25000\n"," })\n"," test: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 3000\n"," })\n","})\n","--------------------------------------------------\n","text: 甄加索是自杀吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 海岸之谜\n","--------------------------------------------------\n","puzzle: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","--------------------------------------------------\n","truth: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","--------------------------------------------------\n","text: 死者受伤了吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 甄庄哭声\n","--------------------------------------------------\n","puzzle: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","--------------------------------------------------\n","truth: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","Evaluating model: internlm/internlm2_5-7b-chat-1m\n"," 0%| | 0/3000 [00:00, ?it/s]Batch output: ['不是']\n","100%|███████████████████████████████████████| 3000/3000 [08:06<00:00, 6.16it/s]\n","(3) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","16.586 GB of memory reserved.\n"," text ... internlm/internlm2_5-7b-chat-1m/InternLM_2_5-7b_LoRA-Adapter_torch.float16\n","0 甄加索是自杀吗 ... 不是 \n","\n","[1 rows x 13 columns]\n","{'accuracy': 0.8013333333333333, 'incorrect_ids': [6, 9, 18, 24, 29, 31, 34, 36, 55, 58, 59, 61, 65, 66, 67, 78, 81, 83, 84, 88, 91, 93, 103, 104, 106, 112, 121, 128, 129, 135, 139, 143, 149, 150, 153, 155, 161, 164, 170, 173, 190, 199, 200, 202, 222, 224, 228, 245, 248, 250, 252, 259, 260, 269, 271, 275, 276, 284, 286, 293, 299, 304, 314, 318, 323, 326, 328, 330, 334, 335, 337, 346, 350, 353, 355, 356, 360, 361, 368, 370, 373, 377, 383, 384, 386, 389, 395, 397, 410, 416, 428, 429, 430, 432, 438, 445, 447, 450, 452, 454, 456, 457, 458, 465, 469, 470, 471, 472, 473, 476, 479, 486, 490, 492, 494, 495, 497, 501, 502, 506, 507, 508, 510, 511, 517, 518, 519, 520, 530, 536, 540, 560, 566, 570, 571, 579, 581, 589, 591, 593, 596, 597, 601, 613, 614, 621, 622, 625, 628, 632, 636, 644, 647, 666, 671, 682, 690, 695, 702, 708, 721, 727, 729, 730, 734, 739, 740, 754, 758, 766, 770, 773, 774, 778, 788, 791, 795, 798, 801, 805, 809, 817, 819, 820, 821, 823, 824, 826, 828, 837, 840, 847, 856, 861, 866, 869, 870, 875, 886, 889, 890, 899, 904, 913, 927, 935, 937, 942, 945, 952, 962, 964, 966, 969, 980, 982, 989, 991, 994, 998, 1001, 1003, 1006, 1011, 1012, 1014, 1015, 1019, 1022, 1036, 1038, 1040, 1043, 1051, 1053, 1057, 1075, 1076, 1077, 1080, 1087, 1089, 1096, 1120, 1125, 1126, 1129, 1158, 1161, 1166, 1172, 1174, 1178, 1180, 1181, 1185, 1198, 1203, 1212, 1228, 1232, 1239, 1240, 1241, 1251, 1252, 1254, 1258, 1259, 1266, 1282, 1289, 1298, 1305, 1308, 1311, 1313, 1315, 1317, 1324, 1331, 1332, 1339, 1342, 1349, 1353, 1357, 1363, 1364, 1367, 1385, 1387, 1388, 1406, 1409, 1417, 1420, 1422, 1426, 1440, 1444, 1451, 1453, 1454, 1459, 1462, 1469, 1476, 1478, 1486, 1490, 1493, 1494, 1496, 1501, 1504, 1512, 1517, 1518, 1525, 1526, 1544, 1547, 1554, 1560, 1562, 1580, 1581, 1585, 1590, 1593, 1594, 1603, 1604, 1605, 1613, 1622, 1627, 1633, 1636, 1637, 1641, 1645, 1647, 1648, 1650, 1654, 1655, 1659, 1662, 1665, 1672, 1673, 1674, 1679, 1686, 1691, 1695, 1712, 1716, 1718, 1721, 1726, 1727, 1728, 1734, 1736, 1751, 1755, 1756, 1766, 1769, 1780, 1786, 1791, 1796, 1798, 1809, 1812, 1827, 1835, 1858, 1860, 1867, 1869, 1877, 1879, 1880, 1888, 1894, 1897, 1907, 1914, 1915, 1919, 1945, 1953, 1958, 1963, 1964, 1971, 1973, 1978, 1981, 1984, 1986, 1989, 1990, 1995, 2017, 2021, 2028, 2029, 2035, 2036, 2049, 2054, 2064, 2067, 2070, 2076, 2077, 2100, 2106, 2109, 2110, 2112, 2119, 2120, 2121, 2126, 2133, 2147, 2159, 2161, 2162, 2164, 2169, 2172, 2174, 2177, 2181, 2185, 2186, 2192, 2193, 2194, 2195, 2196, 2197, 2199, 2212, 2222, 2223, 2226, 2230, 2233, 2237, 2240, 2243, 2262, 2265, 2274, 2278, 2281, 2285, 2297, 2304, 2311, 2317, 2320, 2322, 2324, 2330, 2333, 2339, 2344, 2348, 2360, 2364, 2366, 2367, 2378, 2389, 2395, 2396, 2400, 2404, 2406, 2409, 2410, 2423, 2429, 2437, 2440, 2442, 2454, 2463, 2469, 2488, 2511, 2512, 2515, 2517, 2522, 2529, 2530, 2532, 2535, 2538, 2539, 2546, 2548, 2549, 2553, 2556, 2559, 2560, 2574, 2575, 2577, 2589, 2590, 2616, 2624, 2629, 2632, 2644, 2663, 2664, 2667, 2676, 2678, 2714, 2716, 2727, 2731, 2736, 2742, 2744, 2745, 2746, 2749, 2754, 2757, 2758, 2762, 2764, 2766, 2770, 2777, 2788, 2791, 2797, 2798, 2801, 2806, 2807, 2815, 2816, 2820, 2823, 2837, 2843, 2854, 2857, 2860, 2861, 2867, 2877, 2880, 2882, 2884, 2899, 2902, 2905, 2912, 2913, 2916, 2921, 2933, 2937, 2938, 2942, 2944, 2949, 2953, 2955, 2963, 2968, 2969, 2977, 2979, 2985, 2995, 2998]}\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter False datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","Loading checkpoint shards: 100%|██████████████████| 8/8 [00:02<00:00, 3.31it/s]\n","(2) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","14.486 GB of memory reserved.\n","loading train/test data files\n","DatasetDict({\n"," train: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 25000\n"," })\n"," test: Dataset({\n"," features: ['text', 'label', 'answer', 'title', 'puzzle', 'truth', 'train_text', 'prompt'],\n"," num_rows: 3000\n"," })\n","})\n","--------------------------------------------------\n","text: 甄加索是自杀吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 海岸之谜\n","--------------------------------------------------\n","puzzle: 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","--------------------------------------------------\n","truth: 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在远离城市喧嚣的海边小屋,一天清晨,邻居发现甄加索僵卧在沙滩上,已无生命迹象。现场没有发现任何打斗的迹象。请问甄加索的死因是什么?\n","\n","**谜底:** 甄加索是一位热爱自然的画家,他每年都会来到这个海边小屋寻找灵感。在他生命的最后几天,他一直在创作一幅描绘海洋生物的画作。在画即将完成的前一天晚上,他骑着自行车外出,打算在海边观赏夜景。然而,他在沙滩上意外发现了一只搁浅的海豚,为了救助这只海豚,他耗费了极大的体力,最终成功将其送回海中。筋疲力尽的甄加索在沙滩上睡着了,由于他患有严重的心脏病,却未告知旁人,在寒冷的海风中,他的心脏停止了跳动。因此,警方在现场只发现了车轮痕迹和未完成的画作,而没有发现任何他杀的迹象。\n","\n","**参与者提出的问题:** 甄加索是自杀吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","--------------------------------------------------\n","text: 死者受伤了吗\n","--------------------------------------------------\n","label: 不是\n","--------------------------------------------------\n","answer: nan\n","--------------------------------------------------\n","title: 甄庄哭声\n","--------------------------------------------------\n","puzzle: 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","--------------------------------------------------\n","truth: 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","--------------------------------------------------\n","train_text: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","不是\n","--------------------------------------------------\n","prompt: <|im_start|>system\n","You are an expert in logical reasoning.<|im_end|>\n","<|im_start|>user\n","你是一个情景猜谜游戏的主持人。游戏规则如下:\n","\n","1. 参与者会得到一个谜面,谜面会描述一个简单又难以理解的事件。\n","2. 主持人知道谜底,谜底是谜面的答案。\n","3. 参与者可以询问任何封闭式问题来找寻事件的真相。\n","4. 对于每个问题,主持人将根据实际情况回答以下五个选项之一:是、不是、不重要、回答正确、问法错误。各回答的判断标准如下:\n"," - 若谜面和谜底能找到问题的答案,回答:是或者不是\n"," - 若谜面和谜底不能直接或者间接推断出问题的答案,回答:不重要\n"," - 若参与者提问不是一个封闭式问题或者问题难以理解,回答:问法错误\n"," - 若参与者提问基本还原了谜底真相,回答:回答正确\n","5. 回答中不能添加任何其它信息,也不能省略选项中的任何一个字。例如,不可以把“不是”省略成“不”。\n","\n","请严格按照这些规则回答参与者提出的问题。\n","\n","**谜面:** 在一个安静的夜晚,小村庄的湖边突然传来了阵阵哭泣声。第二天早晨,村长甄锐发现湖边的石头上放着一顶破旧的帽子,但没有人知道这顶帽子是从哪里来的,哭泣声又是为何。请还原故事真相。\n","\n","**谜底:** 原来,这顶破旧的帽子属于一个小男孩,他小时候与爷爷在湖边生活。爷爷教他钓鱼、游泳,还告诉他湖中的海龟是他们的朋友。后来,小男孩随父母去了城市生活,但每年夏天都会回到村子探望爷爷。然而,去年夏天,爷爷因病去世,小男孩伤心欲绝。今年夏天,他回到村子,来到湖边,想起和爷爷的美好回忆,忍不住哭泣。他将爷爷的帽子放在湖边的石头上,希望能让爷爷的在天之灵得到安慰。那晚的哭泣声正是小男孩在祭莫他亲爱的爷爷。\n","\n","**参与者提出的问题:** 死者受伤了吗\n","<|im_end|>\n","<|im_start|>assistant\n","\n","Evaluating model: internlm/internlm2_5-7b-chat-1m\n"," 0%| | 0/3000 [00:00, ?it/s]Batch output: ['不是']\n","100%|███████████████████████████████████████| 3000/3000 [08:08<00:00, 6.15it/s]\n","(3) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","16.613 GB of memory reserved.\n"," text ... internlm/internlm2_5-7b-chat-1m/InternLM_2_5-7b_LoRA-Adapter_torch.bfloat16\n","0 甄加索是自杀吗 ... 不是 \n","\n","[1 rows x 14 columns]\n","{'accuracy': 0.81, 'incorrect_ids': [6, 9, 13, 16, 18, 29, 31, 34, 36, 55, 58, 59, 61, 65, 66, 67, 78, 81, 83, 84, 88, 91, 103, 104, 106, 110, 112, 121, 123, 128, 129, 135, 137, 138, 143, 149, 150, 153, 161, 164, 170, 173, 190, 199, 200, 202, 222, 224, 228, 245, 250, 259, 260, 269, 271, 275, 276, 286, 293, 299, 304, 314, 318, 323, 326, 328, 330, 334, 335, 337, 341, 346, 350, 353, 356, 360, 361, 368, 370, 373, 377, 383, 384, 386, 389, 397, 410, 416, 428, 429, 430, 432, 438, 445, 447, 450, 452, 454, 456, 457, 458, 465, 469, 471, 472, 476, 479, 481, 486, 495, 497, 501, 502, 506, 507, 510, 511, 517, 518, 519, 520, 530, 536, 540, 560, 566, 570, 571, 579, 581, 589, 591, 593, 597, 601, 613, 614, 621, 622, 625, 628, 632, 636, 644, 647, 666, 671, 682, 690, 695, 702, 708, 721, 727, 729, 730, 734, 739, 740, 754, 758, 769, 770, 773, 774, 778, 788, 791, 794, 795, 798, 801, 805, 809, 817, 819, 820, 821, 823, 824, 828, 837, 840, 844, 847, 856, 861, 862, 866, 869, 870, 875, 886, 889, 890, 894, 904, 913, 927, 935, 937, 942, 945, 952, 962, 964, 966, 969, 980, 982, 989, 991, 994, 998, 1001, 1003, 1005, 1006, 1011, 1012, 1014, 1015, 1019, 1036, 1038, 1040, 1043, 1051, 1053, 1057, 1075, 1076, 1077, 1080, 1087, 1089, 1096, 1120, 1125, 1126, 1129, 1158, 1161, 1166, 1172, 1174, 1178, 1180, 1181, 1185, 1198, 1203, 1212, 1228, 1232, 1239, 1240, 1241, 1251, 1252, 1254, 1258, 1259, 1266, 1276, 1282, 1289, 1298, 1305, 1308, 1311, 1313, 1315, 1317, 1324, 1331, 1332, 1339, 1342, 1349, 1353, 1357, 1363, 1364, 1367, 1385, 1387, 1388, 1406, 1409, 1417, 1420, 1422, 1426, 1440, 1451, 1453, 1454, 1459, 1462, 1469, 1476, 1478, 1486, 1490, 1493, 1494, 1496, 1501, 1504, 1512, 1515, 1517, 1518, 1525, 1526, 1544, 1547, 1550, 1554, 1560, 1562, 1580, 1581, 1585, 1590, 1593, 1603, 1604, 1605, 1613, 1622, 1627, 1633, 1636, 1637, 1641, 1645, 1647, 1648, 1650, 1654, 1655, 1659, 1662, 1665, 1672, 1673, 1674, 1679, 1686, 1691, 1695, 1712, 1716, 1718, 1726, 1727, 1734, 1736, 1751, 1755, 1756, 1769, 1780, 1786, 1791, 1796, 1809, 1812, 1827, 1835, 1841, 1858, 1860, 1867, 1869, 1879, 1888, 1894, 1897, 1907, 1914, 1945, 1953, 1958, 1963, 1964, 1973, 1978, 1981, 1984, 1986, 1989, 1990, 1991, 1995, 2017, 2021, 2028, 2029, 2035, 2036, 2049, 2054, 2064, 2067, 2070, 2076, 2077, 2100, 2106, 2109, 2110, 2119, 2120, 2121, 2126, 2133, 2147, 2159, 2161, 2162, 2164, 2172, 2174, 2177, 2181, 2185, 2186, 2192, 2193, 2194, 2195, 2196, 2197, 2199, 2208, 2212, 2223, 2226, 2237, 2240, 2243, 2257, 2262, 2274, 2278, 2280, 2281, 2285, 2297, 2304, 2311, 2317, 2320, 2322, 2324, 2330, 2333, 2339, 2344, 2348, 2354, 2360, 2364, 2366, 2367, 2378, 2389, 2395, 2396, 2400, 2404, 2406, 2409, 2410, 2423, 2429, 2437, 2440, 2442, 2463, 2488, 2511, 2512, 2517, 2522, 2529, 2530, 2532, 2535, 2538, 2546, 2548, 2549, 2553, 2556, 2559, 2574, 2575, 2577, 2581, 2589, 2590, 2616, 2629, 2632, 2644, 2663, 2667, 2676, 2678, 2714, 2716, 2731, 2736, 2742, 2745, 2746, 2749, 2754, 2757, 2758, 2766, 2777, 2788, 2791, 2794, 2797, 2798, 2806, 2807, 2815, 2816, 2820, 2823, 2837, 2843, 2854, 2857, 2860, 2861, 2867, 2877, 2880, 2882, 2884, 2899, 2905, 2912, 2913, 2916, 2921, 2927, 2933, 2937, 2938, 2942, 2944, 2949, 2953, 2955, 2964, 2968, 2977, 2979, 2985, 2995, 2998]}\n","Eval internlm/internlm2_5-7b-chat-1m with inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","loading env vars from: /home/inflaton/code/logical-reasoning/.env\n","Adding /home/inflaton/code/logical-reasoning to sys.path\n","loading /home/inflaton/code/logical-reasoning/llm_toolkit/logical_reasoning_utils.py\n","internlm/internlm2_5-7b-chat-1m inflaton-ai/InternLM_2_5-7b_LoRA-Adapter True datasets/mgtv results/mgtv-results_internlm_best.csv\n","(1) GPU = NVIDIA GeForce RTX 4090. Max memory = 23.988 GB.\n","0.0 GB of memory reserved.\n","loading model: internlm/internlm2_5-7b-chat-1m with adapter: inflaton-ai/InternLM_2_5-7b_LoRA-Adapter\n","Traceback (most recent call last):\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 563, in from_name\n"," return next(cls.discover(name=name))\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n","StopIteration\n","\n","During handling of the above exception, another exception occurred:\n","\n","Traceback (most recent call last):\n"," File \"/home/inflaton/code/logical-reasoning/llm_toolkit/eval_logical_reasoning.py\", line 49, in \n"," model, tokenizer = load_model(\n"," ^^^^^^^^^^^\n"," File \"/home/inflaton/code/logical-reasoning/llm_toolkit/llm_utils.py\", line 53, in load_model\n"," bnb_config = BitsAndBytesConfig(\n"," ^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/quantization_config.py\", line 395, in __init__\n"," self.post_init()\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/site-packages/transformers/utils/quantization_config.py\", line 453, in post_init\n"," if self.load_in_4bit and not version.parse(importlib.metadata.version(\"bitsandbytes\")) >= version.parse(\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 1009, in version\n"," return distribution(distribution_name).version\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 982, in distribution\n"," return Distribution.from_name(distribution_name)\n"," ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n"," File \"/home/inflaton/miniconda3/envs/llm-finetuning/lib/python3.11/importlib/metadata/__init__.py\", line 565, in from_name\n"," raise PackageNotFoundError(name)\n","importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes\n","CPU times: user 7.47 s, sys: 2.41 s, total: 9.88 s\n","Wall time: 17min 13s\n"]}],"source":["%%time\n","\n","os.environ[\"USING_LLAMA_FACTORY\"] = \"false\"\n","!./scripts/eval-mgtv-nv4090.sh"]}],"metadata":{"accelerator":"GPU","application/vnd.databricks.v1+notebook":{"dashboards":[],"environmentMetadata":null,"language":"python","notebookMetadata":{"mostRecentlyExecutedCommandWithImplicitDF":{"commandId":-1,"dataframes":["_sqldf"]},"pythonIndentUnit":4},"notebookName":"10_eval-lf-medium-py3.11","widgets":{}},"colab":{"gpuType":"L4","provenance":[]},"kernelspec":{"display_name":"Python 3","name":"python3"},"language_info":{"codemirror_mode":{"name":"ipython","version":3},"file_extension":".py","mimetype":"text/x-python","name":"python","nbconvert_exporter":"python","pygments_lexer":"ipython3","version":"3.11.9"}},"nbformat":4,"nbformat_minor":0}