_id
stringlengths 24
24
| id
stringlengths 14
79
| author
stringlengths 3
21
| gated
stringclasses 3
values | inference
stringclasses 9
values | lastModified
stringlengths 24
24
| likes
int64 0
4.13k
| private
bool 1
class | sha
stringlengths 40
40
| config
stringlengths 23
5.92k
⌀ | downloads
int64 1.8k
3.13M
| tags
sequencelengths 5
62
| pipeline_tag
stringclasses 6
values | library_name
stringclasses 4
values | createdAt
stringlengths 24
24
| modelId
stringlengths 14
79
| siblings
listlengths 4
285
| model_type
stringclasses 13
values | base_model
stringlengths 13
43
| base_model_downloads
float64 98
28.9k
| base_model_pipeline
stringclasses 1
value | base_model_children_count
float64 98
28.9k
| adapter
float64 0
377
| merge
float64 0
94
| finetune
float64 0
263
| quantized
float64 0
182
| __index_level_0__
int64 86k
992k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
6660ffab694ec45eaa99d251 | cognitivecomputations/dolphin-2.9.3-llama-3-8b | cognitivecomputations | False | not-popular-enough | 2024-06-06T00:16:45.000Z | 9 | false | 27a8c71b28c8ce94ec57057951abb546d80ab378 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}} | 2,445 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-06T00:15:39.000Z | cognitivecomputations/dolphin-2.9.3-llama-3-8b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 1 | 10 | 5 | 684,601 |
6649cc0c4a52f105d78e938c | Metin/LLaMA-3-8B-Instruct-TR-DPO | Metin | False | not-popular-enough | 2024-06-16T09:48:43.000Z | 7 | false | 5a24a2391d417581c30425b4d6053f62cd736dc1 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 2,442 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-19T09:53:16.000Z | Metin/LLaMA-3-8B-Instruct-TR-DPO | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "llama.png"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 10 | 0 | 646,603 |
66de9bfe8f6f3907134e55dc | nekokiku/01-ai-Yi-1.5-9B-Chat-1725864958 | nekokiku | False | pipeline-not-detected | 2024-09-09T07:05:17.000Z | 0 | false | 8978e77d6e990a61cd8e6254bfaa1fbf0877bf3b | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}} | 2,420 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null | peft | 2024-09-09T06:55:58.000Z | nekokiku/01-ai-Yi-1.5-9B-Chat-1725864958 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B-Chat | 134 | text-generation | 134 | 0 | 0 | 0 | 0 | 918,544 |
653f5d1e7bd6a974393bc5fc | FinGPT/fingpt-forecaster_dow30_llama2-7b_lora | FinGPT | False | pipeline-not-detected | 2024-06-11T02:42:34.000Z | 102 | false | 69f77190315afdb03a889d89bf2a0f932b311617 | {"peft": {"base_model_name_or_path": "base_models/Llama-2-7b-chat-hf", "task_type": "CAUSAL_LM"}} | 2,414 | [
"peft",
"safetensors",
"en",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:adapter:meta-llama/Llama-2-7b-chat-hf",
"license:apache-2.0",
"region:us"
] | null | peft | 2023-10-30T07:37:02.000Z | FinGPT/fingpt-forecaster_dow30_llama2-7b_lora | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.bin"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 0 | 0 | 0 | 0 | 361,194 |
6622b0d72d61edba3dbe3aa5 | Mozilla/Meta-Llama-3-8B-Instruct-llamafile | Mozilla | False | library-not-detected | 2024-08-19T10:25:36.000Z | 51 | false | 902a5de8cc6c142515435f5009a46df9d9947d0d | null | 2,406 | [
"llamafile",
"llama",
"llama-3",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"region:us"
] | text-generation | null | 2024-04-19T17:58:47.000Z | Mozilla/Meta-Llama-3-8B-Instruct-llamafile | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.BF16.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.F16.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q2_K.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_L.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_M.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_S.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_0.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_1.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_M.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_S.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_0.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_1.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_M.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_S.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q6_K.llamafile"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q8_0.llamafile"
},
{
"rfilename": "Meta-Llama-3-Community-License-Agreement.txt"
},
{
"rfilename": "README.md"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 594,322 |
64b6ce072a8e3cd59df98e98 | TheBloke/Llama-2-7B-Chat-GGML | TheBloke | False | explicit-opt-out | 2023-09-27T13:00:17.000Z | 843 | false | 76cd63c351ae389e1d4b91cab2cf470aab11864b | {"model_type": "llama"} | 2,399 | [
"transformers",
"llama",
"facebook",
"meta",
"pytorch",
"llama-2",
"text-generation",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-7b-chat-hf",
"base_model:finetune:meta-llama/Llama-2-7b-chat-hf",
"license:other",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-07-18T17:38:15.000Z | TheBloke/Llama-2-7B-Chat-GGML | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q2_K.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q3_K_L.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q3_K_M.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q3_K_S.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q4_0.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q4_1.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q4_K_M.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q4_K_S.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q5_0.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q5_1.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q5_K_M.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q5_K_S.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q6_K.bin"
},
{
"rfilename": "llama-2-7b-chat.ggmlv3.q8_0.bin"
}
] | llama | meta-llama/Llama-2-7b-chat-hf | 1,328 | text-generation | 1,328 | 0 | 0 | 2 | 0 | 247,837 |
668d255070e29925074009dd | curiositytech/MARS | curiositytech | False | not-popular-enough | 2024-07-10T14:18:14.000Z | 5 | false | 29dd2a871c8e1515b2d41e11f29dd0721bfeec46 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 2,393 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"tr",
"en",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-09T11:56:00.000Z | curiositytech/MARS | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "MARS-1.0.png"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00009.safetensors"
},
{
"rfilename": "model-00002-of-00009.safetensors"
},
{
"rfilename": "model-00003-of-00009.safetensors"
},
{
"rfilename": "model-00004-of-00009.safetensors"
},
{
"rfilename": "model-00005-of-00009.safetensors"
},
{
"rfilename": "model-00006-of-00009.safetensors"
},
{
"rfilename": "model-00007-of-00009.safetensors"
},
{
"rfilename": "model-00008-of-00009.safetensors"
},
{
"rfilename": "model-00009-of-00009.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 4 | 3 | 743,499 |
66214dadecdd46cf556f7478 | MaziyarPanahi/Meta-Llama-3-8B-Instruct-GGUF | MaziyarPanahi | False | explicit-opt-out | 2024-04-23T12:55:06.000Z | 76 | false | 4ece958b356f2ec00338e5898ba0a7254d976baf | {"model_type": "mistral"} | 2,384 | [
"transformers",
"gguf",
"mistral",
"facebook",
"meta",
"pytorch",
"llama",
"llama-3",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"16-bit",
"GGUF",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2024-04-18T16:43:25.000Z | MaziyarPanahi/Meta-Llama-3-8B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.IQ1_M.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.IQ1_S.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.IQ2_XS.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.IQ3_XS.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.IQ4_XS.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q2_K.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_L.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_M.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_S.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_M.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_S.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_M.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_S.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q6_K.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.Q8_0.gguf"
},
{
"rfilename": "Meta-Llama-3-8B-Instruct.fp16.gguf"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
}
] | mistral | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 592,378 |
66a62c1f2756d40a3fbf31a6 | mlabonne/FineLlama-3.1-8B | mlabonne | False | not-popular-enough | 2024-08-27T17:21:42.000Z | 7 | false | b5c5e22e496468b59660fc391577eb3c740d22c5 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user\n' + message['value'] + '<|im_end|>\n'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant\n' + message['value'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['value'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|finetune_right_pad_id|>", "unk_token": null}} | 2,381 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"dataset:mlabonne/FineTome-100k",
"base_model:unsloth/Meta-Llama-3.1-8B-bnb-4bit",
"base_model:finetune:unsloth/Meta-Llama-3.1-8B-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-28T11:31:43.000Z | mlabonne/FineLlama-3.1-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/Meta-Llama-3.1-8B-bnb-4bit | 508 | text-generation | 508 | 0 | 0 | 0 | 11 | 783,566 |
66a017b16106bf810ef1c0af | Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1 | Magpie-Align | False | not-popular-enough | 2024-08-20T19:10:51.000Z | 1 | false | 7ee0eb6a6c3e055988e6e3f24606c7fe03e5f5ec | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 2,354 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"axolotl",
"generated_from_trainer",
"conversational",
"en",
"dataset:Magpie-Align/Magpie-Reasoning-150K",
"dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1",
"arxiv:2406.08464",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-23T20:50:57.000Z | Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | llama | meta-llama/Llama-3.1-8B | 254 | text-generation | 254 | 0 | 0 | 1 | 2 | 773,315 |
66e1d68a5b1fc436d9cde6ca | jfranklin-foundry/01-ai-Yi-1.5-9B-1726076646 | jfranklin-foundry | False | pipeline-not-detected | 2024-09-11T17:43:40.000Z | 0 | false | 38ed254a11e2ab2d9071cd7e1c074d3b8d7147d2 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 2,354 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-11T17:42:34.000Z | jfranklin-foundry/01-ai-Yi-1.5-9B-1726076646 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "checkpoint-109/README.md"
},
{
"rfilename": "checkpoint-109/adapter_config.json"
},
{
"rfilename": "checkpoint-109/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-109/optimizer.pt"
},
{
"rfilename": "checkpoint-109/rng_state.pth"
},
{
"rfilename": "checkpoint-109/scheduler.pt"
},
{
"rfilename": "checkpoint-109/special_tokens_map.json"
},
{
"rfilename": "checkpoint-109/tokenizer.json"
},
{
"rfilename": "checkpoint-109/tokenizer.model"
},
{
"rfilename": "checkpoint-109/tokenizer_config.json"
},
{
"rfilename": "checkpoint-109/trainer_state.json"
},
{
"rfilename": "checkpoint-109/training_args.bin"
},
{
"rfilename": "checkpoint-164/README.md"
},
{
"rfilename": "checkpoint-164/adapter_config.json"
},
{
"rfilename": "checkpoint-164/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-164/optimizer.pt"
},
{
"rfilename": "checkpoint-164/rng_state.pth"
},
{
"rfilename": "checkpoint-164/scheduler.pt"
},
{
"rfilename": "checkpoint-164/special_tokens_map.json"
},
{
"rfilename": "checkpoint-164/tokenizer.json"
},
{
"rfilename": "checkpoint-164/tokenizer.model"
},
{
"rfilename": "checkpoint-164/tokenizer_config.json"
},
{
"rfilename": "checkpoint-164/trainer_state.json"
},
{
"rfilename": "checkpoint-164/training_args.bin"
},
{
"rfilename": "checkpoint-219/README.md"
},
{
"rfilename": "checkpoint-219/adapter_config.json"
},
{
"rfilename": "checkpoint-219/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-219/optimizer.pt"
},
{
"rfilename": "checkpoint-219/rng_state.pth"
},
{
"rfilename": "checkpoint-219/scheduler.pt"
},
{
"rfilename": "checkpoint-219/special_tokens_map.json"
},
{
"rfilename": "checkpoint-219/tokenizer.json"
},
{
"rfilename": "checkpoint-219/tokenizer.model"
},
{
"rfilename": "checkpoint-219/tokenizer_config.json"
},
{
"rfilename": "checkpoint-219/trainer_state.json"
},
{
"rfilename": "checkpoint-219/training_args.bin"
},
{
"rfilename": "checkpoint-270/README.md"
},
{
"rfilename": "checkpoint-270/adapter_config.json"
},
{
"rfilename": "checkpoint-270/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-270/optimizer.pt"
},
{
"rfilename": "checkpoint-270/rng_state.pth"
},
{
"rfilename": "checkpoint-270/scheduler.pt"
},
{
"rfilename": "checkpoint-270/special_tokens_map.json"
},
{
"rfilename": "checkpoint-270/tokenizer.json"
},
{
"rfilename": "checkpoint-270/tokenizer.model"
},
{
"rfilename": "checkpoint-270/tokenizer_config.json"
},
{
"rfilename": "checkpoint-270/trainer_state.json"
},
{
"rfilename": "checkpoint-270/training_args.bin"
},
{
"rfilename": "checkpoint-274/README.md"
},
{
"rfilename": "checkpoint-274/adapter_config.json"
},
{
"rfilename": "checkpoint-274/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-274/optimizer.pt"
},
{
"rfilename": "checkpoint-274/rng_state.pth"
},
{
"rfilename": "checkpoint-274/scheduler.pt"
},
{
"rfilename": "checkpoint-274/special_tokens_map.json"
},
{
"rfilename": "checkpoint-274/tokenizer.json"
},
{
"rfilename": "checkpoint-274/tokenizer.model"
},
{
"rfilename": "checkpoint-274/tokenizer_config.json"
},
{
"rfilename": "checkpoint-274/trainer_state.json"
},
{
"rfilename": "checkpoint-274/training_args.bin"
},
{
"rfilename": "checkpoint-329/README.md"
},
{
"rfilename": "checkpoint-329/adapter_config.json"
},
{
"rfilename": "checkpoint-329/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-329/optimizer.pt"
},
{
"rfilename": "checkpoint-329/rng_state.pth"
},
{
"rfilename": "checkpoint-329/scheduler.pt"
},
{
"rfilename": "checkpoint-329/special_tokens_map.json"
},
{
"rfilename": "checkpoint-329/tokenizer.json"
},
{
"rfilename": "checkpoint-329/tokenizer.model"
},
{
"rfilename": "checkpoint-329/tokenizer_config.json"
},
{
"rfilename": "checkpoint-329/trainer_state.json"
},
{
"rfilename": "checkpoint-329/training_args.bin"
},
{
"rfilename": "checkpoint-384/README.md"
},
{
"rfilename": "checkpoint-384/adapter_config.json"
},
{
"rfilename": "checkpoint-384/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-384/optimizer.pt"
},
{
"rfilename": "checkpoint-384/rng_state.pth"
},
{
"rfilename": "checkpoint-384/scheduler.pt"
},
{
"rfilename": "checkpoint-384/special_tokens_map.json"
},
{
"rfilename": "checkpoint-384/tokenizer.json"
},
{
"rfilename": "checkpoint-384/tokenizer.model"
},
{
"rfilename": "checkpoint-384/tokenizer_config.json"
},
{
"rfilename": "checkpoint-384/trainer_state.json"
},
{
"rfilename": "checkpoint-384/training_args.bin"
},
{
"rfilename": "checkpoint-439/README.md"
},
{
"rfilename": "checkpoint-439/adapter_config.json"
},
{
"rfilename": "checkpoint-439/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-439/optimizer.pt"
},
{
"rfilename": "checkpoint-439/rng_state.pth"
},
{
"rfilename": "checkpoint-439/scheduler.pt"
},
{
"rfilename": "checkpoint-439/special_tokens_map.json"
},
{
"rfilename": "checkpoint-439/tokenizer.json"
},
{
"rfilename": "checkpoint-439/tokenizer.model"
},
{
"rfilename": "checkpoint-439/tokenizer_config.json"
},
{
"rfilename": "checkpoint-439/trainer_state.json"
},
{
"rfilename": "checkpoint-439/training_args.bin"
},
{
"rfilename": "checkpoint-493/README.md"
},
{
"rfilename": "checkpoint-493/adapter_config.json"
},
{
"rfilename": "checkpoint-493/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-493/optimizer.pt"
},
{
"rfilename": "checkpoint-493/rng_state.pth"
},
{
"rfilename": "checkpoint-493/scheduler.pt"
},
{
"rfilename": "checkpoint-493/special_tokens_map.json"
},
{
"rfilename": "checkpoint-493/tokenizer.json"
},
{
"rfilename": "checkpoint-493/tokenizer.model"
},
{
"rfilename": "checkpoint-493/tokenizer_config.json"
},
{
"rfilename": "checkpoint-493/trainer_state.json"
},
{
"rfilename": "checkpoint-493/training_args.bin"
},
{
"rfilename": "checkpoint-54/README.md"
},
{
"rfilename": "checkpoint-54/adapter_config.json"
},
{
"rfilename": "checkpoint-54/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-54/optimizer.pt"
},
{
"rfilename": "checkpoint-54/rng_state.pth"
},
{
"rfilename": "checkpoint-54/scheduler.pt"
},
{
"rfilename": "checkpoint-54/special_tokens_map.json"
},
{
"rfilename": "checkpoint-54/tokenizer.json"
},
{
"rfilename": "checkpoint-54/tokenizer.model"
},
{
"rfilename": "checkpoint-54/tokenizer_config.json"
},
{
"rfilename": "checkpoint-54/trainer_state.json"
},
{
"rfilename": "checkpoint-54/training_args.bin"
},
{
"rfilename": "checkpoint-540/README.md"
},
{
"rfilename": "checkpoint-540/adapter_config.json"
},
{
"rfilename": "checkpoint-540/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-540/optimizer.pt"
},
{
"rfilename": "checkpoint-540/rng_state.pth"
},
{
"rfilename": "checkpoint-540/scheduler.pt"
},
{
"rfilename": "checkpoint-540/special_tokens_map.json"
},
{
"rfilename": "checkpoint-540/tokenizer.json"
},
{
"rfilename": "checkpoint-540/tokenizer.model"
},
{
"rfilename": "checkpoint-540/tokenizer_config.json"
},
{
"rfilename": "checkpoint-540/trainer_state.json"
},
{
"rfilename": "checkpoint-540/training_args.bin"
},
{
"rfilename": "checkpoint-548/README.md"
},
{
"rfilename": "checkpoint-548/adapter_config.json"
},
{
"rfilename": "checkpoint-548/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-548/optimizer.pt"
},
{
"rfilename": "checkpoint-548/rng_state.pth"
},
{
"rfilename": "checkpoint-548/scheduler.pt"
},
{
"rfilename": "checkpoint-548/special_tokens_map.json"
},
{
"rfilename": "checkpoint-548/tokenizer.json"
},
{
"rfilename": "checkpoint-548/tokenizer.model"
},
{
"rfilename": "checkpoint-548/tokenizer_config.json"
},
{
"rfilename": "checkpoint-548/trainer_state.json"
},
{
"rfilename": "checkpoint-548/training_args.bin"
},
{
"rfilename": "checkpoint-603/README.md"
},
{
"rfilename": "checkpoint-603/adapter_config.json"
},
{
"rfilename": "checkpoint-603/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-603/optimizer.pt"
},
{
"rfilename": "checkpoint-603/rng_state.pth"
},
{
"rfilename": "checkpoint-603/scheduler.pt"
},
{
"rfilename": "checkpoint-603/special_tokens_map.json"
},
{
"rfilename": "checkpoint-603/tokenizer.json"
},
{
"rfilename": "checkpoint-603/tokenizer.model"
},
{
"rfilename": "checkpoint-603/tokenizer_config.json"
},
{
"rfilename": "checkpoint-603/trainer_state.json"
},
{
"rfilename": "checkpoint-603/training_args.bin"
},
{
"rfilename": "checkpoint-658/README.md"
},
{
"rfilename": "checkpoint-658/adapter_config.json"
},
{
"rfilename": "checkpoint-658/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-658/optimizer.pt"
},
{
"rfilename": "checkpoint-658/rng_state.pth"
},
{
"rfilename": "checkpoint-658/scheduler.pt"
},
{
"rfilename": "checkpoint-658/special_tokens_map.json"
},
{
"rfilename": "checkpoint-658/tokenizer.json"
},
{
"rfilename": "checkpoint-658/tokenizer.model"
},
{
"rfilename": "checkpoint-658/tokenizer_config.json"
},
{
"rfilename": "checkpoint-658/trainer_state.json"
},
{
"rfilename": "checkpoint-658/training_args.bin"
},
{
"rfilename": "checkpoint-713/README.md"
},
{
"rfilename": "checkpoint-713/adapter_config.json"
},
{
"rfilename": "checkpoint-713/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-713/optimizer.pt"
},
{
"rfilename": "checkpoint-713/rng_state.pth"
},
{
"rfilename": "checkpoint-713/scheduler.pt"
},
{
"rfilename": "checkpoint-713/special_tokens_map.json"
},
{
"rfilename": "checkpoint-713/tokenizer.json"
},
{
"rfilename": "checkpoint-713/tokenizer.model"
},
{
"rfilename": "checkpoint-713/tokenizer_config.json"
},
{
"rfilename": "checkpoint-713/trainer_state.json"
},
{
"rfilename": "checkpoint-713/training_args.bin"
},
{
"rfilename": "checkpoint-768/README.md"
},
{
"rfilename": "checkpoint-768/adapter_config.json"
},
{
"rfilename": "checkpoint-768/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-768/optimizer.pt"
},
{
"rfilename": "checkpoint-768/rng_state.pth"
},
{
"rfilename": "checkpoint-768/scheduler.pt"
},
{
"rfilename": "checkpoint-768/special_tokens_map.json"
},
{
"rfilename": "checkpoint-768/tokenizer.json"
},
{
"rfilename": "checkpoint-768/tokenizer.model"
},
{
"rfilename": "checkpoint-768/tokenizer_config.json"
},
{
"rfilename": "checkpoint-768/trainer_state.json"
},
{
"rfilename": "checkpoint-768/training_args.bin"
},
{
"rfilename": "checkpoint-823/README.md"
},
{
"rfilename": "checkpoint-823/adapter_config.json"
},
{
"rfilename": "checkpoint-823/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-823/optimizer.pt"
},
{
"rfilename": "checkpoint-823/rng_state.pth"
},
{
"rfilename": "checkpoint-823/scheduler.pt"
},
{
"rfilename": "checkpoint-823/special_tokens_map.json"
},
{
"rfilename": "checkpoint-823/tokenizer.json"
},
{
"rfilename": "checkpoint-823/tokenizer.model"
},
{
"rfilename": "checkpoint-823/tokenizer_config.json"
},
{
"rfilename": "checkpoint-823/trainer_state.json"
},
{
"rfilename": "checkpoint-823/training_args.bin"
},
{
"rfilename": "checkpoint-878/README.md"
},
{
"rfilename": "checkpoint-878/adapter_config.json"
},
{
"rfilename": "checkpoint-878/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-878/optimizer.pt"
},
{
"rfilename": "checkpoint-878/rng_state.pth"
},
{
"rfilename": "checkpoint-878/scheduler.pt"
},
{
"rfilename": "checkpoint-878/special_tokens_map.json"
},
{
"rfilename": "checkpoint-878/tokenizer.json"
},
{
"rfilename": "checkpoint-878/tokenizer.model"
},
{
"rfilename": "checkpoint-878/tokenizer_config.json"
},
{
"rfilename": "checkpoint-878/trainer_state.json"
},
{
"rfilename": "checkpoint-878/training_args.bin"
},
{
"rfilename": "checkpoint-932/README.md"
},
{
"rfilename": "checkpoint-932/adapter_config.json"
},
{
"rfilename": "checkpoint-932/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-932/optimizer.pt"
},
{
"rfilename": "checkpoint-932/rng_state.pth"
},
{
"rfilename": "checkpoint-932/scheduler.pt"
},
{
"rfilename": "checkpoint-932/special_tokens_map.json"
},
{
"rfilename": "checkpoint-932/tokenizer.json"
},
{
"rfilename": "checkpoint-932/tokenizer.model"
},
{
"rfilename": "checkpoint-932/tokenizer_config.json"
},
{
"rfilename": "checkpoint-932/trainer_state.json"
},
{
"rfilename": "checkpoint-932/training_args.bin"
},
{
"rfilename": "checkpoint-987/README.md"
},
{
"rfilename": "checkpoint-987/adapter_config.json"
},
{
"rfilename": "checkpoint-987/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-987/optimizer.pt"
},
{
"rfilename": "checkpoint-987/rng_state.pth"
},
{
"rfilename": "checkpoint-987/scheduler.pt"
},
{
"rfilename": "checkpoint-987/special_tokens_map.json"
},
{
"rfilename": "checkpoint-987/tokenizer.json"
},
{
"rfilename": "checkpoint-987/tokenizer.model"
},
{
"rfilename": "checkpoint-987/tokenizer_config.json"
},
{
"rfilename": "checkpoint-987/trainer_state.json"
},
{
"rfilename": "checkpoint-987/training_args.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 931,212 |
664a648c4a52f105d7c4d915 | VeriUs/VeriUS-LLM-8b-v0.2 | VeriUs | False | not-popular-enough | 2024-05-30T18:47:52.000Z | 3 | false | 8bb368925c58ff2d837c92d25c1cb6ef8200d32d | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|reserved_special_token_250|>"}} | 2,335 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-19T20:43:56.000Z | VeriUs/VeriUS-LLM-8b-v0.2 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00005.safetensors"
},
{
"rfilename": "model-00002-of-00005.safetensors"
},
{
"rfilename": "model-00003-of-00005.safetensors"
},
{
"rfilename": "model-00004-of-00005.safetensors"
},
{
"rfilename": "model-00005-of-00005.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/llama-3-8b-bnb-4bit | 3,018 | text-generation | 3,018 | 0 | 0 | 0 | 0 | 647,361 |
6651b86a07cc2255ea064efc | notlober/llama3-8b-tr | notlober | False | not-popular-enough | 2024-05-25T10:15:16.000Z | 0 | false | 97bbd310fd817d43a992ab54212c86aef272dc9f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|reserved_special_token_250|>"}} | 2,333 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-25T10:07:38.000Z | notlober/llama3-8b-tr | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/llama-3-8b-bnb-4bit | 3,018 | text-generation | 3,018 | 0 | 0 | 0 | 0 | 658,107 |
66293358292184d586b783c1 | johnsnowlabs/JSL-MedLlama-3-8B-v1.0 | johnsnowlabs | False | not-popular-enough | 2024-04-24T16:49:09.000Z | 6 | false | 5d4fbc297ad7eab0c5a19c98d70dd5f6f5b064fe | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | 2,330 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3-8b",
"sft",
"medical",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:cc-by-nc-nd-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-24T16:29:12.000Z | johnsnowlabs/JSL-MedLlama-3-8B-v1.0 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 9 | 0 | 1 | 603,281 |
66b18dc9023357106d85b90f | ValiantLabs/Llama3.1-8B-ShiningValiant2 | ValiantLabs | False | library-not-detected | 2024-09-19T22:34:56.000Z | 13 | false | 13f058dbc1f5d2bd0ac578de2c4f3dde9d98888b | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}} | 2,322 | [
"safetensors",
"llama",
"shining-valiant",
"shining-valiant-2",
"valiant",
"valiant-labs",
"llama-3.1",
"llama-3.1-instruct",
"llama-3.1-instruct-8b",
"llama-3",
"llama-3-instruct",
"llama-3-instruct-8b",
"8b",
"science",
"physics",
"biology",
"chemistry",
"compsci",
"computer-science",
"engineering",
"technical",
"conversational",
"chat",
"instruct",
"text-generation",
"en",
"dataset:sequelbox/Celestia",
"dataset:sequelbox/Supernova",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"region:us"
] | text-generation | null | 2024-08-06T02:43:21.000Z | ValiantLabs/Llama3.1-8B-ShiningValiant2 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00007.safetensors"
},
{
"rfilename": "model-00002-of-00007.safetensors"
},
{
"rfilename": "model-00003-of-00007.safetensors"
},
{
"rfilename": "model-00004-of-00007.safetensors"
},
{
"rfilename": "model-00005-of-00007.safetensors"
},
{
"rfilename": "model-00006-of-00007.safetensors"
},
{
"rfilename": "model-00007-of-00007.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 1 | 0 | 2 | 802,607 |
663eb1f47d29ead2241a1076 | jondurbin/airoboros-dpo-70b-3.3 | jondurbin | False | not-popular-enough | 2024-05-11T08:40:46.000Z | 5 | false | 0c3cd98374de38854830fa4c9c66985f4cd068bb | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 2,322 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"conversational",
"dataset:jondurbin/airoboros-3.2",
"dataset:bluemoon-fandom-1-1-rp-cleaned",
"dataset:boolq",
"dataset:LDJnr/Capybara",
"dataset:jondurbin/cinematika-v0.1",
"dataset:glaiveai/glaive-function-calling-v2",
"dataset:grimulkan/LimaRP-augmented",
"dataset:piqa",
"dataset:Vezora/Tested-22k-Python-Alpaca",
"dataset:mattpscott/airoboros-summarization",
"dataset:unalignment/toxic-dpo-v0.2",
"dataset:allenai/ultrafeedback_binarized_cleaned",
"dataset:argilla/distilabel-intel-orca-dpo-pairs",
"dataset:jondurbin/contextual-dpo-v0.1",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:jondurbin/py-dpo-v0.1",
"dataset:jondurbin/truthy-dpo-v0.1",
"dataset:lmsys/lmsys-chat-1m",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-10T23:47:00.000Z | jondurbin/airoboros-dpo-70b-3.3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00037.safetensors"
},
{
"rfilename": "model-00002-of-00037.safetensors"
},
{
"rfilename": "model-00003-of-00037.safetensors"
},
{
"rfilename": "model-00004-of-00037.safetensors"
},
{
"rfilename": "model-00005-of-00037.safetensors"
},
{
"rfilename": "model-00006-of-00037.safetensors"
},
{
"rfilename": "model-00007-of-00037.safetensors"
},
{
"rfilename": "model-00008-of-00037.safetensors"
},
{
"rfilename": "model-00009-of-00037.safetensors"
},
{
"rfilename": "model-00010-of-00037.safetensors"
},
{
"rfilename": "model-00011-of-00037.safetensors"
},
{
"rfilename": "model-00012-of-00037.safetensors"
},
{
"rfilename": "model-00013-of-00037.safetensors"
},
{
"rfilename": "model-00014-of-00037.safetensors"
},
{
"rfilename": "model-00015-of-00037.safetensors"
},
{
"rfilename": "model-00016-of-00037.safetensors"
},
{
"rfilename": "model-00017-of-00037.safetensors"
},
{
"rfilename": "model-00018-of-00037.safetensors"
},
{
"rfilename": "model-00019-of-00037.safetensors"
},
{
"rfilename": "model-00020-of-00037.safetensors"
},
{
"rfilename": "model-00021-of-00037.safetensors"
},
{
"rfilename": "model-00022-of-00037.safetensors"
},
{
"rfilename": "model-00023-of-00037.safetensors"
},
{
"rfilename": "model-00024-of-00037.safetensors"
},
{
"rfilename": "model-00025-of-00037.safetensors"
},
{
"rfilename": "model-00026-of-00037.safetensors"
},
{
"rfilename": "model-00027-of-00037.safetensors"
},
{
"rfilename": "model-00028-of-00037.safetensors"
},
{
"rfilename": "model-00029-of-00037.safetensors"
},
{
"rfilename": "model-00030-of-00037.safetensors"
},
{
"rfilename": "model-00031-of-00037.safetensors"
},
{
"rfilename": "model-00032-of-00037.safetensors"
},
{
"rfilename": "model-00033-of-00037.safetensors"
},
{
"rfilename": "model-00034-of-00037.safetensors"
},
{
"rfilename": "model-00035-of-00037.safetensors"
},
{
"rfilename": "model-00036-of-00037.safetensors"
},
{
"rfilename": "model-00037-of-00037.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 2 | 632,142 |
6628caaf1f4dd1e3c56f5454 | umarigan/LLama-3-8B-Instruction-tr | umarigan | False | not-popular-enough | 2024-05-15T14:08:18.000Z | 5 | false | 441b454526650b00918f2bc0ee977efb72581f2f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 2,312 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"en",
"tr",
"dataset:umarigan/GPTeacher-General-Instruct-tr",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-24T09:02:39.000Z | umarigan/LLama-3-8B-Instruction-tr | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/llama-3-8b-bnb-4bit | 3,018 | text-generation | 3,018 | 0 | 0 | 0 | 0 | 602,538 |
662b113e329e32634322660e | dfurman/Llama-3-8B-Orpo-v0.1 | dfurman | False | not-popular-enough | 2024-09-17T16:28:49.000Z | 1 | false | 2b49a35092f4be8a554905a8638aa112d294d391 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}} | 2,307 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"orpo",
"llama 3",
"rlhf",
"sft",
"conversational",
"en",
"dataset:mlabonne/orpo-dpo-mix-40k",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-26T02:28:14.000Z | dfurman/Llama-3-8B-Orpo-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 605,836 |
66e16b5628c211c0b2b3c83e | lovellmnot/Llama-3-8B-sft-lora-mydata3prompt-20epoch | lovellmnot | False | pipeline-not-detected | 2024-09-14T07:40:30.000Z | 0 | false | 3535d0c172fe0dd8511130156d1a817b59c4ea3c | {"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}} | 2,301 | [
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"dataset:generator",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | null | peft | 2024-09-11T10:05:10.000Z | lovellmnot/Llama-3-8B-sft-lora-mydata3prompt-20epoch | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "last-checkpoint/README.md"
},
{
"rfilename": "last-checkpoint/adapter_config.json"
},
{
"rfilename": "last-checkpoint/adapter_model.safetensors"
},
{
"rfilename": "last-checkpoint/optimizer.pt"
},
{
"rfilename": "last-checkpoint/rng_state.pth"
},
{
"rfilename": "last-checkpoint/scheduler.pt"
},
{
"rfilename": "last-checkpoint/special_tokens_map.json"
},
{
"rfilename": "last-checkpoint/tokenizer.json"
},
{
"rfilename": "last-checkpoint/tokenizer_config.json"
},
{
"rfilename": "last-checkpoint/trainer_state.json"
},
{
"rfilename": "last-checkpoint/training_args.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 929,349 |
663c7c6807e058aa732cd96e | jondurbin/airoboros-70b-3.3 | jondurbin | False | not-popular-enough | 2024-05-12T10:15:05.000Z | 14 | false | 03a0445fb8d7968a84294a93ec1e17edde323eec | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 2,291 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"conversational",
"dataset:jondurbin/airoboros-3.2",
"dataset:bluemoon-fandom-1-1-rp-cleaned",
"dataset:boolq",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:LDJnr/Capybara",
"dataset:jondurbin/cinematika-v0.1",
"dataset:glaiveai/glaive-function-calling-v2",
"dataset:grimulkan/LimaRP-augmented",
"dataset:piqa",
"dataset:Vezora/Tested-22k-Python-Alpaca",
"dataset:mattpscott/airoboros-summarization",
"dataset:unalignment/toxic-dpo-v0.2",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-09T07:34:00.000Z | jondurbin/airoboros-70b-3.3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00037.safetensors"
},
{
"rfilename": "model-00002-of-00037.safetensors"
},
{
"rfilename": "model-00003-of-00037.safetensors"
},
{
"rfilename": "model-00004-of-00037.safetensors"
},
{
"rfilename": "model-00005-of-00037.safetensors"
},
{
"rfilename": "model-00006-of-00037.safetensors"
},
{
"rfilename": "model-00007-of-00037.safetensors"
},
{
"rfilename": "model-00008-of-00037.safetensors"
},
{
"rfilename": "model-00009-of-00037.safetensors"
},
{
"rfilename": "model-00010-of-00037.safetensors"
},
{
"rfilename": "model-00011-of-00037.safetensors"
},
{
"rfilename": "model-00012-of-00037.safetensors"
},
{
"rfilename": "model-00013-of-00037.safetensors"
},
{
"rfilename": "model-00014-of-00037.safetensors"
},
{
"rfilename": "model-00015-of-00037.safetensors"
},
{
"rfilename": "model-00016-of-00037.safetensors"
},
{
"rfilename": "model-00017-of-00037.safetensors"
},
{
"rfilename": "model-00018-of-00037.safetensors"
},
{
"rfilename": "model-00019-of-00037.safetensors"
},
{
"rfilename": "model-00020-of-00037.safetensors"
},
{
"rfilename": "model-00021-of-00037.safetensors"
},
{
"rfilename": "model-00022-of-00037.safetensors"
},
{
"rfilename": "model-00023-of-00037.safetensors"
},
{
"rfilename": "model-00024-of-00037.safetensors"
},
{
"rfilename": "model-00025-of-00037.safetensors"
},
{
"rfilename": "model-00026-of-00037.safetensors"
},
{
"rfilename": "model-00027-of-00037.safetensors"
},
{
"rfilename": "model-00028-of-00037.safetensors"
},
{
"rfilename": "model-00029-of-00037.safetensors"
},
{
"rfilename": "model-00030-of-00037.safetensors"
},
{
"rfilename": "model-00031-of-00037.safetensors"
},
{
"rfilename": "model-00032-of-00037.safetensors"
},
{
"rfilename": "model-00033-of-00037.safetensors"
},
{
"rfilename": "model-00034-of-00037.safetensors"
},
{
"rfilename": "model-00035-of-00037.safetensors"
},
{
"rfilename": "model-00036-of-00037.safetensors"
},
{
"rfilename": "model-00037-of-00037.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 2 | 629,057 |
637fb52f4611b79277348d11 | kpriyanshu256/gpt-ya2-v2 | kpriyanshu256 | False | not-popular-enough | 2023-09-15T03:05:40.000Z | 0 | false | 5e21761b81fe384a68c670c950de0cabd64900ab | {"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | 2,288 | [
"transformers",
"pytorch",
"tensorboard",
"gpt2",
"text-generation",
"generated_from_trainer",
"dataset:kpriyanshu256/the_verge-linustechtips-two_min",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2022-11-24T18:17:19.000Z | kpriyanshu256/gpt-ya2-v2 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": ".gitignore"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "pytorch_model.bin"
},
{
"rfilename": "runs/Nov24_18-20-11_18a182a10601/1669314161.671456/events.out.tfevents.1669314161.18a182a10601.1202.1"
},
{
"rfilename": "runs/Nov24_18-20-11_18a182a10601/events.out.tfevents.1669314161.18a182a10601.1202.0"
},
{
"rfilename": "runs/Nov24_18-20-11_18a182a10601/events.out.tfevents.1669333034.18a182a10601.1202.2"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | gpt2 | openai-community/gpt2 | 2,686 | text-generation | 2,686 | 0 | 0 | 0 | 0 | 86,033 |
66486cc3fe822b08e61f427c | collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v1 | collaiborateorg | False | not-popular-enough | 2024-09-22T11:24:41.000Z | 4 | false | fce235ed6f6a979cbc6fafa00aa3a5911dd00e46 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 2,286 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"medical",
"Healthcare & Lifesciences",
"BioMed",
"conversational",
"dataset:collaiborateorg/BioMedData",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-18T08:54:27.000Z | collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 2 | 644,969 |
66b310a3262e90e7f15e7651 | ChatGpt1/model | ChatGpt1 | False | pipeline-not-detected | 2024-08-07T06:45:45.000Z | 0 | false | 3fb3c505540299f3f9b4380858e03976d67c1875 | {"model_type": "llama"} | 2,280 | [
"transformers",
"gguf",
"llama",
"text-generation-inference",
"unsloth",
"en",
"base_model:unsloth/llama-3-8b-Instruct-bnb-4bit",
"base_model:quantized:unsloth/llama-3-8b-Instruct-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | transformers | 2024-08-07T06:13:55.000Z | ChatGpt1/model | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "unsloth.F16.gguf"
},
{
"rfilename": "unsloth.Q4_K_M.gguf"
},
{
"rfilename": "unsloth.Q8_0.gguf"
}
] | llama | unsloth/llama-3-8b-Instruct-bnb-4bit | 1,183 | text-generation | 1,183 | 0 | 0 | 0 | 0 | 804,955 |
665633f37491c508a19c1f29 | lighteternal/Llama3-merge-biomed-8b | lighteternal | False | not-popular-enough | 2024-05-30T06:52:13.000Z | 4 | false | 3b97d1fe9bdb9a3c1b32d869ec08cb63945bcf3d | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 2,277 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2311.03099",
"arxiv:2306.01708",
"base_model:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:merge:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:aaditya/Llama3-OpenBioLLM-8B",
"base_model:merge:aaditya/Llama3-OpenBioLLM-8B",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:merge:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-28T19:43:47.000Z | lighteternal/Llama3-merge-biomed-8b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | NousResearch/Hermes-2-Pro-Llama-3-8B | 131 | text-generation | 131 | 0 | 6 | 0 | 0 | 665,816 |
665047a5f88564a89a6b900b | cypienai/cymist-2-v03-SFT | cypienai | False | not-popular-enough | 2024-06-03T16:24:01.000Z | 1 | false | 1c1652647237f57eb0b275edd127053aa65069c5 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{% endif %}{% endfor %}", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,271 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"turkish",
"general tasks",
"RAG",
"SFT",
"text2text-generation",
"tr",
"en",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:finetune:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text2text-generation | transformers | 2024-05-24T07:54:13.000Z | cypienai/cymist-2-v03-SFT | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.3 | 338 | text-generation | 338 | 0 | 0 | 0 | 0 | 656,240 |
66bbc257cfaa68f770042cd6 | datek/Qwen-Qwen1.5-7B-1723581015 | datek | False | pipeline-not-detected | 2024-08-13T20:30:19.000Z | 0 | false | b0972e4980167b7aaad0fd696b74dc69e744155a | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}} | 2,271 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null | peft | 2024-08-13T20:30:15.000Z | datek/Qwen-Qwen1.5-7B-1723581015 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen1.5-7B | 6,515 | text-generation | 6,515 | 0 | 0 | 0 | 0 | 820,221 |
664e3581ca285a85fb52782a | cypienai/cymist-2-v02-SFT | cypienai | False | not-popular-enough | 2024-06-03T16:19:44.000Z | 0 | false | 336fe06722266bff2a8d21260d5a49e0279323c6 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{% endif %}{% endfor %}", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,269 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"turkish",
"general tasks",
"RAG",
"SFT",
"text2text-generation",
"tr",
"en",
"arxiv:1910.09700",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text2text-generation | transformers | 2024-05-22T18:12:17.000Z | cypienai/cymist-2-v02-SFT | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 653,553 |
6642495e511bb1bb8883e0bf | ruslanmv/ai-medical-model-32bit | ruslanmv | False | not-popular-enough | 2024-05-23T11:49:41.000Z | 1 | false | c1cc57388efd0723f653f55a37623878eeacf1c4 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}} | 2,259 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"ruslanmv",
"trl",
"conversational",
"en",
"dataset:ruslanmv/ai-medical-dataset",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-13T17:09:50.000Z | ruslanmv/ai-medical-model-32bit | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "future.jpg"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00001-of-00007.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00007.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00007.safetensors"
},
{
"rfilename": "model-00004-of-00007.safetensors"
},
{
"rfilename": "model-00005-of-00007.safetensors"
},
{
"rfilename": "model-00006-of-00007.safetensors"
},
{
"rfilename": "model-00007-of-00007.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 2 | 0 | 0 | 636,839 |
661441a1e84c106921b38245 | johnsnowlabs/BioLing-7B-Dare | johnsnowlabs | False | not-popular-enough | 2024-04-18T19:26:43.000Z | 1 | false | bd9d49ab8822358a5adf7125845319cd906b3b43 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,249 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merge",
"mergekit",
"lazymergekit",
"BioMistral/BioMistral-7B",
"Nexusflow/Starling-LM-7B-beta",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:merge:BioMistral/BioMistral-7B",
"base_model:Nexusflow/Starling-LM-7B-beta",
"base_model:merge:Nexusflow/Starling-LM-7B-beta",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-08T19:12:33.000Z | johnsnowlabs/BioLing-7B-Dare | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | BioMistral/BioMistral-7B | 124 | text-generation | 124 | 0 | 0 | 0 | 0 | 574,850 |
6551713e8cc59d5b49589b98 | allenai/tulu-2-dpo-7b | allenai | False | not-popular-enough | 2024-05-14T03:06:00.000Z | 17 | false | b57ef95260b6d4e726adf64518af038e5673f126 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"pad_token": null, "bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}"}} | 2,246 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"dataset:allenai/tulu-v2-sft-mixture",
"arxiv:2305.18290",
"arxiv:2311.10702",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2023-11-13T00:43:42.000Z | allenai/tulu-2-dpo-7b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.md"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "llama2_license.txt"
},
{
"rfilename": "notice.txt"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 5 | 5 | 5 | 376,483 |
663365138581c392afae2bff | FairMind/Phi-3-mini-4k-instruct-bnb-4bit-Ita | FairMind | False | not-popular-enough | 2024-05-02T12:01:29.000Z | 0 | false | d6f3e1b61705279c6747031a73e0e49ae6aeefb8 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] in ['user', 'system']) %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif message['role'] == 'assistant' %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,228 | [
"transformers",
"pytorch",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"phi-3",
"phi-3-mini",
"italian",
"conversational",
"it",
"dataset:mchl-labs/stambecco_data_it",
"base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"base_model:finetune:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"license:cc-by-nc-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-02T10:04:03.000Z | FairMind/Phi-3-mini-4k-instruct-bnb-4bit-Ita | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | unsloth/Phi-3-mini-4k-instruct-bnb-4bit | 626 | text-generation | 626 | 0 | 0 | 0 | 0 | 616,588 |
65fbe197187c4ac021e06739 | FinancialSupport/hellfire-2b | FinancialSupport | False | not-popular-enough | 2024-05-27T12:27:51.000Z | 1 | false | 557438aa9f21b00ecc93a7eb5e52077aaf21a8dd | {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "quantization_config": {"load_in_4bit": true, "load_in_8bit": false, "quant_method": "bitsandbytes"}, "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,221 | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:unsloth/gemma-2b-it-bnb-4bit",
"base_model:quantized:unsloth/gemma-2b-it-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | transformers | 2024-03-21T07:28:23.000Z | FinancialSupport/hellfire-2b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | gemma | unsloth/gemma-2b-it-bnb-4bit | 98 | text-generation | 98 | 0 | 0 | 0 | 0 | 545,471 |
65a69fd8e211a95509997569 | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO-GGUF | NousResearch | False | library-not-detected | 2024-02-20T09:17:51.000Z | 64 | false | 39db2cdbfcd893093ae199e6ea24c8ffe0eca60c | null | 2,202 | [
"gguf",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"region:us"
] | null | null | 2024-01-16T15:25:12.000Z | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q2_K.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_L.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_M.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_S.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_0.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_K_M.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_K_S.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_0.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_K_M.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_K_S.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q6_K.gguf"
},
{
"rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null | mistralai/Mixtral-8x7B-v0.1 | 182 | text-generation | 182 | 0 | 0 | 0 | 0 | 455,566 |
66aab5553dbde0b78e316813 | mlc-ai/gemma-2-2b-it-q4f16_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-08-01T14:37:54.000Z | 6 | false | 4f3984195840df4234e4e34c7904e63d40154190 | {"tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,199 | [
"mlc-llm",
"web-llm",
"base_model:google/gemma-2-2b-it",
"base_model:finetune:google/gemma-2-2b-it",
"region:us"
] | null | mlc-llm | 2024-07-31T22:06:13.000Z | mlc-ai/gemma-2-2b-it-q4f16_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_38.bin"
},
{
"rfilename": "params_shard_39.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_40.bin"
},
{
"rfilename": "params_shard_41.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | null | google/gemma-2-2b-it | 265 | text-generation | 265 | 0 | 0 | 0 | 0 | 791,794 |
66e47c15b8911bbcbd2d8a3a | silent666/01-ai-Yi-1.5-9B-1726250005 | silent666 | False | pipeline-not-detected | 2024-09-13T17:55:59.000Z | 0 | false | d20bb06c42453c45dafefd1f98b9d813e2460ff6 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 2,184 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-13T17:53:25.000Z | silent666/01-ai-Yi-1.5-9B-1726250005 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 940,744 |
65e195f69fac7d008ef1c7e4 | HuggingFaceH4/zephyr-7b-gemma-sft-v0.1 | HuggingFaceH4 | False | not-popular-enough | 2024-03-01T16:47:18.000Z | 12 | false | bfef69be8fbbd6a26fd570c337b4b829c24de736 | {"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'user' or messages[0]['role'] == 'system' %}{{ bos_token }}{% endif %}{% for message in messages %}{{ '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n' }}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% elif messages[-1]['role'] == 'assistant' %}{{ eos_token }}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,117 | [
"transformers",
"tensorboard",
"safetensors",
"gemma",
"text-generation",
"alignment-handbook",
"trl",
"sft",
"generated_from_trainer",
"conversational",
"en",
"dataset:HuggingFaceH4/deita-10k-v0-sft",
"base_model:google/gemma-7b",
"base_model:finetune:google/gemma-7b",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-01T08:46:46.000Z | HuggingFaceH4/zephyr-7b-gemma-sft-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "runs/Mar01_08-52-15_ip-26-0-171-168/events.out.tfevents.1709283363.ip-26-0-171-168.1403296.0"
},
{
"rfilename": "runs/Mar01_09-11-13_ip-26-0-164-18/events.out.tfevents.1709284309.ip-26-0-164-18.560442.0"
},
{
"rfilename": "runs/Mar01_09-22-26_ip-26-0-165-24/events.out.tfevents.1709284978.ip-26-0-165-24.1220823.0"
},
{
"rfilename": "runs/Mar01_09-22-26_ip-26-0-165-24/events.out.tfevents.1709290184.ip-26-0-165-24.1220823.1"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] | gemma | google/gemma-7b | 9,278 | text-generation | 9,278 | 1 | 0 | 32 | 4 | 514,772 |
66ba330a32be421cd87d6dac | nbeerbower/mistral-nemo-wissenschaft-12B | nbeerbower | False | not-popular-enough | 2024-09-03T09:11:19.000Z | 3 | false | 2b6df0c588713611c4bbf554504cb865e18ae166 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{%- for message in loop_messages | rejectattr(\"role\", \"equalto\", \"tool\") | rejectattr(\"role\", \"equalto\", \"tool_results\") | selectattr(\"tool_calls\", \"undefined\") %}\n {%- if (message[\"role\"] == \"user\") != (loop.index0 % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n{%- endfor %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS][\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST]\" + system_message + \"\\n\\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST]\" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif message[\"role\"] == \"tool_calls\" or message.tool_calls is defined %}\n {%- if message.tool_calls is defined %}\n {%- set tool_calls = message.tool_calls %}\n {%- else %}\n {%- set tool_calls = message.content %}\n {%- endif %}\n {{- \"[TOOL_CALLS][\" }}\n {%- for tool_call in tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + eos_token }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- message[\"content\"] + eos_token}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS]{\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n", "eos_token": "</s>", "unk_token": "<unk>"}} | 2,087 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"dataset:tasksource/ScienceQA_text_only",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:finetune:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-08-12T16:06:34.000Z | nbeerbower/mistral-nemo-wissenschaft-12B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00005.safetensors"
},
{
"rfilename": "model-00002-of-00005.safetensors"
},
{
"rfilename": "model-00003-of-00005.safetensors"
},
{
"rfilename": "model-00004-of-00005.safetensors"
},
{
"rfilename": "model-00005-of-00005.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-Nemo-Instruct-2407 | 101 | text-generation | 101 | 0 | 7 | 0 | 2 | 816,569 |
65d3a65cea28ba508b5a30c1 | predibase/customer_support | predibase | False | not-popular-enough | 2024-02-26T19:46:49.000Z | 5 | false | 9ced8fd8cfdf0738c88308a3b56463c751e171f2 | {"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}} | 2,003 | [
"peft",
"safetensors",
"text-generation",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:adapter:mistralai/Mistral-7B-v0.1",
"region:us"
] | text-generation | peft | 2024-02-19T19:05:00.000Z | predibase/customer_support | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 499,425 |
66c5592ffaf1c11450cd46f4 | mlc-ai/Phi-3.5-mini-instruct-q4f16_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-08-23T22:58:26.000Z | 2 | false | f7e8c1f37dffe44eb9e081341bce0cae7960905a | {"tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 1,991 | [
"mlc-llm",
"web-llm",
"base_model:microsoft/Phi-3.5-mini-instruct",
"base_model:finetune:microsoft/Phi-3.5-mini-instruct",
"region:us"
] | null | mlc-llm | 2024-08-21T03:04:15.000Z | mlc-ai/Phi-3.5-mini-instruct-q4f16_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_38.bin"
},
{
"rfilename": "params_shard_39.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_40.bin"
},
{
"rfilename": "params_shard_41.bin"
},
{
"rfilename": "params_shard_42.bin"
},
{
"rfilename": "params_shard_43.bin"
},
{
"rfilename": "params_shard_44.bin"
},
{
"rfilename": "params_shard_45.bin"
},
{
"rfilename": "params_shard_46.bin"
},
{
"rfilename": "params_shard_47.bin"
},
{
"rfilename": "params_shard_48.bin"
},
{
"rfilename": "params_shard_49.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_50.bin"
},
{
"rfilename": "params_shard_51.bin"
},
{
"rfilename": "params_shard_52.bin"
},
{
"rfilename": "params_shard_53.bin"
},
{
"rfilename": "params_shard_54.bin"
},
{
"rfilename": "params_shard_55.bin"
},
{
"rfilename": "params_shard_56.bin"
},
{
"rfilename": "params_shard_57.bin"
},
{
"rfilename": "params_shard_58.bin"
},
{
"rfilename": "params_shard_59.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_60.bin"
},
{
"rfilename": "params_shard_61.bin"
},
{
"rfilename": "params_shard_62.bin"
},
{
"rfilename": "params_shard_63.bin"
},
{
"rfilename": "params_shard_64.bin"
},
{
"rfilename": "params_shard_65.bin"
},
{
"rfilename": "params_shard_66.bin"
},
{
"rfilename": "params_shard_67.bin"
},
{
"rfilename": "params_shard_68.bin"
},
{
"rfilename": "params_shard_69.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_70.bin"
},
{
"rfilename": "params_shard_71.bin"
},
{
"rfilename": "params_shard_72.bin"
},
{
"rfilename": "params_shard_73.bin"
},
{
"rfilename": "params_shard_74.bin"
},
{
"rfilename": "params_shard_75.bin"
},
{
"rfilename": "params_shard_76.bin"
},
{
"rfilename": "params_shard_77.bin"
},
{
"rfilename": "params_shard_78.bin"
},
{
"rfilename": "params_shard_79.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_80.bin"
},
{
"rfilename": "params_shard_81.bin"
},
{
"rfilename": "params_shard_82.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | null | microsoft/Phi-3.5-mini-instruct | 147 | text-generation | 147 | 0 | 0 | 0 | 0 | 843,171 |
66ab11019a318d1bc0319f0b | Saxo/Linkbricks-Horizon-AI-Korean-llama-3.1-sft-dpo-8B | Saxo | False | not-popular-enough | 2024-09-02T00:46:24.000Z | 4 | false | 4a543d074904d38c89ea0c5f5c042c55494096a7 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}} | 1,939 | [
"transformers",
"safetensors",
"gguf",
"llama",
"text-generation",
"conversational",
"ko",
"en",
"jp",
"cn",
"dataset:Saxo/ko_cn_translation_tech_social_science_linkbricks_single_dataset",
"dataset:Saxo/ko_jp_translation_tech_social_science_linkbricks_single_dataset",
"dataset:Saxo/en_ko_translation_tech_science_linkbricks_single_dataset_with_prompt_text_huggingface",
"dataset:Saxo/en_ko_translation_social_science_linkbricks_single_dataset_with_prompt_text_huggingface",
"dataset:Saxo/ko_aspect_sentiment_sns_mall_sentiment_linkbricks_single_dataset_with_prompt_text_huggingface",
"dataset:Saxo/ko_summarization_linkbricks_single_dataset_with_prompt_text_huggingface",
"dataset:Saxo/OpenOrca_cleaned_kor_linkbricks_single_dataset_with_prompt_text_huggingface",
"dataset:Saxo/ko_government_qa_total_linkbricks_single_dataset_with_prompt_text_huggingface_sampled",
"dataset:maywell/ko_Ultrafeedback_binarized",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-08-01T04:37:21.000Z | Saxo/Linkbricks-Horizon-AI-Korean-llama-3.1-sft-dpo-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "ggml-model-bf16.gguf"
},
{
"rfilename": "ggml-model-q4_k_m.gguf"
},
{
"rfilename": "ggml-model-q5_k_m.gguf"
},
{
"rfilename": "ggml-model-q6_k.gguf"
},
{
"rfilename": "ggml-model-q8_0.gguf"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 1 | 0 | 1 | 792,355 |
65d3e1472af7a199b8e33df1 | predibase/dbpedia | predibase | False | not-popular-enough | 2024-02-21T19:14:00.000Z | 8 | false | 150202d7bc480c707c10f63e18a27ea742525919 | {"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}} | 1,906 | [
"peft",
"safetensors",
"text-generation",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:adapter:mistralai/Mistral-7B-v0.1",
"region:us"
] | text-generation | peft | 2024-02-19T23:16:23.000Z | predibase/dbpedia | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 499,689 |
66dc5c7e79d36ca08592c520 | gerbyk/Qwen-Qwen1.5-0.5B-1725717630 | gerbyk | False | pipeline-not-detected | 2024-09-07T14:00:31.000Z | 0 | false | 7bff538242d48525b816640d5e4c6640a3fccc1e | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-0.5B", "task_type": "CAUSAL_LM"}} | 1,902 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-0.5B",
"base_model:adapter:Qwen/Qwen1.5-0.5B",
"region:us"
] | null | peft | 2024-09-07T14:00:30.000Z | gerbyk/Qwen-Qwen1.5-0.5B-1725717630 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen1.5-0.5B | 28,947 | text-generation | 28,947 | 0 | 0 | 0 | 0 | 910,551 |
6632e900e97d286a1d74f360 | 12thD/ko-Llama-3-8B-sft-v0.3 | 12thD | False | not-popular-enough | 2024-05-02T01:24:46.000Z | 0 | false | 134a44b329a37805306c77e45e932d839cae8baa | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 1,871 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"korean",
"gemma",
"conversational",
"ko",
"en",
"arxiv:1910.09700",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-02T01:14:40.000Z | 12thD/ko-Llama-3-8B-sft-v0.3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 615,907 |
66e02101e1bfac4b460f103a | HF1BitLLM/Llama3-8B-1.58-100B-tokens | HF1BitLLM | False | not-popular-enough | 2024-09-19T04:09:11.000Z | 43 | false | 5c35ae1f2c622b75a9c28e3603074863d74e4792 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"quant_method": "bitnet"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 1,870 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:2402.17764",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"bitnet",
"region:us"
] | text-generation | transformers | 2024-09-10T10:35:45.000Z | HF1BitLLM/Llama3-8B-1.58-100B-tokens | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 2 | 924,743 |
66cf01c89a7e2c98a06284a3 | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | BAAI | False | library-not-detected | 2024-09-05T02:56:26.000Z | 8 | false | fd6d02d300e3b9015e07c217e26c6f1b4823963a | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] | trim + '\n\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{% endif %}{% for message in messages %}{% if loop.index0 == 0 %}{% set content = system_message + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + content | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<end_of_turn>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 1,859 | [
"safetensors",
"gemma2",
"en",
"dataset:BAAI/Infinity-Instruct",
"base_model:google/gemma-2-9b-it",
"base_model:finetune:google/gemma-2-9b-it",
"region:us"
] | null | null | 2024-08-28T10:54:00.000Z | BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | gemma2 | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 4 | 0 | 2 | 869,266 |
664fc1b1e823e13f74323d39 | solidrust/Mistral-7B-Instruct-v0.3-AWQ | solidrust | False | explicit-opt-out | 2024-09-03T08:50:39.000Z | 2 | false | 95b1295ddd1a8673117cdc7bd2a4da2a457bb3f7 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 1,841 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"4-bit",
"AWQ",
"autotrain_compatible",
"endpoints_compatible",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.3",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.3",
"license:apache-2.0",
"text-generation-inference",
"awq",
"region:us"
] | text-generation | transformers | 2024-05-23T22:22:41.000Z | solidrust/Mistral-7B-Instruct-v0.3-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-Instruct-v0.3 | 320 | text-generation | 320 | 0 | 0 | 0 | 0 | 655,731 |
66cfbe7e8b9d49dccb3c88de | jfranklin-foundry/01-ai-Yi-1.5-9B-Chat-1724890832 | jfranklin-foundry | False | pipeline-not-detected | 2024-08-29T00:19:56.000Z | 0 | false | 3585c30d90852b0135b939801df7f46e146524ba | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}} | 1,818 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null | peft | 2024-08-29T00:19:10.000Z | jfranklin-foundry/01-ai-Yi-1.5-9B-Chat-1724890832 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B-Chat | 134 | text-generation | 134 | 0 | 0 | 0 | 0 | 871,497 |
65771780c793e5b72b0e8d68 | TheBloke/Mixtral-8x7B-v0.1-GPTQ | TheBloke | False | explicit-opt-out | 2023-12-14T14:30:54.000Z | 127 | false | 7d1eb57b65f823458e27509cd0aac7172f54a260 | {"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 1,817 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2023-12-11T14:06:56.000Z | TheBloke/Mixtral-8x7B-v0.1-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mixtral | mistralai/Mixtral-8x7B-v0.1 | 182 | text-generation | 182 | 0 | 0 | 0 | 0 | 414,316 |
64f614a19c2aeb2a3dea1312 | TheBloke/Llama-2-13B-GGUF | TheBloke | False | explicit-opt-out | 2023-09-27T12:47:13.000Z | 61 | false | b106d1c018ac999af9130b83134fb6b7c5331dea | {"model_type": "llama"} | 1,815 | [
"transformers",
"gguf",
"llama",
"facebook",
"meta",
"pytorch",
"llama-2",
"text-generation",
"en",
"arxiv:2307.09288",
"base_model:meta-llama/Llama-2-13b-hf",
"base_model:quantized:meta-llama/Llama-2-13b-hf",
"license:llama2",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-09-04T17:32:17.000Z | TheBloke/Llama-2-13B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "llama-2-13b.Q2_K.gguf"
},
{
"rfilename": "llama-2-13b.Q3_K_L.gguf"
},
{
"rfilename": "llama-2-13b.Q3_K_M.gguf"
},
{
"rfilename": "llama-2-13b.Q3_K_S.gguf"
},
{
"rfilename": "llama-2-13b.Q4_0.gguf"
},
{
"rfilename": "llama-2-13b.Q4_K_M.gguf"
},
{
"rfilename": "llama-2-13b.Q4_K_S.gguf"
},
{
"rfilename": "llama-2-13b.Q5_0.gguf"
},
{
"rfilename": "llama-2-13b.Q5_K_M.gguf"
},
{
"rfilename": "llama-2-13b.Q5_K_S.gguf"
},
{
"rfilename": "llama-2-13b.Q6_K.gguf"
},
{
"rfilename": "llama-2-13b.Q8_0.gguf"
}
] | llama | meta-llama/Llama-2-13b-hf | 180 | text-generation | 180 | 0 | 0 | 0 | 0 | 303,243 |
666154aba52d8a6aa5efa07c | Qwen/Qwen2-7B-Instruct-GPTQ-Int8 | Qwen | False | not-popular-enough | 2024-08-21T10:33:47.000Z | 16 | false | b018e4540ccdb307df57b1a06918be08d29e6464 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 8, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 1,802 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"8-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2024-06-06T06:18:19.000Z | Qwen/Qwen2-7B-Instruct-GPTQ-Int8 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | qwen2 | Qwen/Qwen2-7B-Instruct | 161 | text-generation | 161 | 2 | 0 | 0 | 0 | 685,153 |