_id
stringlengths
24
24
id
stringlengths
14
79
author
stringlengths
3
21
gated
stringclasses
3 values
inference
stringclasses
9 values
lastModified
stringlengths
24
24
likes
int64
0
4.13k
private
bool
1 class
sha
stringlengths
40
40
config
stringlengths
23
5.92k
downloads
int64
1.8k
3.13M
tags
sequencelengths
5
62
pipeline_tag
stringclasses
6 values
library_name
stringclasses
4 values
createdAt
stringlengths
24
24
modelId
stringlengths
14
79
siblings
listlengths
4
285
model_type
stringclasses
13 values
base_model
stringlengths
13
43
base_model_downloads
float64
98
28.9k
base_model_pipeline
stringclasses
1 value
base_model_children_count
float64
98
28.9k
adapter
float64
0
377
merge
float64
0
94
finetune
float64
0
263
quantized
float64
0
182
__index_level_0__
int64
86k
992k
6660ffab694ec45eaa99d251
cognitivecomputations/dolphin-2.9.3-llama-3-8b
cognitivecomputations
False
not-popular-enough
2024-06-06T00:16:45.000Z
9
false
27a8c71b28c8ce94ec57057951abb546d80ab378
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}}
2,445
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-06T00:15:39.000Z
cognitivecomputations/dolphin-2.9.3-llama-3-8b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
1
10
5
684,601
6649cc0c4a52f105d78e938c
Metin/LLaMA-3-8B-Instruct-TR-DPO
Metin
False
not-popular-enough
2024-06-16T09:48:43.000Z
7
false
5a24a2391d417581c30425b4d6053f62cd736dc1
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
2,442
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "tr", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-19T09:53:16.000Z
Metin/LLaMA-3-8B-Instruct-TR-DPO
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "llama.png" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
10
0
646,603
66de9bfe8f6f3907134e55dc
nekokiku/01-ai-Yi-1.5-9B-Chat-1725864958
nekokiku
False
pipeline-not-detected
2024-09-09T07:05:17.000Z
0
false
8978e77d6e990a61cd8e6254bfaa1fbf0877bf3b
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
2,420
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B-Chat", "base_model:adapter:01-ai/Yi-1.5-9B-Chat", "region:us" ]
null
peft
2024-09-09T06:55:58.000Z
nekokiku/01-ai-Yi-1.5-9B-Chat-1725864958
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B-Chat
134
text-generation
134
0
0
0
0
918,544
653f5d1e7bd6a974393bc5fc
FinGPT/fingpt-forecaster_dow30_llama2-7b_lora
FinGPT
False
pipeline-not-detected
2024-06-11T02:42:34.000Z
102
false
69f77190315afdb03a889d89bf2a0f932b311617
{"peft": {"base_model_name_or_path": "base_models/Llama-2-7b-chat-hf", "task_type": "CAUSAL_LM"}}
2,414
[ "peft", "safetensors", "en", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:adapter:meta-llama/Llama-2-7b-chat-hf", "license:apache-2.0", "region:us" ]
null
peft
2023-10-30T07:37:02.000Z
FinGPT/fingpt-forecaster_dow30_llama2-7b_lora
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.bin" }, { "rfilename": "adapter_model.safetensors" } ]
null
meta-llama/Llama-2-7b-chat-hf
1,328
text-generation
1,328
0
0
0
0
361,194
6622b0d72d61edba3dbe3aa5
Mozilla/Meta-Llama-3-8B-Instruct-llamafile
Mozilla
False
library-not-detected
2024-08-19T10:25:36.000Z
51
false
902a5de8cc6c142515435f5009a46df9d9947d0d
null
2,406
[ "llamafile", "llama", "llama-3", "text-generation", "en", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:other", "region:us" ]
text-generation
null
2024-04-19T17:58:47.000Z
Mozilla/Meta-Llama-3-8B-Instruct-llamafile
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.BF16.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.F16.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q2_K.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_L.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_M.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_S.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_0.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_1.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_M.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_S.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_0.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_1.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_M.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_S.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q6_K.llamafile" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q8_0.llamafile" }, { "rfilename": "Meta-Llama-3-Community-License-Agreement.txt" }, { "rfilename": "README.md" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
594,322
64b6ce072a8e3cd59df98e98
TheBloke/Llama-2-7B-Chat-GGML
TheBloke
False
explicit-opt-out
2023-09-27T13:00:17.000Z
843
false
76cd63c351ae389e1d4b91cab2cf470aab11864b
{"model_type": "llama"}
2,399
[ "transformers", "llama", "facebook", "meta", "pytorch", "llama-2", "text-generation", "en", "arxiv:2307.09288", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:finetune:meta-llama/Llama-2-7b-chat-hf", "license:other", "text-generation-inference", "region:us" ]
text-generation
transformers
2023-07-18T17:38:15.000Z
TheBloke/Llama-2-7B-Chat-GGML
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "Notice" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "config.json" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q2_K.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q3_K_L.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q3_K_M.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q3_K_S.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q4_0.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q4_1.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q4_K_M.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q4_K_S.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q5_0.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q5_1.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q5_K_M.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q5_K_S.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q6_K.bin" }, { "rfilename": "llama-2-7b-chat.ggmlv3.q8_0.bin" } ]
llama
meta-llama/Llama-2-7b-chat-hf
1,328
text-generation
1,328
0
0
2
0
247,837
668d255070e29925074009dd
curiositytech/MARS
curiositytech
False
not-popular-enough
2024-07-10T14:18:14.000Z
5
false
29dd2a871c8e1515b2d41e11f29dd0721bfeec46
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
2,393
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "tr", "en", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-09T11:56:00.000Z
curiositytech/MARS
[ { "rfilename": ".gitattributes" }, { "rfilename": "MARS-1.0.png" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00009.safetensors" }, { "rfilename": "model-00002-of-00009.safetensors" }, { "rfilename": "model-00003-of-00009.safetensors" }, { "rfilename": "model-00004-of-00009.safetensors" }, { "rfilename": "model-00005-of-00009.safetensors" }, { "rfilename": "model-00006-of-00009.safetensors" }, { "rfilename": "model-00007-of-00009.safetensors" }, { "rfilename": "model-00008-of-00009.safetensors" }, { "rfilename": "model-00009-of-00009.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
4
3
743,499
66214dadecdd46cf556f7478
MaziyarPanahi/Meta-Llama-3-8B-Instruct-GGUF
MaziyarPanahi
False
explicit-opt-out
2024-04-23T12:55:06.000Z
76
false
4ece958b356f2ec00338e5898ba0a7254d976baf
{"model_type": "mistral"}
2,384
[ "transformers", "gguf", "mistral", "facebook", "meta", "pytorch", "llama", "llama-3", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "16-bit", "GGUF", "text-generation", "en", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "text-generation-inference", "region:us" ]
text-generation
transformers
2024-04-18T16:43:25.000Z
MaziyarPanahi/Meta-Llama-3-8B-Instruct-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Meta-Llama-3-8B-Instruct.IQ1_M.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.IQ1_S.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.IQ2_XS.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.IQ3_XS.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.IQ4_XS.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q2_K.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_L.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_M.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q3_K_S.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_M.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q4_K_S.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_M.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q5_K_S.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q6_K.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.Q8_0.gguf" }, { "rfilename": "Meta-Llama-3-8B-Instruct.fp16.gguf" }, { "rfilename": "README.md" }, { "rfilename": "config.json" } ]
mistral
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
592,378
66a62c1f2756d40a3fbf31a6
mlabonne/FineLlama-3.1-8B
mlabonne
False
not-popular-enough
2024-08-27T17:21:42.000Z
7
false
b5c5e22e496468b59660fc391577eb3c740d22c5
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% for message in messages %}{% if message['from'] == 'human' %}{{'<|im_start|>user\n' + message['value'] + '<|im_end|>\n'}}{% elif message['from'] == 'gpt' %}{{'<|im_start|>assistant\n' + message['value'] + '<|im_end|>\n' }}{% else %}{{ '<|im_start|>system\n' + message['value'] + '<|im_end|>\n' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|finetune_right_pad_id|>", "unk_token": null}}
2,381
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "dataset:mlabonne/FineTome-100k", "base_model:unsloth/Meta-Llama-3.1-8B-bnb-4bit", "base_model:finetune:unsloth/Meta-Llama-3.1-8B-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-28T11:31:43.000Z
mlabonne/FineLlama-3.1-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/Meta-Llama-3.1-8B-bnb-4bit
508
text-generation
508
0
0
0
11
783,566
66a017b16106bf810ef1c0af
Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
Magpie-Align
False
not-popular-enough
2024-08-20T19:10:51.000Z
1
false
7ee0eb6a6c3e055988e6e3f24606c7fe03e5f5ec
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
2,354
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "axolotl", "generated_from_trainer", "conversational", "en", "dataset:Magpie-Align/Magpie-Reasoning-150K", "dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1", "arxiv:2406.08464", "base_model:meta-llama/Llama-3.1-8B", "base_model:finetune:meta-llama/Llama-3.1-8B", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-23T20:50:57.000Z
Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model-00001-of-00004.bin" }, { "rfilename": "pytorch_model-00002-of-00004.bin" }, { "rfilename": "pytorch_model-00003-of-00004.bin" }, { "rfilename": "pytorch_model-00004-of-00004.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
llama
meta-llama/Llama-3.1-8B
254
text-generation
254
0
0
1
2
773,315
66e1d68a5b1fc436d9cde6ca
jfranklin-foundry/01-ai-Yi-1.5-9B-1726076646
jfranklin-foundry
False
pipeline-not-detected
2024-09-11T17:43:40.000Z
0
false
38ed254a11e2ab2d9071cd7e1c074d3b8d7147d2
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
2,354
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-11T17:42:34.000Z
jfranklin-foundry/01-ai-Yi-1.5-9B-1726076646
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "checkpoint-109/README.md" }, { "rfilename": "checkpoint-109/adapter_config.json" }, { "rfilename": "checkpoint-109/adapter_model.safetensors" }, { "rfilename": "checkpoint-109/optimizer.pt" }, { "rfilename": "checkpoint-109/rng_state.pth" }, { "rfilename": "checkpoint-109/scheduler.pt" }, { "rfilename": "checkpoint-109/special_tokens_map.json" }, { "rfilename": "checkpoint-109/tokenizer.json" }, { "rfilename": "checkpoint-109/tokenizer.model" }, { "rfilename": "checkpoint-109/tokenizer_config.json" }, { "rfilename": "checkpoint-109/trainer_state.json" }, { "rfilename": "checkpoint-109/training_args.bin" }, { "rfilename": "checkpoint-164/README.md" }, { "rfilename": "checkpoint-164/adapter_config.json" }, { "rfilename": "checkpoint-164/adapter_model.safetensors" }, { "rfilename": "checkpoint-164/optimizer.pt" }, { "rfilename": "checkpoint-164/rng_state.pth" }, { "rfilename": "checkpoint-164/scheduler.pt" }, { "rfilename": "checkpoint-164/special_tokens_map.json" }, { "rfilename": "checkpoint-164/tokenizer.json" }, { "rfilename": "checkpoint-164/tokenizer.model" }, { "rfilename": "checkpoint-164/tokenizer_config.json" }, { "rfilename": "checkpoint-164/trainer_state.json" }, { "rfilename": "checkpoint-164/training_args.bin" }, { "rfilename": "checkpoint-219/README.md" }, { "rfilename": "checkpoint-219/adapter_config.json" }, { "rfilename": "checkpoint-219/adapter_model.safetensors" }, { "rfilename": "checkpoint-219/optimizer.pt" }, { "rfilename": "checkpoint-219/rng_state.pth" }, { "rfilename": "checkpoint-219/scheduler.pt" }, { "rfilename": "checkpoint-219/special_tokens_map.json" }, { "rfilename": "checkpoint-219/tokenizer.json" }, { "rfilename": "checkpoint-219/tokenizer.model" }, { "rfilename": "checkpoint-219/tokenizer_config.json" }, { "rfilename": "checkpoint-219/trainer_state.json" }, { "rfilename": "checkpoint-219/training_args.bin" }, { "rfilename": "checkpoint-270/README.md" }, { "rfilename": "checkpoint-270/adapter_config.json" }, { "rfilename": "checkpoint-270/adapter_model.safetensors" }, { "rfilename": "checkpoint-270/optimizer.pt" }, { "rfilename": "checkpoint-270/rng_state.pth" }, { "rfilename": "checkpoint-270/scheduler.pt" }, { "rfilename": "checkpoint-270/special_tokens_map.json" }, { "rfilename": "checkpoint-270/tokenizer.json" }, { "rfilename": "checkpoint-270/tokenizer.model" }, { "rfilename": "checkpoint-270/tokenizer_config.json" }, { "rfilename": "checkpoint-270/trainer_state.json" }, { "rfilename": "checkpoint-270/training_args.bin" }, { "rfilename": "checkpoint-274/README.md" }, { "rfilename": "checkpoint-274/adapter_config.json" }, { "rfilename": "checkpoint-274/adapter_model.safetensors" }, { "rfilename": "checkpoint-274/optimizer.pt" }, { "rfilename": "checkpoint-274/rng_state.pth" }, { "rfilename": "checkpoint-274/scheduler.pt" }, { "rfilename": "checkpoint-274/special_tokens_map.json" }, { "rfilename": "checkpoint-274/tokenizer.json" }, { "rfilename": "checkpoint-274/tokenizer.model" }, { "rfilename": "checkpoint-274/tokenizer_config.json" }, { "rfilename": "checkpoint-274/trainer_state.json" }, { "rfilename": "checkpoint-274/training_args.bin" }, { "rfilename": "checkpoint-329/README.md" }, { "rfilename": "checkpoint-329/adapter_config.json" }, { "rfilename": "checkpoint-329/adapter_model.safetensors" }, { "rfilename": "checkpoint-329/optimizer.pt" }, { "rfilename": "checkpoint-329/rng_state.pth" }, { "rfilename": "checkpoint-329/scheduler.pt" }, { "rfilename": "checkpoint-329/special_tokens_map.json" }, { "rfilename": "checkpoint-329/tokenizer.json" }, { "rfilename": "checkpoint-329/tokenizer.model" }, { "rfilename": "checkpoint-329/tokenizer_config.json" }, { "rfilename": "checkpoint-329/trainer_state.json" }, { "rfilename": "checkpoint-329/training_args.bin" }, { "rfilename": "checkpoint-384/README.md" }, { "rfilename": "checkpoint-384/adapter_config.json" }, { "rfilename": "checkpoint-384/adapter_model.safetensors" }, { "rfilename": "checkpoint-384/optimizer.pt" }, { "rfilename": "checkpoint-384/rng_state.pth" }, { "rfilename": "checkpoint-384/scheduler.pt" }, { "rfilename": "checkpoint-384/special_tokens_map.json" }, { "rfilename": "checkpoint-384/tokenizer.json" }, { "rfilename": "checkpoint-384/tokenizer.model" }, { "rfilename": "checkpoint-384/tokenizer_config.json" }, { "rfilename": "checkpoint-384/trainer_state.json" }, { "rfilename": "checkpoint-384/training_args.bin" }, { "rfilename": "checkpoint-439/README.md" }, { "rfilename": "checkpoint-439/adapter_config.json" }, { "rfilename": "checkpoint-439/adapter_model.safetensors" }, { "rfilename": "checkpoint-439/optimizer.pt" }, { "rfilename": "checkpoint-439/rng_state.pth" }, { "rfilename": "checkpoint-439/scheduler.pt" }, { "rfilename": "checkpoint-439/special_tokens_map.json" }, { "rfilename": "checkpoint-439/tokenizer.json" }, { "rfilename": "checkpoint-439/tokenizer.model" }, { "rfilename": "checkpoint-439/tokenizer_config.json" }, { "rfilename": "checkpoint-439/trainer_state.json" }, { "rfilename": "checkpoint-439/training_args.bin" }, { "rfilename": "checkpoint-493/README.md" }, { "rfilename": "checkpoint-493/adapter_config.json" }, { "rfilename": "checkpoint-493/adapter_model.safetensors" }, { "rfilename": "checkpoint-493/optimizer.pt" }, { "rfilename": "checkpoint-493/rng_state.pth" }, { "rfilename": "checkpoint-493/scheduler.pt" }, { "rfilename": "checkpoint-493/special_tokens_map.json" }, { "rfilename": "checkpoint-493/tokenizer.json" }, { "rfilename": "checkpoint-493/tokenizer.model" }, { "rfilename": "checkpoint-493/tokenizer_config.json" }, { "rfilename": "checkpoint-493/trainer_state.json" }, { "rfilename": "checkpoint-493/training_args.bin" }, { "rfilename": "checkpoint-54/README.md" }, { "rfilename": "checkpoint-54/adapter_config.json" }, { "rfilename": "checkpoint-54/adapter_model.safetensors" }, { "rfilename": "checkpoint-54/optimizer.pt" }, { "rfilename": "checkpoint-54/rng_state.pth" }, { "rfilename": "checkpoint-54/scheduler.pt" }, { "rfilename": "checkpoint-54/special_tokens_map.json" }, { "rfilename": "checkpoint-54/tokenizer.json" }, { "rfilename": "checkpoint-54/tokenizer.model" }, { "rfilename": "checkpoint-54/tokenizer_config.json" }, { "rfilename": "checkpoint-54/trainer_state.json" }, { "rfilename": "checkpoint-54/training_args.bin" }, { "rfilename": "checkpoint-540/README.md" }, { "rfilename": "checkpoint-540/adapter_config.json" }, { "rfilename": "checkpoint-540/adapter_model.safetensors" }, { "rfilename": "checkpoint-540/optimizer.pt" }, { "rfilename": "checkpoint-540/rng_state.pth" }, { "rfilename": "checkpoint-540/scheduler.pt" }, { "rfilename": "checkpoint-540/special_tokens_map.json" }, { "rfilename": "checkpoint-540/tokenizer.json" }, { "rfilename": "checkpoint-540/tokenizer.model" }, { "rfilename": "checkpoint-540/tokenizer_config.json" }, { "rfilename": "checkpoint-540/trainer_state.json" }, { "rfilename": "checkpoint-540/training_args.bin" }, { "rfilename": "checkpoint-548/README.md" }, { "rfilename": "checkpoint-548/adapter_config.json" }, { "rfilename": "checkpoint-548/adapter_model.safetensors" }, { "rfilename": "checkpoint-548/optimizer.pt" }, { "rfilename": "checkpoint-548/rng_state.pth" }, { "rfilename": "checkpoint-548/scheduler.pt" }, { "rfilename": "checkpoint-548/special_tokens_map.json" }, { "rfilename": "checkpoint-548/tokenizer.json" }, { "rfilename": "checkpoint-548/tokenizer.model" }, { "rfilename": "checkpoint-548/tokenizer_config.json" }, { "rfilename": "checkpoint-548/trainer_state.json" }, { "rfilename": "checkpoint-548/training_args.bin" }, { "rfilename": "checkpoint-603/README.md" }, { "rfilename": "checkpoint-603/adapter_config.json" }, { "rfilename": "checkpoint-603/adapter_model.safetensors" }, { "rfilename": "checkpoint-603/optimizer.pt" }, { "rfilename": "checkpoint-603/rng_state.pth" }, { "rfilename": "checkpoint-603/scheduler.pt" }, { "rfilename": "checkpoint-603/special_tokens_map.json" }, { "rfilename": "checkpoint-603/tokenizer.json" }, { "rfilename": "checkpoint-603/tokenizer.model" }, { "rfilename": "checkpoint-603/tokenizer_config.json" }, { "rfilename": "checkpoint-603/trainer_state.json" }, { "rfilename": "checkpoint-603/training_args.bin" }, { "rfilename": "checkpoint-658/README.md" }, { "rfilename": "checkpoint-658/adapter_config.json" }, { "rfilename": "checkpoint-658/adapter_model.safetensors" }, { "rfilename": "checkpoint-658/optimizer.pt" }, { "rfilename": "checkpoint-658/rng_state.pth" }, { "rfilename": "checkpoint-658/scheduler.pt" }, { "rfilename": "checkpoint-658/special_tokens_map.json" }, { "rfilename": "checkpoint-658/tokenizer.json" }, { "rfilename": "checkpoint-658/tokenizer.model" }, { "rfilename": "checkpoint-658/tokenizer_config.json" }, { "rfilename": "checkpoint-658/trainer_state.json" }, { "rfilename": "checkpoint-658/training_args.bin" }, { "rfilename": "checkpoint-713/README.md" }, { "rfilename": "checkpoint-713/adapter_config.json" }, { "rfilename": "checkpoint-713/adapter_model.safetensors" }, { "rfilename": "checkpoint-713/optimizer.pt" }, { "rfilename": "checkpoint-713/rng_state.pth" }, { "rfilename": "checkpoint-713/scheduler.pt" }, { "rfilename": "checkpoint-713/special_tokens_map.json" }, { "rfilename": "checkpoint-713/tokenizer.json" }, { "rfilename": "checkpoint-713/tokenizer.model" }, { "rfilename": "checkpoint-713/tokenizer_config.json" }, { "rfilename": "checkpoint-713/trainer_state.json" }, { "rfilename": "checkpoint-713/training_args.bin" }, { "rfilename": "checkpoint-768/README.md" }, { "rfilename": "checkpoint-768/adapter_config.json" }, { "rfilename": "checkpoint-768/adapter_model.safetensors" }, { "rfilename": "checkpoint-768/optimizer.pt" }, { "rfilename": "checkpoint-768/rng_state.pth" }, { "rfilename": "checkpoint-768/scheduler.pt" }, { "rfilename": "checkpoint-768/special_tokens_map.json" }, { "rfilename": "checkpoint-768/tokenizer.json" }, { "rfilename": "checkpoint-768/tokenizer.model" }, { "rfilename": "checkpoint-768/tokenizer_config.json" }, { "rfilename": "checkpoint-768/trainer_state.json" }, { "rfilename": "checkpoint-768/training_args.bin" }, { "rfilename": "checkpoint-823/README.md" }, { "rfilename": "checkpoint-823/adapter_config.json" }, { "rfilename": "checkpoint-823/adapter_model.safetensors" }, { "rfilename": "checkpoint-823/optimizer.pt" }, { "rfilename": "checkpoint-823/rng_state.pth" }, { "rfilename": "checkpoint-823/scheduler.pt" }, { "rfilename": "checkpoint-823/special_tokens_map.json" }, { "rfilename": "checkpoint-823/tokenizer.json" }, { "rfilename": "checkpoint-823/tokenizer.model" }, { "rfilename": "checkpoint-823/tokenizer_config.json" }, { "rfilename": "checkpoint-823/trainer_state.json" }, { "rfilename": "checkpoint-823/training_args.bin" }, { "rfilename": "checkpoint-878/README.md" }, { "rfilename": "checkpoint-878/adapter_config.json" }, { "rfilename": "checkpoint-878/adapter_model.safetensors" }, { "rfilename": "checkpoint-878/optimizer.pt" }, { "rfilename": "checkpoint-878/rng_state.pth" }, { "rfilename": "checkpoint-878/scheduler.pt" }, { "rfilename": "checkpoint-878/special_tokens_map.json" }, { "rfilename": "checkpoint-878/tokenizer.json" }, { "rfilename": "checkpoint-878/tokenizer.model" }, { "rfilename": "checkpoint-878/tokenizer_config.json" }, { "rfilename": "checkpoint-878/trainer_state.json" }, { "rfilename": "checkpoint-878/training_args.bin" }, { "rfilename": "checkpoint-932/README.md" }, { "rfilename": "checkpoint-932/adapter_config.json" }, { "rfilename": "checkpoint-932/adapter_model.safetensors" }, { "rfilename": "checkpoint-932/optimizer.pt" }, { "rfilename": "checkpoint-932/rng_state.pth" }, { "rfilename": "checkpoint-932/scheduler.pt" }, { "rfilename": "checkpoint-932/special_tokens_map.json" }, { "rfilename": "checkpoint-932/tokenizer.json" }, { "rfilename": "checkpoint-932/tokenizer.model" }, { "rfilename": "checkpoint-932/tokenizer_config.json" }, { "rfilename": "checkpoint-932/trainer_state.json" }, { "rfilename": "checkpoint-932/training_args.bin" }, { "rfilename": "checkpoint-987/README.md" }, { "rfilename": "checkpoint-987/adapter_config.json" }, { "rfilename": "checkpoint-987/adapter_model.safetensors" }, { "rfilename": "checkpoint-987/optimizer.pt" }, { "rfilename": "checkpoint-987/rng_state.pth" }, { "rfilename": "checkpoint-987/scheduler.pt" }, { "rfilename": "checkpoint-987/special_tokens_map.json" }, { "rfilename": "checkpoint-987/tokenizer.json" }, { "rfilename": "checkpoint-987/tokenizer.model" }, { "rfilename": "checkpoint-987/tokenizer_config.json" }, { "rfilename": "checkpoint-987/trainer_state.json" }, { "rfilename": "checkpoint-987/training_args.bin" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
931,212
664a648c4a52f105d7c4d915
VeriUs/VeriUS-LLM-8b-v0.2
VeriUs
False
not-popular-enough
2024-05-30T18:47:52.000Z
3
false
8bb368925c58ff2d837c92d25c1cb6ef8200d32d
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|reserved_special_token_250|>"}}
2,335
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-19T20:43:56.000Z
VeriUs/VeriUS-LLM-8b-v0.2
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00005.safetensors" }, { "rfilename": "model-00002-of-00005.safetensors" }, { "rfilename": "model-00003-of-00005.safetensors" }, { "rfilename": "model-00004-of-00005.safetensors" }, { "rfilename": "model-00005-of-00005.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/llama-3-8b-bnb-4bit
3,018
text-generation
3,018
0
0
0
0
647,361
6651b86a07cc2255ea064efc
notlober/llama3-8b-tr
notlober
False
not-popular-enough
2024-05-25T10:15:16.000Z
0
false
97bbd310fd817d43a992ab54212c86aef272dc9f
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|reserved_special_token_250|>"}}
2,333
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-25T10:07:38.000Z
notlober/llama3-8b-tr
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "pytorch_model-00001-of-00004.bin" }, { "rfilename": "pytorch_model-00002-of-00004.bin" }, { "rfilename": "pytorch_model-00003-of-00004.bin" }, { "rfilename": "pytorch_model-00004-of-00004.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/llama-3-8b-bnb-4bit
3,018
text-generation
3,018
0
0
0
0
658,107
66293358292184d586b783c1
johnsnowlabs/JSL-MedLlama-3-8B-v1.0
johnsnowlabs
False
not-popular-enough
2024-04-24T16:49:09.000Z
6
false
5d4fbc297ad7eab0c5a19c98d70dd5f6f5b064fe
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
2,330
[ "transformers", "safetensors", "llama", "text-generation", "llama-3-8b", "sft", "medical", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:cc-by-nc-nd-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-24T16:29:12.000Z
johnsnowlabs/JSL-MedLlama-3-8B-v1.0
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
9
0
1
603,281
66b18dc9023357106d85b90f
ValiantLabs/Llama3.1-8B-ShiningValiant2
ValiantLabs
False
library-not-detected
2024-09-19T22:34:56.000Z
13
false
13f058dbc1f5d2bd0ac578de2c4f3dde9d98888b
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}}
2,322
[ "safetensors", "llama", "shining-valiant", "shining-valiant-2", "valiant", "valiant-labs", "llama-3.1", "llama-3.1-instruct", "llama-3.1-instruct-8b", "llama-3", "llama-3-instruct", "llama-3-instruct-8b", "8b", "science", "physics", "biology", "chemistry", "compsci", "computer-science", "engineering", "technical", "conversational", "chat", "instruct", "text-generation", "en", "dataset:sequelbox/Celestia", "dataset:sequelbox/Supernova", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "region:us" ]
text-generation
null
2024-08-06T02:43:21.000Z
ValiantLabs/Llama3.1-8B-ShiningValiant2
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00007.safetensors" }, { "rfilename": "model-00002-of-00007.safetensors" }, { "rfilename": "model-00003-of-00007.safetensors" }, { "rfilename": "model-00004-of-00007.safetensors" }, { "rfilename": "model-00005-of-00007.safetensors" }, { "rfilename": "model-00006-of-00007.safetensors" }, { "rfilename": "model-00007-of-00007.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
1
0
2
802,607
663eb1f47d29ead2241a1076
jondurbin/airoboros-dpo-70b-3.3
jondurbin
False
not-popular-enough
2024-05-11T08:40:46.000Z
5
false
0c3cd98374de38854830fa4c9c66985f4cd068bb
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
2,322
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "conversational", "dataset:jondurbin/airoboros-3.2", "dataset:bluemoon-fandom-1-1-rp-cleaned", "dataset:boolq", "dataset:LDJnr/Capybara", "dataset:jondurbin/cinematika-v0.1", "dataset:glaiveai/glaive-function-calling-v2", "dataset:grimulkan/LimaRP-augmented", "dataset:piqa", "dataset:Vezora/Tested-22k-Python-Alpaca", "dataset:mattpscott/airoboros-summarization", "dataset:unalignment/toxic-dpo-v0.2", "dataset:allenai/ultrafeedback_binarized_cleaned", "dataset:argilla/distilabel-intel-orca-dpo-pairs", "dataset:jondurbin/contextual-dpo-v0.1", "dataset:jondurbin/gutenberg-dpo-v0.1", "dataset:jondurbin/py-dpo-v0.1", "dataset:jondurbin/truthy-dpo-v0.1", "dataset:lmsys/lmsys-chat-1m", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-10T23:47:00.000Z
jondurbin/airoboros-dpo-70b-3.3
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00037.safetensors" }, { "rfilename": "model-00002-of-00037.safetensors" }, { "rfilename": "model-00003-of-00037.safetensors" }, { "rfilename": "model-00004-of-00037.safetensors" }, { "rfilename": "model-00005-of-00037.safetensors" }, { "rfilename": "model-00006-of-00037.safetensors" }, { "rfilename": "model-00007-of-00037.safetensors" }, { "rfilename": "model-00008-of-00037.safetensors" }, { "rfilename": "model-00009-of-00037.safetensors" }, { "rfilename": "model-00010-of-00037.safetensors" }, { "rfilename": "model-00011-of-00037.safetensors" }, { "rfilename": "model-00012-of-00037.safetensors" }, { "rfilename": "model-00013-of-00037.safetensors" }, { "rfilename": "model-00014-of-00037.safetensors" }, { "rfilename": "model-00015-of-00037.safetensors" }, { "rfilename": "model-00016-of-00037.safetensors" }, { "rfilename": "model-00017-of-00037.safetensors" }, { "rfilename": "model-00018-of-00037.safetensors" }, { "rfilename": "model-00019-of-00037.safetensors" }, { "rfilename": "model-00020-of-00037.safetensors" }, { "rfilename": "model-00021-of-00037.safetensors" }, { "rfilename": "model-00022-of-00037.safetensors" }, { "rfilename": "model-00023-of-00037.safetensors" }, { "rfilename": "model-00024-of-00037.safetensors" }, { "rfilename": "model-00025-of-00037.safetensors" }, { "rfilename": "model-00026-of-00037.safetensors" }, { "rfilename": "model-00027-of-00037.safetensors" }, { "rfilename": "model-00028-of-00037.safetensors" }, { "rfilename": "model-00029-of-00037.safetensors" }, { "rfilename": "model-00030-of-00037.safetensors" }, { "rfilename": "model-00031-of-00037.safetensors" }, { "rfilename": "model-00032-of-00037.safetensors" }, { "rfilename": "model-00033-of-00037.safetensors" }, { "rfilename": "model-00034-of-00037.safetensors" }, { "rfilename": "model-00035-of-00037.safetensors" }, { "rfilename": "model-00036-of-00037.safetensors" }, { "rfilename": "model-00037-of-00037.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
2
632,142
6628caaf1f4dd1e3c56f5454
umarigan/LLama-3-8B-Instruction-tr
umarigan
False
not-popular-enough
2024-05-15T14:08:18.000Z
5
false
441b454526650b00918f2bc0ee977efb72581f2f
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
2,312
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "tr", "dataset:umarigan/GPTeacher-General-Instruct-tr", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-24T09:02:39.000Z
umarigan/LLama-3-8B-Instruction-tr
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "pytorch_model-00001-of-00004.bin" }, { "rfilename": "pytorch_model-00002-of-00004.bin" }, { "rfilename": "pytorch_model-00003-of-00004.bin" }, { "rfilename": "pytorch_model-00004-of-00004.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/llama-3-8b-bnb-4bit
3,018
text-generation
3,018
0
0
0
0
602,538
662b113e329e32634322660e
dfurman/Llama-3-8B-Orpo-v0.1
dfurman
False
not-popular-enough
2024-09-17T16:28:49.000Z
1
false
2b49a35092f4be8a554905a8638aa112d294d391
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}}
2,307
[ "transformers", "safetensors", "llama", "text-generation", "orpo", "llama 3", "rlhf", "sft", "conversational", "en", "dataset:mlabonne/orpo-dpo-mix-40k", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-26T02:28:14.000Z
dfurman/Llama-3-8B-Orpo-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
605,836
66e16b5628c211c0b2b3c83e
lovellmnot/Llama-3-8B-sft-lora-mydata3prompt-20epoch
lovellmnot
False
pipeline-not-detected
2024-09-14T07:40:30.000Z
0
false
3535d0c172fe0dd8511130156d1a817b59c4ea3c
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}}
2,301
[ "peft", "safetensors", "trl", "sft", "generated_from_trainer", "dataset:generator", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "region:us" ]
null
peft
2024-09-11T10:05:10.000Z
lovellmnot/Llama-3-8B-sft-lora-mydata3prompt-20epoch
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "last-checkpoint/README.md" }, { "rfilename": "last-checkpoint/adapter_config.json" }, { "rfilename": "last-checkpoint/adapter_model.safetensors" }, { "rfilename": "last-checkpoint/optimizer.pt" }, { "rfilename": "last-checkpoint/rng_state.pth" }, { "rfilename": "last-checkpoint/scheduler.pt" }, { "rfilename": "last-checkpoint/special_tokens_map.json" }, { "rfilename": "last-checkpoint/tokenizer.json" }, { "rfilename": "last-checkpoint/tokenizer_config.json" }, { "rfilename": "last-checkpoint/trainer_state.json" }, { "rfilename": "last-checkpoint/training_args.bin" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
929,349
663c7c6807e058aa732cd96e
jondurbin/airoboros-70b-3.3
jondurbin
False
not-popular-enough
2024-05-12T10:15:05.000Z
14
false
03a0445fb8d7968a84294a93ec1e17edde323eec
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
2,291
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "conversational", "dataset:jondurbin/airoboros-3.2", "dataset:bluemoon-fandom-1-1-rp-cleaned", "dataset:boolq", "dataset:jondurbin/gutenberg-dpo-v0.1", "dataset:LDJnr/Capybara", "dataset:jondurbin/cinematika-v0.1", "dataset:glaiveai/glaive-function-calling-v2", "dataset:grimulkan/LimaRP-augmented", "dataset:piqa", "dataset:Vezora/Tested-22k-Python-Alpaca", "dataset:mattpscott/airoboros-summarization", "dataset:unalignment/toxic-dpo-v0.2", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-09T07:34:00.000Z
jondurbin/airoboros-70b-3.3
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00037.safetensors" }, { "rfilename": "model-00002-of-00037.safetensors" }, { "rfilename": "model-00003-of-00037.safetensors" }, { "rfilename": "model-00004-of-00037.safetensors" }, { "rfilename": "model-00005-of-00037.safetensors" }, { "rfilename": "model-00006-of-00037.safetensors" }, { "rfilename": "model-00007-of-00037.safetensors" }, { "rfilename": "model-00008-of-00037.safetensors" }, { "rfilename": "model-00009-of-00037.safetensors" }, { "rfilename": "model-00010-of-00037.safetensors" }, { "rfilename": "model-00011-of-00037.safetensors" }, { "rfilename": "model-00012-of-00037.safetensors" }, { "rfilename": "model-00013-of-00037.safetensors" }, { "rfilename": "model-00014-of-00037.safetensors" }, { "rfilename": "model-00015-of-00037.safetensors" }, { "rfilename": "model-00016-of-00037.safetensors" }, { "rfilename": "model-00017-of-00037.safetensors" }, { "rfilename": "model-00018-of-00037.safetensors" }, { "rfilename": "model-00019-of-00037.safetensors" }, { "rfilename": "model-00020-of-00037.safetensors" }, { "rfilename": "model-00021-of-00037.safetensors" }, { "rfilename": "model-00022-of-00037.safetensors" }, { "rfilename": "model-00023-of-00037.safetensors" }, { "rfilename": "model-00024-of-00037.safetensors" }, { "rfilename": "model-00025-of-00037.safetensors" }, { "rfilename": "model-00026-of-00037.safetensors" }, { "rfilename": "model-00027-of-00037.safetensors" }, { "rfilename": "model-00028-of-00037.safetensors" }, { "rfilename": "model-00029-of-00037.safetensors" }, { "rfilename": "model-00030-of-00037.safetensors" }, { "rfilename": "model-00031-of-00037.safetensors" }, { "rfilename": "model-00032-of-00037.safetensors" }, { "rfilename": "model-00033-of-00037.safetensors" }, { "rfilename": "model-00034-of-00037.safetensors" }, { "rfilename": "model-00035-of-00037.safetensors" }, { "rfilename": "model-00036-of-00037.safetensors" }, { "rfilename": "model-00037-of-00037.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
2
629,057
637fb52f4611b79277348d11
kpriyanshu256/gpt-ya2-v2
kpriyanshu256
False
not-popular-enough
2023-09-15T03:05:40.000Z
0
false
5e21761b81fe384a68c670c950de0cabd64900ab
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
2,288
[ "transformers", "pytorch", "tensorboard", "gpt2", "text-generation", "generated_from_trainer", "dataset:kpriyanshu256/the_verge-linustechtips-two_min", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "license:mit", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2022-11-24T18:17:19.000Z
kpriyanshu256/gpt-ya2-v2
[ { "rfilename": ".gitattributes" }, { "rfilename": ".gitignore" }, { "rfilename": "README.md" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "eval_results.json" }, { "rfilename": "merges.txt" }, { "rfilename": "pytorch_model.bin" }, { "rfilename": "runs/Nov24_18-20-11_18a182a10601/1669314161.671456/events.out.tfevents.1669314161.18a182a10601.1202.1" }, { "rfilename": "runs/Nov24_18-20-11_18a182a10601/events.out.tfevents.1669314161.18a182a10601.1202.0" }, { "rfilename": "runs/Nov24_18-20-11_18a182a10601/events.out.tfevents.1669333034.18a182a10601.1202.2" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "train_results.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
gpt2
openai-community/gpt2
2,686
text-generation
2,686
0
0
0
0
86,033
66486cc3fe822b08e61f427c
collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v1
collaiborateorg
False
not-popular-enough
2024-09-22T11:24:41.000Z
4
false
fce235ed6f6a979cbc6fafa00aa3a5911dd00e46
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
2,286
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "medical", "Healthcare & Lifesciences", "BioMed", "conversational", "dataset:collaiborateorg/BioMedData", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-18T08:54:27.000Z
collaiborateorg/Collaiborator-MEDLLM-Llama-3-8B-v1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
2
644,969
66b310a3262e90e7f15e7651
ChatGpt1/model
ChatGpt1
False
pipeline-not-detected
2024-08-07T06:45:45.000Z
0
false
3fb3c505540299f3f9b4380858e03976d67c1875
{"model_type": "llama"}
2,280
[ "transformers", "gguf", "llama", "text-generation-inference", "unsloth", "en", "base_model:unsloth/llama-3-8b-Instruct-bnb-4bit", "base_model:quantized:unsloth/llama-3-8b-Instruct-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
transformers
2024-08-07T06:13:55.000Z
ChatGpt1/model
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "unsloth.F16.gguf" }, { "rfilename": "unsloth.Q4_K_M.gguf" }, { "rfilename": "unsloth.Q8_0.gguf" } ]
llama
unsloth/llama-3-8b-Instruct-bnb-4bit
1,183
text-generation
1,183
0
0
0
0
804,955
665633f37491c508a19c1f29
lighteternal/Llama3-merge-biomed-8b
lighteternal
False
not-popular-enough
2024-05-30T06:52:13.000Z
4
false
3b97d1fe9bdb9a3c1b32d869ec08cb63945bcf3d
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
2,277
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "conversational", "arxiv:2311.03099", "arxiv:2306.01708", "base_model:NousResearch/Hermes-2-Pro-Llama-3-8B", "base_model:merge:NousResearch/Hermes-2-Pro-Llama-3-8B", "base_model:aaditya/Llama3-OpenBioLLM-8B", "base_model:merge:aaditya/Llama3-OpenBioLLM-8B", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:merge:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-28T19:43:47.000Z
lighteternal/Llama3-merge-biomed-8b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
NousResearch/Hermes-2-Pro-Llama-3-8B
131
text-generation
131
0
6
0
0
665,816
665047a5f88564a89a6b900b
cypienai/cymist-2-v03-SFT
cypienai
False
not-popular-enough
2024-06-03T16:24:01.000Z
1
false
1c1652647237f57eb0b275edd127053aa65069c5
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{% endif %}{% endfor %}", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,271
[ "transformers", "safetensors", "mistral", "text-generation", "turkish", "general tasks", "RAG", "SFT", "text2text-generation", "tr", "en", "arxiv:1910.09700", "base_model:mistralai/Mistral-7B-v0.3", "base_model:finetune:mistralai/Mistral-7B-v0.3", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
transformers
2024-05-24T07:54:13.000Z
cypienai/cymist-2-v03-SFT
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.3
338
text-generation
338
0
0
0
0
656,240
66bbc257cfaa68f770042cd6
datek/Qwen-Qwen1.5-7B-1723581015
datek
False
pipeline-not-detected
2024-08-13T20:30:19.000Z
0
false
b0972e4980167b7aaad0fd696b74dc69e744155a
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}}
2,271
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-7B", "base_model:adapter:Qwen/Qwen1.5-7B", "region:us" ]
null
peft
2024-08-13T20:30:15.000Z
datek/Qwen-Qwen1.5-7B-1723581015
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "added_tokens.json" }, { "rfilename": "merges.txt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen1.5-7B
6,515
text-generation
6,515
0
0
0
0
820,221
664e3581ca285a85fb52782a
cypienai/cymist-2-v02-SFT
cypienai
False
not-popular-enough
2024-06-03T16:19:44.000Z
0
false
336fe06722266bff2a8d21260d5a49e0279323c6
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{% endif %}{% endfor %}", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,269
[ "transformers", "safetensors", "mistral", "text-generation", "turkish", "general tasks", "RAG", "SFT", "text2text-generation", "tr", "en", "arxiv:1910.09700", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text2text-generation
transformers
2024-05-22T18:12:17.000Z
cypienai/cymist-2-v02-SFT
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
653,553
6642495e511bb1bb8883e0bf
ruslanmv/ai-medical-model-32bit
ruslanmv
False
not-popular-enough
2024-05-23T11:49:41.000Z
1
false
c1cc57388efd0723f653f55a37623878eeacf1c4
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}}
2,259
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "ruslanmv", "trl", "conversational", "en", "dataset:ruslanmv/ai-medical-dataset", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-13T17:09:50.000Z
ruslanmv/ai-medical-model-32bit
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "future.jpg" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00001-of-00007.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00002-of-00007.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model-00003-of-00007.safetensors" }, { "rfilename": "model-00004-of-00007.safetensors" }, { "rfilename": "model-00005-of-00007.safetensors" }, { "rfilename": "model-00006-of-00007.safetensors" }, { "rfilename": "model-00007-of-00007.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
2
0
0
636,839
661441a1e84c106921b38245
johnsnowlabs/BioLing-7B-Dare
johnsnowlabs
False
not-popular-enough
2024-04-18T19:26:43.000Z
1
false
bd9d49ab8822358a5adf7125845319cd906b3b43
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
2,249
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "BioMistral/BioMistral-7B", "Nexusflow/Starling-LM-7B-beta", "conversational", "base_model:BioMistral/BioMistral-7B", "base_model:merge:BioMistral/BioMistral-7B", "base_model:Nexusflow/Starling-LM-7B-beta", "base_model:merge:Nexusflow/Starling-LM-7B-beta", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-08T19:12:33.000Z
johnsnowlabs/BioLing-7B-Dare
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
BioMistral/BioMistral-7B
124
text-generation
124
0
0
0
0
574,850
6551713e8cc59d5b49589b98
allenai/tulu-2-dpo-7b
allenai
False
not-popular-enough
2024-05-14T03:06:00.000Z
17
false
b57ef95260b6d4e726adf64518af038e5673f126
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"pad_token": null, "bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": true, "rstrip": false, "single_word": false}, "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}"}}
2,246
[ "transformers", "pytorch", "llama", "text-generation", "conversational", "en", "dataset:HuggingFaceH4/ultrafeedback_binarized", "dataset:allenai/tulu-v2-sft-mixture", "arxiv:2305.18290", "arxiv:2311.10702", "base_model:meta-llama/Llama-2-7b-hf", "base_model:finetune:meta-llama/Llama-2-7b-hf", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2023-11-13T00:43:42.000Z
allenai/tulu-2-dpo-7b
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.md" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "llama2_license.txt" }, { "rfilename": "notice.txt" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-2-7b-hf
1,595
text-generation
1,595
0
5
5
5
376,483
663365138581c392afae2bff
FairMind/Phi-3-mini-4k-instruct-bnb-4bit-Ita
FairMind
False
not-popular-enough
2024-05-02T12:01:29.000Z
0
false
d6f3e1b61705279c6747031a73e0e49ae6aeefb8
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] in ['user', 'system']) %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif message['role'] == 'assistant' %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,228
[ "transformers", "pytorch", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "phi-3", "phi-3-mini", "italian", "conversational", "it", "dataset:mchl-labs/stambecco_data_it", "base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit", "base_model:finetune:unsloth/Phi-3-mini-4k-instruct-bnb-4bit", "license:cc-by-nc-sa-4.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-02T10:04:03.000Z
FairMind/Phi-3-mini-4k-instruct-bnb-4bit-Ita
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
unsloth/Phi-3-mini-4k-instruct-bnb-4bit
626
text-generation
626
0
0
0
0
616,588
65fbe197187c4ac021e06739
FinancialSupport/hellfire-2b
FinancialSupport
False
not-popular-enough
2024-05-27T12:27:51.000Z
1
false
557438aa9f21b00ecc93a7eb5e52077aaf21a8dd
{"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "quantization_config": {"load_in_4bit": true, "load_in_8bit": false, "quant_method": "bitsandbytes"}, "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,221
[ "transformers", "safetensors", "gemma", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:unsloth/gemma-2b-it-bnb-4bit", "base_model:quantized:unsloth/gemma-2b-it-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
transformers
2024-03-21T07:28:23.000Z
FinancialSupport/hellfire-2b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
gemma
unsloth/gemma-2b-it-bnb-4bit
98
text-generation
98
0
0
0
0
545,471
65a69fd8e211a95509997569
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO-GGUF
NousResearch
False
library-not-detected
2024-02-20T09:17:51.000Z
64
false
39db2cdbfcd893093ae199e6ea24c8ffe0eca60c
null
2,202
[ "gguf", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "en", "dataset:teknium/OpenHermes-2.5", "base_model:mistralai/Mixtral-8x7B-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-v0.1", "license:apache-2.0", "region:us" ]
null
null
2024-01-16T15:25:12.000Z
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q2_K.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_L.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_M.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q3_K_S.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_0.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_K_M.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q4_K_S.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_0.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_K_M.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q5_K_S.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q6_K.gguf" }, { "rfilename": "Nous-Hermes-2-Mixtral-8x7B-DPO.Q8_0.gguf" }, { "rfilename": "README.md" } ]
null
mistralai/Mixtral-8x7B-v0.1
182
text-generation
182
0
0
0
0
455,566
66aab5553dbde0b78e316813
mlc-ai/gemma-2-2b-it-q4f16_1-MLC
mlc-ai
False
pipeline-not-detected
2024-08-01T14:37:54.000Z
6
false
4f3984195840df4234e4e34c7904e63d40154190
{"tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,199
[ "mlc-llm", "web-llm", "base_model:google/gemma-2-2b-it", "base_model:finetune:google/gemma-2-2b-it", "region:us" ]
null
mlc-llm
2024-07-31T22:06:13.000Z
mlc-ai/gemma-2-2b-it-q4f16_1-MLC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "mlc-chat-config.json" }, { "rfilename": "ndarray-cache.json" }, { "rfilename": "params_shard_0.bin" }, { "rfilename": "params_shard_1.bin" }, { "rfilename": "params_shard_10.bin" }, { "rfilename": "params_shard_11.bin" }, { "rfilename": "params_shard_12.bin" }, { "rfilename": "params_shard_13.bin" }, { "rfilename": "params_shard_14.bin" }, { "rfilename": "params_shard_15.bin" }, { "rfilename": "params_shard_16.bin" }, { "rfilename": "params_shard_17.bin" }, { "rfilename": "params_shard_18.bin" }, { "rfilename": "params_shard_19.bin" }, { "rfilename": "params_shard_2.bin" }, { "rfilename": "params_shard_20.bin" }, { "rfilename": "params_shard_21.bin" }, { "rfilename": "params_shard_22.bin" }, { "rfilename": "params_shard_23.bin" }, { "rfilename": "params_shard_24.bin" }, { "rfilename": "params_shard_25.bin" }, { "rfilename": "params_shard_26.bin" }, { "rfilename": "params_shard_27.bin" }, { "rfilename": "params_shard_28.bin" }, { "rfilename": "params_shard_29.bin" }, { "rfilename": "params_shard_3.bin" }, { "rfilename": "params_shard_30.bin" }, { "rfilename": "params_shard_31.bin" }, { "rfilename": "params_shard_32.bin" }, { "rfilename": "params_shard_33.bin" }, { "rfilename": "params_shard_34.bin" }, { "rfilename": "params_shard_35.bin" }, { "rfilename": "params_shard_36.bin" }, { "rfilename": "params_shard_37.bin" }, { "rfilename": "params_shard_38.bin" }, { "rfilename": "params_shard_39.bin" }, { "rfilename": "params_shard_4.bin" }, { "rfilename": "params_shard_40.bin" }, { "rfilename": "params_shard_41.bin" }, { "rfilename": "params_shard_5.bin" }, { "rfilename": "params_shard_6.bin" }, { "rfilename": "params_shard_7.bin" }, { "rfilename": "params_shard_8.bin" }, { "rfilename": "params_shard_9.bin" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
null
google/gemma-2-2b-it
265
text-generation
265
0
0
0
0
791,794
66e47c15b8911bbcbd2d8a3a
silent666/01-ai-Yi-1.5-9B-1726250005
silent666
False
pipeline-not-detected
2024-09-13T17:55:59.000Z
0
false
d20bb06c42453c45dafefd1f98b9d813e2460ff6
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
2,184
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-13T17:53:25.000Z
silent666/01-ai-Yi-1.5-9B-1726250005
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
940,744
65e195f69fac7d008ef1c7e4
HuggingFaceH4/zephyr-7b-gemma-sft-v0.1
HuggingFaceH4
False
not-popular-enough
2024-03-01T16:47:18.000Z
12
false
bfef69be8fbbd6a26fd570c337b4b829c24de736
{"architectures": ["GemmaForCausalLM"], "model_type": "gemma", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'user' or messages[0]['role'] == 'system' %}{{ bos_token }}{% endif %}{% for message in messages %}{{ '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n' }}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% elif messages[-1]['role'] == 'assistant' %}{{ eos_token }}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,117
[ "transformers", "tensorboard", "safetensors", "gemma", "text-generation", "alignment-handbook", "trl", "sft", "generated_from_trainer", "conversational", "en", "dataset:HuggingFaceH4/deita-10k-v0-sft", "base_model:google/gemma-7b", "base_model:finetune:google/gemma-7b", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-01T08:46:46.000Z
HuggingFaceH4/zephyr-7b-gemma-sft-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "eval_results.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "runs/Mar01_08-52-15_ip-26-0-171-168/events.out.tfevents.1709283363.ip-26-0-171-168.1403296.0" }, { "rfilename": "runs/Mar01_09-11-13_ip-26-0-164-18/events.out.tfevents.1709284309.ip-26-0-164-18.560442.0" }, { "rfilename": "runs/Mar01_09-22-26_ip-26-0-165-24/events.out.tfevents.1709284978.ip-26-0-165-24.1220823.0" }, { "rfilename": "runs/Mar01_09-22-26_ip-26-0-165-24/events.out.tfevents.1709290184.ip-26-0-165-24.1220823.1" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "train_results.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" } ]
gemma
google/gemma-7b
9,278
text-generation
9,278
1
0
32
4
514,772
66ba330a32be421cd87d6dac
nbeerbower/mistral-nemo-wissenschaft-12B
nbeerbower
False
not-popular-enough
2024-09-03T09:11:19.000Z
3
false
2b6df0c588713611c4bbf554504cb865e18ae166
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{%- if messages[0][\"role\"] == \"system\" %}\n {%- set system_message = messages[0][\"content\"] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n{%- set user_messages = loop_messages | selectattr(\"role\", \"equalto\", \"user\") | list %}\n\n{%- for message in loop_messages | rejectattr(\"role\", \"equalto\", \"tool\") | rejectattr(\"role\", \"equalto\", \"tool_results\") | selectattr(\"tool_calls\", \"undefined\") %}\n {%- if (message[\"role\"] == \"user\") != (loop.index0 % 2 == 0) %}\n {{- raise_exception(\"After the optional system message, conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif %}\n{%- endfor %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if message[\"role\"] == \"user\" %}\n {%- if tools is not none and (message == user_messages[-1]) %}\n {{- \"[AVAILABLE_TOOLS][\" }}\n {%- for tool in tools %}\n {%- set tool = tool.function %}\n {{- '{\"type\": \"function\", \"function\": {' }}\n {%- for key, val in tool.items() if key != \"return\" %}\n {%- if val is string %}\n {{- '\"' + key + '\": \"' + val + '\"' }}\n {%- else %}\n {{- '\"' + key + '\": ' + val|tojson }}\n {%- endif %}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \"}}\" }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" }}\n {%- endif %}\n {%- endfor %}\n {{- \"[/AVAILABLE_TOOLS]\" }}\n {%- endif %}\n {%- if loop.last and system_message is defined %}\n {{- \"[INST]\" + system_message + \"\\n\\n\" + message[\"content\"] + \"[/INST]\" }}\n {%- else %}\n {{- \"[INST]\" + message[\"content\"] + \"[/INST]\" }}\n {%- endif %}\n {%- elif message[\"role\"] == \"tool_calls\" or message.tool_calls is defined %}\n {%- if message.tool_calls is defined %}\n {%- set tool_calls = message.tool_calls %}\n {%- else %}\n {%- set tool_calls = message.content %}\n {%- endif %}\n {{- \"[TOOL_CALLS][\" }}\n {%- for tool_call in tool_calls %}\n {%- set out = tool_call.function|tojson %}\n {{- out[:-1] }}\n {%- if not tool_call.id is defined or tool_call.id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- ', \"id\": \"' + tool_call.id + '\"}' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- else %}\n {{- \"]\" + eos_token }}\n {%- endif %}\n {%- endfor %}\n {%- elif message[\"role\"] == \"assistant\" %}\n {{- message[\"content\"] + eos_token}}\n {%- elif message[\"role\"] == \"tool_results\" or message[\"role\"] == \"tool\" %}\n {%- if message.content is defined and message.content.content is defined %}\n {%- set content = message.content.content %}\n {%- else %}\n {%- set content = message.content %}\n {%- endif %}\n {{- '[TOOL_RESULTS]{\"content\": ' + content|string + \", \" }}\n {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}\n {{- raise_exception(\"Tool call IDs should be alphanumeric strings with length 9!\") }}\n {%- endif %}\n {{- '\"call_id\": \"' + message.tool_call_id + '\"}[/TOOL_RESULTS]' }}\n {%- else %}\n {{- raise_exception(\"Only user and assistant roles are supported, with the exception of an initial optional system message!\") }}\n {%- endif %}\n{%- endfor %}\n", "eos_token": "</s>", "unk_token": "<unk>"}}
2,087
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "dataset:tasksource/ScienceQA_text_only", "base_model:mistralai/Mistral-Nemo-Instruct-2407", "base_model:finetune:mistralai/Mistral-Nemo-Instruct-2407", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-08-12T16:06:34.000Z
nbeerbower/mistral-nemo-wissenschaft-12B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00005.safetensors" }, { "rfilename": "model-00002-of-00005.safetensors" }, { "rfilename": "model-00003-of-00005.safetensors" }, { "rfilename": "model-00004-of-00005.safetensors" }, { "rfilename": "model-00005-of-00005.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-Nemo-Instruct-2407
101
text-generation
101
0
7
0
2
816,569
65d3a65cea28ba508b5a30c1
predibase/customer_support
predibase
False
not-popular-enough
2024-02-26T19:46:49.000Z
5
false
9ced8fd8cfdf0738c88308a3b56463c751e171f2
{"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}}
2,003
[ "peft", "safetensors", "text-generation", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "region:us" ]
text-generation
peft
2024-02-19T19:05:00.000Z
predibase/customer_support
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" } ]
null
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
499,425
66c5592ffaf1c11450cd46f4
mlc-ai/Phi-3.5-mini-instruct-q4f16_1-MLC
mlc-ai
False
pipeline-not-detected
2024-08-23T22:58:26.000Z
2
false
f7e8c1f37dffe44eb9e081341bce0cae7960905a
{"tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
1,991
[ "mlc-llm", "web-llm", "base_model:microsoft/Phi-3.5-mini-instruct", "base_model:finetune:microsoft/Phi-3.5-mini-instruct", "region:us" ]
null
mlc-llm
2024-08-21T03:04:15.000Z
mlc-ai/Phi-3.5-mini-instruct-q4f16_1-MLC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "mlc-chat-config.json" }, { "rfilename": "ndarray-cache.json" }, { "rfilename": "params_shard_0.bin" }, { "rfilename": "params_shard_1.bin" }, { "rfilename": "params_shard_10.bin" }, { "rfilename": "params_shard_11.bin" }, { "rfilename": "params_shard_12.bin" }, { "rfilename": "params_shard_13.bin" }, { "rfilename": "params_shard_14.bin" }, { "rfilename": "params_shard_15.bin" }, { "rfilename": "params_shard_16.bin" }, { "rfilename": "params_shard_17.bin" }, { "rfilename": "params_shard_18.bin" }, { "rfilename": "params_shard_19.bin" }, { "rfilename": "params_shard_2.bin" }, { "rfilename": "params_shard_20.bin" }, { "rfilename": "params_shard_21.bin" }, { "rfilename": "params_shard_22.bin" }, { "rfilename": "params_shard_23.bin" }, { "rfilename": "params_shard_24.bin" }, { "rfilename": "params_shard_25.bin" }, { "rfilename": "params_shard_26.bin" }, { "rfilename": "params_shard_27.bin" }, { "rfilename": "params_shard_28.bin" }, { "rfilename": "params_shard_29.bin" }, { "rfilename": "params_shard_3.bin" }, { "rfilename": "params_shard_30.bin" }, { "rfilename": "params_shard_31.bin" }, { "rfilename": "params_shard_32.bin" }, { "rfilename": "params_shard_33.bin" }, { "rfilename": "params_shard_34.bin" }, { "rfilename": "params_shard_35.bin" }, { "rfilename": "params_shard_36.bin" }, { "rfilename": "params_shard_37.bin" }, { "rfilename": "params_shard_38.bin" }, { "rfilename": "params_shard_39.bin" }, { "rfilename": "params_shard_4.bin" }, { "rfilename": "params_shard_40.bin" }, { "rfilename": "params_shard_41.bin" }, { "rfilename": "params_shard_42.bin" }, { "rfilename": "params_shard_43.bin" }, { "rfilename": "params_shard_44.bin" }, { "rfilename": "params_shard_45.bin" }, { "rfilename": "params_shard_46.bin" }, { "rfilename": "params_shard_47.bin" }, { "rfilename": "params_shard_48.bin" }, { "rfilename": "params_shard_49.bin" }, { "rfilename": "params_shard_5.bin" }, { "rfilename": "params_shard_50.bin" }, { "rfilename": "params_shard_51.bin" }, { "rfilename": "params_shard_52.bin" }, { "rfilename": "params_shard_53.bin" }, { "rfilename": "params_shard_54.bin" }, { "rfilename": "params_shard_55.bin" }, { "rfilename": "params_shard_56.bin" }, { "rfilename": "params_shard_57.bin" }, { "rfilename": "params_shard_58.bin" }, { "rfilename": "params_shard_59.bin" }, { "rfilename": "params_shard_6.bin" }, { "rfilename": "params_shard_60.bin" }, { "rfilename": "params_shard_61.bin" }, { "rfilename": "params_shard_62.bin" }, { "rfilename": "params_shard_63.bin" }, { "rfilename": "params_shard_64.bin" }, { "rfilename": "params_shard_65.bin" }, { "rfilename": "params_shard_66.bin" }, { "rfilename": "params_shard_67.bin" }, { "rfilename": "params_shard_68.bin" }, { "rfilename": "params_shard_69.bin" }, { "rfilename": "params_shard_7.bin" }, { "rfilename": "params_shard_70.bin" }, { "rfilename": "params_shard_71.bin" }, { "rfilename": "params_shard_72.bin" }, { "rfilename": "params_shard_73.bin" }, { "rfilename": "params_shard_74.bin" }, { "rfilename": "params_shard_75.bin" }, { "rfilename": "params_shard_76.bin" }, { "rfilename": "params_shard_77.bin" }, { "rfilename": "params_shard_78.bin" }, { "rfilename": "params_shard_79.bin" }, { "rfilename": "params_shard_8.bin" }, { "rfilename": "params_shard_80.bin" }, { "rfilename": "params_shard_81.bin" }, { "rfilename": "params_shard_82.bin" }, { "rfilename": "params_shard_9.bin" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
null
microsoft/Phi-3.5-mini-instruct
147
text-generation
147
0
0
0
0
843,171
66ab11019a318d1bc0319f0b
Saxo/Linkbricks-Horizon-AI-Korean-llama-3.1-sft-dpo-8B
Saxo
False
not-popular-enough
2024-09-02T00:46:24.000Z
4
false
4a543d074904d38c89ea0c5f5c042c55494096a7
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}}
1,939
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "conversational", "ko", "en", "jp", "cn", "dataset:Saxo/ko_cn_translation_tech_social_science_linkbricks_single_dataset", "dataset:Saxo/ko_jp_translation_tech_social_science_linkbricks_single_dataset", "dataset:Saxo/en_ko_translation_tech_science_linkbricks_single_dataset_with_prompt_text_huggingface", "dataset:Saxo/en_ko_translation_social_science_linkbricks_single_dataset_with_prompt_text_huggingface", "dataset:Saxo/ko_aspect_sentiment_sns_mall_sentiment_linkbricks_single_dataset_with_prompt_text_huggingface", "dataset:Saxo/ko_summarization_linkbricks_single_dataset_with_prompt_text_huggingface", "dataset:Saxo/OpenOrca_cleaned_kor_linkbricks_single_dataset_with_prompt_text_huggingface", "dataset:Saxo/ko_government_qa_total_linkbricks_single_dataset_with_prompt_text_huggingface_sampled", "dataset:maywell/ko_Ultrafeedback_binarized", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:quantized:meta-llama/Llama-3.1-8B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-08-01T04:37:21.000Z
Saxo/Linkbricks-Horizon-AI-Korean-llama-3.1-sft-dpo-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "ggml-model-bf16.gguf" }, { "rfilename": "ggml-model-q4_k_m.gguf" }, { "rfilename": "ggml-model-q5_k_m.gguf" }, { "rfilename": "ggml-model-q6_k.gguf" }, { "rfilename": "ggml-model-q8_0.gguf" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
1
0
1
792,355
65d3e1472af7a199b8e33df1
predibase/dbpedia
predibase
False
not-popular-enough
2024-02-21T19:14:00.000Z
8
false
150202d7bc480c707c10f63e18a27ea742525919
{"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}}
1,906
[ "peft", "safetensors", "text-generation", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "region:us" ]
text-generation
peft
2024-02-19T23:16:23.000Z
predibase/dbpedia
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" } ]
null
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
499,689
66dc5c7e79d36ca08592c520
gerbyk/Qwen-Qwen1.5-0.5B-1725717630
gerbyk
False
pipeline-not-detected
2024-09-07T14:00:31.000Z
0
false
7bff538242d48525b816640d5e4c6640a3fccc1e
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-0.5B", "task_type": "CAUSAL_LM"}}
1,902
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-0.5B", "base_model:adapter:Qwen/Qwen1.5-0.5B", "region:us" ]
null
peft
2024-09-07T14:00:30.000Z
gerbyk/Qwen-Qwen1.5-0.5B-1725717630
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "added_tokens.json" }, { "rfilename": "merges.txt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen1.5-0.5B
28,947
text-generation
28,947
0
0
0
0
910,551
6632e900e97d286a1d74f360
12thD/ko-Llama-3-8B-sft-v0.3
12thD
False
not-popular-enough
2024-05-02T01:24:46.000Z
0
false
134a44b329a37805306c77e45e932d839cae8baa
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
1,871
[ "transformers", "safetensors", "llama", "text-generation", "korean", "gemma", "conversational", "ko", "en", "arxiv:1910.09700", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-02T01:14:40.000Z
12thD/ko-Llama-3-8B-sft-v0.3
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
615,907
66e02101e1bfac4b460f103a
HF1BitLLM/Llama3-8B-1.58-100B-tokens
HF1BitLLM
False
not-popular-enough
2024-09-19T04:09:11.000Z
43
false
5c35ae1f2c622b75a9c28e3603074863d74e4792
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"quant_method": "bitnet"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
1,870
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:2402.17764", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "bitnet", "region:us" ]
text-generation
transformers
2024-09-10T10:35:45.000Z
HF1BitLLM/Llama3-8B-1.58-100B-tokens
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
2
924,743
66cf01c89a7e2c98a06284a3
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
BAAI
False
library-not-detected
2024-09-05T02:56:26.000Z
8
false
fd6d02d300e3b9015e07c217e26c6f1b4823963a
{"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] | trim + '\n\n' %}{% set messages = messages[1:] %}{% else %}{% set system_message = '' %}{% endif %}{% for message in messages %}{% if loop.index0 == 0 %}{% set content = system_message + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + content | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<end_of_turn>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
1,859
[ "safetensors", "gemma2", "en", "dataset:BAAI/Infinity-Instruct", "base_model:google/gemma-2-9b-it", "base_model:finetune:google/gemma-2-9b-it", "region:us" ]
null
null
2024-08-28T10:54:00.000Z
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "eval_results.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
gemma2
google/gemma-2-9b-it
131
text-generation
131
0
4
0
2
869,266
664fc1b1e823e13f74323d39
solidrust/Mistral-7B-Instruct-v0.3-AWQ
solidrust
False
explicit-opt-out
2024-09-03T08:50:39.000Z
2
false
95b1295ddd1a8673117cdc7bd2a4da2a457bb3f7
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
1,841
[ "transformers", "safetensors", "mistral", "text-generation", "4-bit", "AWQ", "autotrain_compatible", "endpoints_compatible", "conversational", "base_model:mistralai/Mistral-7B-Instruct-v0.3", "base_model:quantized:mistralai/Mistral-7B-Instruct-v0.3", "license:apache-2.0", "text-generation-inference", "awq", "region:us" ]
text-generation
transformers
2024-05-23T22:22:41.000Z
solidrust/Mistral-7B-Instruct-v0.3-AWQ
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "quant_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-Instruct-v0.3
320
text-generation
320
0
0
0
0
655,731
66cfbe7e8b9d49dccb3c88de
jfranklin-foundry/01-ai-Yi-1.5-9B-Chat-1724890832
jfranklin-foundry
False
pipeline-not-detected
2024-08-29T00:19:56.000Z
0
false
3585c30d90852b0135b939801df7f46e146524ba
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
1,818
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B-Chat", "base_model:adapter:01-ai/Yi-1.5-9B-Chat", "region:us" ]
null
peft
2024-08-29T00:19:10.000Z
jfranklin-foundry/01-ai-Yi-1.5-9B-Chat-1724890832
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B-Chat
134
text-generation
134
0
0
0
0
871,497
65771780c793e5b72b0e8d68
TheBloke/Mixtral-8x7B-v0.1-GPTQ
TheBloke
False
explicit-opt-out
2023-12-14T14:30:54.000Z
127
false
7d1eb57b65f823458e27509cd0aac7172f54a260
{"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
1,817
[ "transformers", "safetensors", "mixtral", "text-generation", "fr", "it", "de", "es", "en", "base_model:mistralai/Mixtral-8x7B-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
transformers
2023-12-11T14:06:56.000Z
TheBloke/Mixtral-8x7B-v0.1-GPTQ
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "quantize_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mixtral
mistralai/Mixtral-8x7B-v0.1
182
text-generation
182
0
0
0
0
414,316
64f614a19c2aeb2a3dea1312
TheBloke/Llama-2-13B-GGUF
TheBloke
False
explicit-opt-out
2023-09-27T12:47:13.000Z
61
false
b106d1c018ac999af9130b83134fb6b7c5331dea
{"model_type": "llama"}
1,815
[ "transformers", "gguf", "llama", "facebook", "meta", "pytorch", "llama-2", "text-generation", "en", "arxiv:2307.09288", "base_model:meta-llama/Llama-2-13b-hf", "base_model:quantized:meta-llama/Llama-2-13b-hf", "license:llama2", "text-generation-inference", "region:us" ]
text-generation
transformers
2023-09-04T17:32:17.000Z
TheBloke/Llama-2-13B-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.txt" }, { "rfilename": "Notice" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "config.json" }, { "rfilename": "llama-2-13b.Q2_K.gguf" }, { "rfilename": "llama-2-13b.Q3_K_L.gguf" }, { "rfilename": "llama-2-13b.Q3_K_M.gguf" }, { "rfilename": "llama-2-13b.Q3_K_S.gguf" }, { "rfilename": "llama-2-13b.Q4_0.gguf" }, { "rfilename": "llama-2-13b.Q4_K_M.gguf" }, { "rfilename": "llama-2-13b.Q4_K_S.gguf" }, { "rfilename": "llama-2-13b.Q5_0.gguf" }, { "rfilename": "llama-2-13b.Q5_K_M.gguf" }, { "rfilename": "llama-2-13b.Q5_K_S.gguf" }, { "rfilename": "llama-2-13b.Q6_K.gguf" }, { "rfilename": "llama-2-13b.Q8_0.gguf" } ]
llama
meta-llama/Llama-2-13b-hf
180
text-generation
180
0
0
0
0
303,243
666154aba52d8a6aa5efa07c
Qwen/Qwen2-7B-Instruct-GPTQ-Int8
Qwen
False
not-popular-enough
2024-08-21T10:33:47.000Z
16
false
b018e4540ccdb307df57b1a06918be08d29e6464
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 8, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
1,802
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "base_model:Qwen/Qwen2-7B-Instruct", "base_model:quantized:Qwen/Qwen2-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
transformers
2024-06-06T06:18:19.000Z
Qwen/Qwen2-7B-Instruct-GPTQ-Int8
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
qwen2
Qwen/Qwen2-7B-Instruct
161
text-generation
161
2
0
0
0
685,153