id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
MaziyarPanahi/llama3-8B-aifeifei-1.0-GGUF
null
null
"2024-11-06T14:57:15Z"
null
null
112
null
null
null
null
0
null
[ "gguf", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:aifeifei798/llama3-8B-aifeifei-1.0", "base_model:quantized:aifeifei798/llama3-8B-aifeifei-1.0", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
Youlln/ECE-PRYMMAL-YL-7B-SLERP-V4
null
null
"2024-11-06T14:59:27Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
float32
4.46.2
true
152,064
null
0
null
null
null
null
Marsouuu/general3Bv2-ECE-PRYMMAL-Martial
null
null
"2024-11-06T15:11:40Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "base_model:Tsunami-th/Tsunami-0.5x-7B-Instruct", "base_model:merge:Tsunami-th/Tsunami-0.5x-7B-Instruct", "base_model:fblgit/cybertron-v4-qw7B-MGS", "base_model:merge:fblgit/cybertron-v4-qw7B-MGS", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial
null
null
"2024-11-06T15:31:06Z"
null
null
35
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "base_model:Tsunami-th/Tsunami-0.5x-7B-Instruct", "base_model:merge:Tsunami-th/Tsunami-0.5x-7B-Instruct", "base_model:fblgit/cybertron-v4-qw7B-MGS", "base_model:merge:fblgit/cybertron-v4-qw7B-MGS", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
c01zaut/gemma-2-9b-rk3588-1.1.1
null
null
"2024-11-06T15:32:13Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:2009.03300", "arxiv:1905.07830", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1905.10044", "arxiv:1907.10641", "arxiv:1811.00937", "arxiv:1809.02789", "arxiv:1911.01547", "arxiv:1705.03551", "arxiv:2107.03374", "arxiv:2108.07732", "arxiv:2110.14168", "arxiv:2009.11462", "arxiv:2101.11718", "arxiv:2110.08193", "arxiv:1804.09301", "arxiv:2109.07958", "arxiv:1804.06876", "arxiv:2103.03874", "arxiv:2304.06364", "arxiv:2206.04615", "arxiv:2203.09509", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
float32
4.42.0.dev0
true
256,000
false
0
256
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-06T15:37:23Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
MaziyarPanahi/Kancil-V1-llama3-fp16-GGUF
null
null
"2024-11-06T15:40:45Z"
null
null
108
null
null
null
null
0
null
[ "gguf", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:afrizalha/Kancil-V1-llama3-fp16", "base_model:quantized:afrizalha/Kancil-V1-llama3-fp16", "region:us", "imatrix" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter2
null
null
"2024-11-06T15:50:19Z"
null
null
65
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "generated_from_trainer", "trl", "dpo", "conversational", "arxiv:2305.18290", "base_model:RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter1", "base_model:finetune:RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
12,288
32,768
mistral
32
36
8
0.00001
100,000,000
32,768
false
bfloat16
4.45.2
false
131,073
null
0
128
null
null
null
jukofyork/creative-writer-v0.2-bravo-35b
null
null
"2024-11-06T15:50:57Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "cohere", "text-generation", "creative-writing", "creative-writer", "multiplicative-lora", "conversational", "license:cc-by-nc-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "CohereForCausalLM" ]
5
255,001
silu
8,192
0.02
22,528
8,192
cohere
64
40
64
null
8,000,000
null
true
float16
4.38.2
true
256,000
false
0
null
null
1
null
Diplomkaazvposlednimsemestru/my_small_gpt2_csknihy
null
null
"2024-11-06T15:52:02Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
daliabess7/llama3-8b-oig-unsloth-merged
null
null
"2024-11-06T16:21:12Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float16
4.46.2
true
128,256
false
0
128
false
1
null
netcat420/MFANN3bv0.23
null
null
"2024-11-06T16:21:52Z"
null
null
32
null
null
null
null
0
transformers
[ "transformers", "safetensors", "phi", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "PhiForCausalLM" ]
50,256
50,256
gelu_new
2,560
0.02
10,240
131,072
phi
32
32
32
null
10,000
null
false
float32
4.46.2
true
51,200
null
0
null
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-06T16:29:52Z"
null
null
7
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
MaziyarPanahi/CosmicBun-8B-GGUF
null
null
"2024-11-06T16:33:49Z"
null
null
117
null
null
null
null
0
null
[ "gguf", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:aloobun/CosmicBun-8B", "base_model:quantized:aloobun/CosmicBun-8B", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
davidbzyk/QuantQwen2.5-32b-merged_16bit
null
null
"2024-11-06T17:25:17Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
MaziyarPanahi/llama-3-8b-English-to-Hinglish-GGUF
null
null
"2024-11-06T17:25:42Z"
null
null
112
null
null
null
null
0
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:Hinglish-Project/llama-3-8b-English-to-Hinglish", "base_model:quantized:Hinglish-Project/llama-3-8b-English-to-Hinglish", "region:us", "imatrix" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
davidbzyk/QuantQwen2.5-7b-Coder-merged_16bit
null
null
"2024-11-06T17:26:58Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit", "base_model:finetune:unsloth/Qwen2.5-Coder-7B-Instruct-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-06T17:32:09Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
MaziyarPanahi/MeowGPT-ll3-GGUF
null
null
"2024-11-06T18:11:12Z"
null
null
109
null
null
null
null
0
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:cutycat2000x/MeowGPT-ll3", "base_model:quantized:cutycat2000x/MeowGPT-ll3", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
2point5p/krx-qwen2.5-7b-it-X-Two
null
null
"2024-11-06T18:13:58Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "conversational", "en", "base_model:unsloth/Qwen2.5-7B-Instruct", "base_model:finetune:unsloth/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.0
true
152,064
null
0
null
null
null
null
skrishna/llama3-8b-hellaswag
null
null
"2024-11-06T18:16:40Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float32
4.44.0
true
128,256
false
0
null
false
1
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-06T18:24:55Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
zelk12/Gemma-2-DMv1A-9B
null
null
"2024-11-06T18:34:29Z"
null
null
16
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:lemon07r/Gemma-2-Ataraxy-9B", "base_model:merge:lemon07r/Gemma-2-Ataraxy-9B", "base_model:sam-paech/Darkest-muse-v1", "base_model:merge:sam-paech/Darkest-muse-v1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
false
256,000
false
0
256
null
null
null
toastloaf/smollm2-135m-it-orca-agentinstruct-creative
null
null
"2024-11-06T18:40:16Z"
null
null
7
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "llama", "text-generation", "autotrain", "text-generation-inference", "conversational", "dataset:microsoft/orca-agentinstruct-1M-v1", "base_model:HuggingFaceTB/SmolLM2-135M-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-135M-Instruct", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
576
0.041667
1,536
8,192
llama
9
30
3
0.00001
100,000
null
true
float32
4.45.0
true
49,152
false
0
64
false
1
null
emozilla/smol-15b-init
null
null
"2024-11-06T18:42:23Z"
null
null
33
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
0
0
silu
5,120
0.02
17,920
4,096
llama
40
42
8
0.00001
130,000
null
false
bfloat16
4.46.1
true
49,152
false
0
128
false
1
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-06T19:09:00Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
zelk12/Gemma-2-IAv2-9B
null
null
"2024-11-06T19:10:37Z"
null
null
17
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:ifable/gemma-2-Ifable-9B", "base_model:merge:ifable/gemma-2-Ifable-9B", "base_model:lemon07r/Gemma-2-Ataraxy-v2-9B", "base_model:merge:lemon07r/Gemma-2-Ataraxy-v2-9B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
false
256,000
false
0
256
null
null
null
Zekunli/qwen2.5-1.5b-lora-wo-cot-wo-cor
null
null
"2024-11-06T19:34:53Z"
null
null
77
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
null
true
float32
4.45.0.dev0
true
151,936
null
0
null
null
null
null
loremipsum3658/Qwen2.5-7B-Instruct-AWQ
null
null
"2024-11-06T19:43:30Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:quantized:Qwen/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "awq", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
float16
4.41.1
true
152,064
null
0
null
null
null
null
zelk12/Gemma-2-MT1MT1g1-9B
null
null
"2024-11-06T19:55:00Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:zelk12/MT1-Gen1-gemma-2-9B", "base_model:merge:zelk12/MT1-Gen1-gemma-2-9B", "base_model:zelk12/MT1-gemma-2-9B", "base_model:merge:zelk12/MT1-gemma-2-9B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
true
256,000
false
0
256
null
null
null
broalantap/GPT2-large-4-40000steps
null
null
"2024-11-06T19:57:53Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
bfloat16
4.41.2
true
50,321
null
null
null
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-06T20:01:52Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
fayez-siddiqui/Llama3_8b_HQQ-1bit
null
null
"2024-11-06T20:02:13Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "hqq", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.46.2
true
128,256
false
0
128
false
1
null
zelk12/Gemma-2-Tv3Tv1-9B
null
null
"2024-11-06T20:04:43Z"
null
null
13
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:TheDrummer/Tiger-Gemma-9B-v1", "base_model:merge:TheDrummer/Tiger-Gemma-9B-v1", "base_model:TheDrummer/Tiger-Gemma-9B-v3", "base_model:merge:TheDrummer/Tiger-Gemma-9B-v3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
false
256,000
false
0
256
null
null
null
fayez-siddiqui/Llama3_8b_HQQ-4bit
null
null
"2024-11-06T20:14:04Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "hqq", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.46.2
true
128,256
false
0
128
false
1
null
zelk12/Gemma-2-T1-9B
null
null
"2024-11-06T20:15:46Z"
null
null
13
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:zelk12/Gemma-2-MT1MT1g1-9B", "base_model:merge:zelk12/Gemma-2-MT1MT1g1-9B", "base_model:zelk12/Gemma-2-Tv3Tv1-9B", "base_model:merge:zelk12/Gemma-2-Tv3Tv1-9B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
true
256,000
false
0
256
null
null
null
fayez-siddiqui/Llama3_8b_HQQ-2bit
null
null
"2024-11-06T20:20:17Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "hqq", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.46.2
true
128,256
false
0
128
false
1
null
gaspardsinequa/SineAgent-v0.5-data-fail
null
null
"2024-11-06T20:22:29Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
null
true
float16
4.45.1
true
151,936
null
0
null
null
null
null
zelk12/Gemma-2-T2-9B
null
null
"2024-11-06T20:23:22Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:zelk12/Gemma-2-DMv1A-9B", "base_model:merge:zelk12/Gemma-2-DMv1A-9B", "base_model:zelk12/Gemma-2-IAv2-9B", "base_model:merge:zelk12/Gemma-2-IAv2-9B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
false
256,000
false
0
256
null
null
null
zelk12/Gemma-2-TM-9B
null
null
"2024-11-06T20:32:16Z"
null
null
24
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "mergekit", "merge", "conversational", "base_model:zelk12/Gemma-2-T1-9B", "base_model:merge:zelk12/Gemma-2-T1-9B", "base_model:zelk12/Gemma-2-T2-9B", "base_model:merge:zelk12/Gemma-2-T2-9B", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.45.1
false
256,000
false
0
256
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-06T20:45:40Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
hsuwill000/Qwen2.5-3B-Instruct-openvino
null
null
"2024-11-06T21:04:04Z"
null
null
26
null
null
null
null
0
transformers
[ "transformers", "safetensors", "openvino", "qwen2", "text-generation", "chat", "openvino-export", "conversational", "en", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
32,768
true
bfloat16
4.42.4
true
151,936
null
0
null
null
null
null
MaziyarPanahi/OpenChat-3.5-0106_32K-PoSE-GGUF
null
null
"2024-11-06T21:30:31Z"
null
null
59
null
null
null
null
0
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:Pretergeek/OpenChat-3.5-0106_32K-PoSE", "base_model:quantized:Pretergeek/OpenChat-3.5-0106_32K-PoSE", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-06T21:38:51Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
dondongwonlee/GELI
null
null
"2024-11-06T21:42:13Z"
null
null
13
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "Conversation", "Social", "conversational", "en", "arxiv:2403.11330", "base_model:meta-llama/Llama-2-7b-chat-hf", "base_model:finetune:meta-llama/Llama-2-7b-chat-hf", "license:llama2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float32
4.36.0
true
32,000
false
0
null
null
1
null
Xu-Ouyang/pythia-6.9b-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-06T22:31:52Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
4,096
0.02
16,384
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,432
true
0
null
null
null
null
yiran-wang3/qwen2_coder7b_adamw_iter1
null
null
"2024-11-06T22:52:03Z"
null
null
51
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2coder_7b_original_cn_mining_oj_iter0-binarized", "base_model:Qwen/Qwen2.5-Coder-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-Coder-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step1-GPTQ-wikitext2
null
null
"2024-11-06T23:05:54Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
yiran-wang3/qwen2_chat7b_adamw_iter1
null
null
"2024-11-06T23:14:21Z"
null
null
52
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2_7b_original_cn_mining_oj_iter0-binarized", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step2-GPTQ-wikitext2
null
null
"2024-11-06T23:37:20Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Kaballas/T35Model
null
null
"2024-11-06T23:38:24Z"
null
null
28
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "en", "base_model:Kaballas/T21Model", "base_model:finetune:Kaballas/T21Model", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.47.0.dev0
true
152,064
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep8
null
null
"2024-11-06T23:38:46Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
yjwon/mp_gemma9b_sft_dpo_beta5e-2_epoch4
null
null
"2024-11-06T23:39:06Z"
null
null
30
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.46.1
true
256,000
false
0
256
null
null
null
yjwon/mp_gemma9b_sft_dpo_beta5e-2_epoch2
null
null
"2024-11-06T23:39:09Z"
null
null
28
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.46.1
true
256,000
false
0
256
null
null
null
yjwon/mp_gemma9b_sft_dpo_beta5e-2_epoch5
null
null
"2024-11-06T23:39:10Z"
null
null
30
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.46.1
true
256,000
false
0
256
null
null
null
yjwon/mp_gemma9b_sft_dpo_beta5e-2_epoch1
null
null
"2024-11-06T23:39:11Z"
null
null
52
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.46.1
true
256,000
false
0
256
null
null
null
yjwon/mp_gemma9b_sft_dpo_beta5e-2_epoch3
null
null
"2024-11-06T23:39:12Z"
null
null
26
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.46.1
true
256,000
false
0
256
null
null
null
rewicks/monolingual_de_8k-shared_ep9
null
null
"2024-11-06T23:42:21Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep10
null
null
"2024-11-06T23:45:40Z"
null
null
19
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep11
null
null
"2024-11-06T23:58:41Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-07T00:02:47Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep12
null
null
"2024-11-07T00:07:13Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
yiran-wang3/qwen2_coder7b_reflct_adamw_iter1
null
null
"2024-11-07T00:07:27Z"
null
null
73
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2coder_reflct7b_original_cn_mining_oj_iter0-binarized-reflection-scored", "base_model:Qwen/Qwen2.5-Coder-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-Coder-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep13
null
null
"2024-11-07T00:10:45Z"
null
null
20
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep14
null
null
"2024-11-07T00:13:41Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep15
null
null
"2024-11-07T00:17:14Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
kayfour/T3Q-Qwen2.5-7B-it-ko-safe
null
null
"2024-11-07T00:18:11Z"
null
null
22
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.46.2
true
152,064
null
0
null
null
null
null
yiran-wang3/qwen2_coder7b_adamw_iter2
null
null
"2024-11-07T00:20:30Z"
null
null
47
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2coder_7b_sppo_hard_new_cn_mining_oj_iter1-binarized", "base_model:yiran-wang3/qwen2_coder7b_adamw_iter1", "base_model:finetune:yiran-wang3/qwen2_coder7b_adamw_iter1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep16
null
null
"2024-11-07T00:20:34Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep17
null
null
"2024-11-07T00:28:52Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-07T00:30:13Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep18
null
null
"2024-11-07T00:37:17Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
yiran-wang3/qwen2_chat7b_adamw_iter2
null
null
"2024-11-07T00:43:24Z"
null
null
49
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2_7b_sppo_hard_new_cn_mining_oj_iter1-binarized", "base_model:yiran-wang3/qwen2_chat7b_adamw_iter1", "base_model:finetune:yiran-wang3/qwen2_chat7b_adamw_iter1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep19
null
null
"2024-11-07T00:45:45Z"
null
null
80
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-07T01:01:27Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-07T01:23:06Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Grayx/rnd_18
null
null
"2024-11-07T01:26:44Z"
null
null
264
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
50,256
50,256
silu
1,408
0.02
4,340
2,048
llama
22
24
22
0.00001
10,000
null
false
float32
4.44.1
false
50,257
false
0
64
false
1
null
Hachipo/Qwen2.5-0.5B_2epoch_inst_2
null
null
"2024-11-07T01:28:49Z"
null
null
16
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float32
4.46.2
true
151,667
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-07T01:44:56Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
yiran-wang3/qwen2_chat7b_reflct_adamw_iter1
null
null
"2024-11-07T01:45:25Z"
null
null
33
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2_reflct7b_original_cn_mining_oj_iter0-binarized-reflection-scored", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
emozilla/smol-7b-init
null
null
"2024-11-07T01:48:45Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
0
0
silu
4,096
0.02
14,336
2,048
llama
32
32
8
0.00001
130,000
null
false
bfloat16
4.46.1
true
49,152
false
0
128
false
1
null
yiran-wang3/qwen2_coder7b_adamw_iter3
null
null
"2024-11-07T01:52:54Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2coder_7b_sppo_hard_new_cn_mining_oj_iter2-binarized", "base_model:yiran-wang3/qwen2_coder7b_adamw_iter2", "base_model:finetune:yiran-wang3/qwen2_coder7b_adamw_iter2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
Grayx/rnd_25
null
null
"2024-11-07T01:58:32Z"
null
null
258
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
50,256
50,256
silu
1,408
0.02
4,340
2,048
llama
22
24
22
0.00001
10,000
null
false
float32
4.44.1
false
50,257
false
0
64
false
1
null
yiran-wang3/qwen2_chat7b_adamw_iter3
null
null
"2024-11-07T02:05:53Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2_7b_sppo_hard_new_cn_mining_oj_iter2-binarized", "base_model:yiran-wang3/qwen2_chat7b_adamw_iter2", "base_model:finetune:yiran-wang3/qwen2_chat7b_adamw_iter2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-07T02:06:48Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
vrglx33/TinyLlama-1.1B-32k-Instruct-mlx
null
null
"2024-11-07T02:19:53Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mlx", "conversational", "en", "dataset:LDJnr/Capybara", "dataset:jondurbin/airoboros-3.2", "dataset:unalignment/toxic-dpo-v0.1", "dataset:LDJnr/Verified-Camel", "dataset:HuggingFaceH4/no_robots", "dataset:Doctor-Shotgun/no-robots-sharegpt", "dataset:Doctor-Shotgun/capybara-sharegpt", "base_model:Doctor-Shotgun/TinyLlama-1.1B-32k-Instruct", "base_model:quantized:Doctor-Shotgun/TinyLlama-1.1B-32k-Instruct", "autotrain_compatible", "text-generation-inference", "4-bit", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
5,632
32,768
llama
32
22
4
0.00001
1,000,000
null
false
bfloat16
4.37.0.dev0
false
32,000
false
0
null
null
1
null
yiran-wang3/qwen2_coder7b_reflct_adamw_iter2
null
null
"2024-11-07T02:23:58Z"
null
null
47
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/qw2coder_reflct7b_sppo_hard_new_cn_mining_oj_iter1-binarized-reflection-scored", "base_model:yiran-wang3/qwen2_coder7b_reflct_adamw_iter1", "base_model:finetune:yiran-wang3/qwen2_coder7b_reflct_adamw_iter1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-07T02:40:12Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
vickyTEC/results
null
null
"2024-11-07T02:41:57Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.44.2
true
50,257
null
null
null
null
null
null
RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter2-s42
null
null
"2024-11-07T02:42:40Z"
null
null
47
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "generated_from_trainer", "trl", "dpo", "conversational", "arxiv:2305.18290", "base_model:RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter1", "base_model:finetune:RyanYr/self-correct_ministral8Bit_mMQA_dpo_iter1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
12,288
32,768
mistral
32
36
8
0.00001
100,000,000
32,768
false
bfloat16
4.45.2
false
131,073
null
0
128
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-07T03:05:39Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-2.8b-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-07T03:32:40Z"
null
null
9
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,560
0.02
10,240
2,048
gpt_neox
32
32
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
tinhpx2911/Vistral_IL_M10_sft_chat_2ep
null
null
"2024-11-07T03:47:26Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.46.2
true
38,369
null
0
128
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step1-GPTQ-wikitext2
null
null
"2024-11-07T03:50:53Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
emozilla/smol-3b-init
null
null
"2024-11-07T03:52:09Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
0
0
silu
3,072
0.02
8,192
2,048
llama
24
28
8
0.00001
130,000
null
false
bfloat16
4.46.1
true
49,152
false
0
128
false
1
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step2-GPTQ-wikitext2
null
null
"2024-11-07T04:03:40Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
twodigit/hades_9b3-38000
null
null
"2024-11-07T04:07:05Z"
null
null
19
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.43.4
true
256,000
false
0
256
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-07T04:17:25Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
RollLBJ/test-model5
null
null
"2024-11-07T04:19:22Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "en", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.46.2
true
128,256
false
0
128
false
1
null
sunwoo2/sunwoo-Qwen2.5-0.5B-241107
null
null
"2024-11-07T04:20:35Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "pytorch", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float16
4.46.2
true
151,936
null
0
null
null
null
null