id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
Takvmi/model_pmc_gamma_0.1_noise0.2_0.2_epoch0
null
null
"2024-11-09T18:16:29Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_kl_0.1_noise0.2_0.2_epoch0
null
null
"2024-11-09T18:17:17Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_beta_0.1_noise0.2_0.2_epoch0
null
null
"2024-11-09T18:18:21Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_beta_0.1_noise0.2_0.2_epoch1
null
null
"2024-11-09T18:21:55Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
AndreyRzhaksinskiy/CDS-starcoder2-Ins-7b-E2E-20241109
null
null
"2024-11-09T18:24:02Z"
null
null
27
null
null
null
null
0
transformers
[ "transformers", "safetensors", "starcoder2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Starcoder2ForCausalLM" ]
0
0
gelu_pytorch_tanh
4,608
0.018042
18,432
16,384
starcoder2
36
32
4
null
1,000,000
4,096
null
float16
4.46.0
true
49,152
null
0.1
null
null
null
null
VLKVLK/media-file-recognizer
null
null
"2024-11-09T18:25:28Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "llama-factory", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,536
0.02
8,960
32,768
qwen2
12
28
2
0.000001
1,000,000
null
true
bfloat16
4.44.2
true
151,936
null
0
null
null
null
null
Takvmi/model_pmc_kl_0.1_noise0.2_0.2_epoch1
null
null
"2024-11-09T18:26:46Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_gamma_0.1_noise0.2_0.2_epoch1
null
null
"2024-11-09T18:27:43Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Hachipo/chico-1.3b-4epoch_inst
null
null
"2024-11-09T18:33:56Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
7
7
silu
1,536
0.02
8,192
2,048
llama
24
24
24
0.000001
10,000
null
false
float32
4.46.2
true
50,570
false
0
64
false
1
null
Uynaity/AutoTrain-Qwen-Rui-SHLR
null
null
"2024-11-09T19:06:46Z"
null
null
22
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "qwen2", "text-generation", "autotrain", "text-generation-inference", "peft", "conversational", "dataset:Uynaity/Rui-Pro", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float16
4.46.2
true
151,665
null
0
null
null
null
null
kanishka/opt-babylm2-rewritten-clean-spacy-32k-earlystop_seed-42_1e-3
null
null
"2024-11-09T19:15:11Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "opt", "text-generation", "generated_from_trainer", "dataset:kanishka/babylm2-rewritten-clean-spacy", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "OPTForCausalLM" ]
1
1
null
768
null
null
256
opt
12
12
null
null
null
null
null
float32
4.45.1
true
32,768
null
0
null
null
null
null
pavl0/LLaMA5M
null
null
"2024-11-09T19:15:18Z"
null
null
48
null
null
null
null
0
transformers
[ "transformers", "safetensors", "pavl0_llama", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "LLaMA" ]
null
null
null
null
null
null
null
pavl0_llama
null
null
null
null
null
null
null
float32
4.46.2
null
32,000
null
null
null
null
null
null
KaraKaraWitch/HiTZ-GoLLIE-13B-AsSafeTensors
null
null
"2024-11-09T19:44:39Z"
null
null
20
null
null
null
null
0
null
[ "safetensors", "llama", "code", "text-generation-inference", "Information Extraction", "IE", "Named Entity Recogniton", "Event Extraction", "Relation Extraction", "LLaMA", "text-generation", "custom_code", "en", "dataset:ACE05", "dataset:bc5cdr", "dataset:conll2003", "dataset:ncbi_disease", "dataset:conll2012_ontonotesv5", "dataset:rams", "dataset:tacred", "dataset:wnut_17", "arxiv:2310.03668", "license:llama2", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
5,120
0.02
13,824
16,384
llama
40
40
40
0.00001
1,000,000
null
false
float32
4.44.2
true
32,016
false
0
null
false
1
null
pavl0/LLaMA5M_config
null
null
"2024-11-09T19:59:04Z"
null
null
9
null
null
null
null
0
transformers
[ "transformers", "pavl0_llama", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "LLaMA" ]
null
null
null
null
null
null
null
pavl0_llama
null
null
null
null
null
null
null
float32
4.46.2
null
32,000
null
null
null
null
null
null
RyanYr/self-reflect_ministral8Bit_math_gsm8k_dpo_iter1
null
null
"2024-11-09T20:18:14Z"
null
null
33
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "generated_from_trainer", "trl", "dpo", "conversational", "arxiv:2305.18290", "base_model:mistralai/Ministral-8B-Instruct-2410", "base_model:finetune:mistralai/Ministral-8B-Instruct-2410", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
12,288
32,768
mistral
32
36
8
0.00001
100,000,000
32,768
false
bfloat16
4.45.2
false
131,073
null
0
128
null
null
null
ThatsGroes/gemma-2-27b-it-SkoleGPT
null
null
"2024-11-09T21:08:51Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:unsloth/gemma-2-27b-it", "base_model:finetune:unsloth/gemma-2-27b-it", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
4,608
0.02
36,864
8,192
gemma2
32
46
16
0.000001
10,000
4,096
null
bfloat16
4.47.0.dev0
true
256,000
false
0
128
null
null
null
Limby/qwenfloat16
null
null
"2024-11-09T21:27:08Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float16
4.46.2
true
151,936
null
0
null
null
null
null
yacioo/yassio-17-ben-belgacem
null
null
"2024-11-09T21:56:56Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "phi", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "PhiForCausalLM" ]
50,256
50,256
gelu_new
2,560
0.02
10,240
2,048
phi
32
32
32
null
10,000
null
false
float16
4.46.2
true
51,200
null
0
null
null
null
null
Takvmi/model_pmc_kl_0.1_noise0.1_0.1_epoch0
null
null
"2024-11-09T22:09:38Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_kl_0.1_noise0.2_0.1_epoch0
null
null
"2024-11-09T22:11:05Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
Takvmi/model_pmc_beta_0.05_noise0.2_0.1_epoch0
null
null
"2024-11-09T22:13:35Z"
null
null
5
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,258
null
null
null
null
null
null
tanquangduong/Qwen2.5-3B-Instruct-TinyStories
null
null
"2024-11-09T22:17:58Z"
null
null
52
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "en", "base_model:unsloth/Qwen2.5-3B", "base_model:finetune:unsloth/Qwen2.5-3B", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
bfloat16
4.46.2
true
151,936
null
0
null
null
null
null
TDN-M/tknv1
null
null
"2024-11-09T22:26:37Z"
null
null
30
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float32
4.44.2
true
151,936
null
0
null
null
null
null
yakazimir/simpo-exps
null
null
"2024-11-09T22:50:09Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "alignment-handbook", "trl", "simpo", "generated_from_trainer", "conversational", "dataset:princeton-nlp/llama3-ultrafeedback-armorm", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.44.2
true
128,256
false
0
null
false
1
null
thirdeyeai/SmolLM2-1.7B-Instruct-Uncensored
null
null
"2024-11-09T23:01:51Z"
null
null
9
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-1.7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
8,192
8,192
llama
32
24
32
0.00001
130,000
null
true
bfloat16
4.44.2
true
49,152
false
0
null
false
1
null
henryz2004/Qwen-1_8B-Chat-lens
null
null
"2024-11-09T23:43:48Z"
null
null
2
null
null
null
null
0
null
[ "safetensors", "qwen", "text-generation", "custom_code", "zh", "en", "arxiv:2309.16609", "arxiv:2305.08322", "arxiv:2009.03300", "region:us" ]
text-generation
null
null
0
[ "QWenLMHeadModel" ]
null
null
null
2,048
0.02
11,008
8,192
qwen
16
24
null
null
null
null
false
null
4.32.0
true
151,936
null
null
null
null
null
null
harishnair04/Gemma-medtr-2b-sft-v2
null
null
"2024-11-09T23:54:50Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
256,000
256,001
gelu_pytorch_tanh
2,304
0.02
9,216
8,192
gemma2
8
26
4
0.000001
10,000
4,096
null
bfloat16
4.47.0.dev0
true
256,002
false
0
256
null
null
null
Zekunli/qwen2.5-7b-alpaca-discrim
null
null
"2024-11-10T00:10:41Z"
null
null
164
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
float32
4.45.0.dev0
true
152,064
null
0
null
null
null
null
rawsh/mirrorqwen2.5-0.5b-SimPO
null
null
"2024-11-10T00:10:51Z"
null
null
19
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "trl", "cpo", "unsloth", "arxiv:2401.08417", "base_model:rawsh/mirrorqwen2.5-0.5b-SFT", "base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float16
4.46.2
false
151,936
null
0
null
null
null
null
Treza12/Florence3
null
null
"2024-11-10T00:25:19Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "florence2", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "Florence2ForConditionalGeneration" ]
2
1
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float32
4.47.0.dev0
null
51,289
null
null
null
null
null
null
ahmed-tammam/telecom-gemma-2-finetuned-model
null
null
"2024-11-10T00:29:52Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu_pytorch_tanh
2,048
0.02
16,384
8,192
gemma
8
18
1
0.000001
10,000
null
null
float32
4.46.2
true
256,000
false
0
256
null
null
null
SolomonMartin/llama-2-7b-miniplatypus
null
null
"2024-11-10T01:13:54Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.46.2
true
32,000
false
0
128
false
1
null
Zekunli/gemma2-9b-alpaca-discrim
null
null
"2024-11-10T01:24:59Z"
null
null
53
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
float32
4.45.0.dev0
true
256,000
false
0
256
null
null
null
Clark88/Tinyllama-QNA-V1.4-40epochcolab
null
null
"2024-11-10T01:34:03Z"
null
null
64
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
5,632
2,048
llama
32
22
4
0.00001
10,000
null
false
float16
4.46.2
true
32,000
false
0
64
false
1
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft
null
null
"2024-11-10T01:59:11Z"
null
null
9
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float32
4.41.2
false
32,000
false
0
null
false
1
null
Treza12/Florence4
null
null
"2024-11-10T02:15:21Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "florence2", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "Florence2ForConditionalGeneration" ]
2
1
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float32
4.47.0.dev0
null
51,289
null
null
null
null
null
null
KSU-HW-SEC/LLama3_8B_assertion_ablation_cwe
null
null
"2024-11-10T02:16:18Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "llama-factory", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,009
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.46.1
true
128,256
false
0
128
false
1
null
itorgov/Qwen2-0.5B-Instruct-MLX
null
null
"2024-11-10T02:27:27Z"
null
null
17
null
null
null
null
0
null
[ "safetensors", "qwen2", "chat", "text-generation", "conversational", "en", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
32,768
true
bfloat16
4.40.1
true
151,936
null
0
null
null
null
null
namannn/short-llama
null
null
"2024-11-10T02:34:07Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.44.2
true
32,000
false
0
null
false
1
null
Treza12/Florence2
null
null
"2024-11-10T02:37:34Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "florence2", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "Florence2ForConditionalGeneration" ]
2
1
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float32
4.47.0.dev0
null
51,289
null
null
null
null
null
null
aalleavitch/HobieLLMFull
null
null
"2024-11-10T02:50:59Z"
null
null
42
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
131,072
mistral
48
56
8
0.00001
1,000,000
null
false
float16
4.46.2
true
32,768
null
0
128
null
null
null
Treza12/Florence1
null
null
"2024-11-10T03:29:29Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "florence2", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "region:us" ]
text-generation
null
null
0
[ "Florence2ForConditionalGeneration" ]
2
1
null
null
null
null
null
florence2
null
null
null
null
null
null
null
float32
4.47.0.dev0
null
51,289
null
null
null
null
null
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_lora_positive_dataset_v3
null
null
"2024-11-10T03:51:10Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
cgus/Qwen2.5-14B-Instruct-abliterated-v2-exl2
null
null
"2024-11-10T04:01:25Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "qwen2", "text-generation", "chat", "abliterated", "uncensored", "conversational", "en", "base_model:huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2", "base_model:quantized:huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "4-bit", "exl2", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
13,824
32,768
qwen2
40
48
8
0.000001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
rawsh/simpo-math-model
null
null
"2024-11-10T05:07:15Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "trl", "cpo", "unsloth", "arxiv:2401.08417", "base_model:rawsh/mirrorqwen2.5-0.5b-SFT", "base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float16
4.46.2
false
151,936
null
0
null
null
null
null
rawsh/mirrorqwen2.5-0.5b-SimPO-0
null
null
"2024-11-10T05:12:52Z"
null
null
20
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "trl", "cpo", "unsloth", "arxiv:2401.08417", "base_model:rawsh/mirrorqwen2.5-0.5b-SFT", "base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float16
4.46.2
false
151,936
null
0
null
null
null
null
pwork7/gemma7b_meta_math_1epoch
null
null
"2024-11-10T05:23:17Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu_pytorch_tanh
3,072
0.02
24,576
8,192
gemma
16
28
16
0.000001
10,000
null
null
bfloat16
4.43.3
false
256,000
false
0
256
null
null
null
yiran-wang3/ds_coder_reflct_adamw_iter5
null
null
"2024-11-10T06:07:38Z"
null
null
76
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/ds_coder_reflct_sppo_hard_new_cn_mining_oj_iter4-binarized-reflection-scored", "base_model:yiran-wang3/ds_coder_reflct_adamw_iter4", "base_model:finetune:yiran-wang3/ds_coder_reflct_adamw_iter4", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
100,000
100,015
silu
4,096
0.02
11,008
4,096
llama
32
30
32
0.000001
10,000
null
false
bfloat16
4.45.0
true
102,400
false
0
128
false
1
null
Rich-J/subnet29_upload_c02_N10_0
null
null
"2024-11-10T06:23:07Z"
null
null
106
null
null
null
null
0
transformers
[ "transformers", "safetensors", "phi3", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Phi3ForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
131,072
phi3
32
48
32
0.00001
10,000
262,144
false
bfloat16
4.44.0
false
32,064
false
0
null
null
null
null
kanishka/opt-babylm2-rewritten-clean-spacy-32k-earlystop_seed-42_3e-4
null
null
"2024-11-10T06:42:25Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "opt", "text-generation", "generated_from_trainer", "dataset:kanishka/babylm2-rewritten-clean-spacy", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "OPTForCausalLM" ]
1
1
null
768
null
null
256
opt
12
12
null
null
null
null
null
float32
4.45.1
true
32,768
null
0
null
null
null
null
t2ance/pretrain_inner_500
null
null
"2024-11-10T06:49:52Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
NeerjaK/Urdu_Model
null
null
"2024-11-10T06:50:03Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
256
0.02
688
64
llama
4
12
2
0.000001
10,000
null
false
float32
4.44.2
true
32,769
false
0
null
false
1
null
teka38/Qwen0.5-finetuned_v2
null
null
"2024-11-10T06:55:05Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
1,024
0.02
2,816
32,768
qwen2
16
24
16
0.000001
1,000,000
null
true
float32
4.44.2
true
151,936
null
0
null
null
null
null
t2ance/pretrain_inner_1000
null
null
"2024-11-10T07:00:25Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
c01zaut/gemma-2-27b-rk3588-1.1.2
null
null
"2024-11-10T07:05:09Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "arxiv:2009.03300", "arxiv:1905.07830", "arxiv:1911.11641", "arxiv:1904.09728", "arxiv:1905.10044", "arxiv:1907.10641", "arxiv:1811.00937", "arxiv:1809.02789", "arxiv:1911.01547", "arxiv:1705.03551", "arxiv:2107.03374", "arxiv:2108.07732", "arxiv:2110.14168", "arxiv:2009.11462", "arxiv:2101.11718", "arxiv:2110.08193", "arxiv:1804.09301", "arxiv:2109.07958", "arxiv:1804.06876", "arxiv:2103.03874", "arxiv:2304.06364", "arxiv:2206.04615", "arxiv:2203.09509", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
4,608
0.02
36,864
8,192
gemma2
32
46
16
0.000001
10,000
4,096
null
float32
4.42.0.dev0
true
256,000
false
0
128
null
null
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_gdr_lora_positive_dataset_v3
null
null
"2024-11-10T07:10:09Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
openerotica/Mistral-Small-Instruct-2409-abliterated-4.5bpw-exl2
null
null
"2024-11-10T07:52:50Z"
null
null
7
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "base_model:mistralai/Mistral-Small-Instruct-2409", "base_model:quantized:mistralai/Mistral-Small-Instruct-2409", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "exl2", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
32,768
mistral
48
56
8
0.00001
1,000,000
null
false
bfloat16
4.44.2
true
32,768
null
0
128
null
null
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_lora_positive_dataset_v1
null
null
"2024-11-10T08:04:25Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
DavidAU/MN-DARKEST-UNIVERSE-29B
null
null
"2024-11-10T08:07:50Z"
null
null
11
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
102
8
0.00001
1,000,000
null
false
bfloat16
4.46.0
true
131,074
null
0
128
null
null
null
moonloon/summary_gemma_ko
null
null
"2024-11-10T08:08:06Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu
2,048
0.02
16,384
8,192
gemma
8
18
1
0.000001
10,000
null
null
float32
4.46.2
true
256,000
false
0
256
null
null
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_klr_lora_positive_dataset_v4
null
null
"2024-11-10T08:09:44Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_gdr_lora_positive_dataset_v4
null
null
"2024-11-10T08:09:55Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
KR-X-AI/krx-qwen2.5-7b-instruct-v2
null
null
"2024-11-10T08:30:40Z"
null
null
57
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
pwork7/gemma7b_meta_math_2epoch
null
null
"2024-11-10T08:36:24Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu_pytorch_tanh
3,072
0.02
24,576
8,192
gemma
16
28
16
0.000001
10,000
null
null
bfloat16
4.43.3
false
256,000
false
0
256
null
null
null
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_klr_lora_positive_dataset_v4
null
null
"2024-11-10T08:46:11Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.41.2
false
32,000
false
0
null
false
1
null
Aurora-Gem/Opt_lora16_qwen2.5_14B_model_25k-1109
null
null
"2024-11-10T08:56:37Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "unsloth", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
13,824
131,072
qwen2
40
48
8
0.00001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
yiran-wang3/ds_coder_reflct_adamw_iter6
null
null
"2024-11-10T08:59:29Z"
null
null
9
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "alignment-handbook", "generated_from_trainer", "trl", "dpo", "conversational", "dataset:self-generate/ds_coder_reflct_sppo_hard_new_cn_mining_oj_iter5-binarized-reflection-scored", "base_model:yiran-wang3/ds_coder_reflct_adamw_iter5", "base_model:finetune:yiran-wang3/ds_coder_reflct_adamw_iter5", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
100,000
100,015
silu
4,096
0.02
11,008
4,096
llama
32
30
32
0.000001
10,000
null
false
bfloat16
4.45.0
true
102,400
false
0
128
false
1
null
pwork7/gemma7b_meta_math_3epoch
null
null
"2024-11-10T09:01:28Z"
null
null
11
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu_pytorch_tanh
3,072
0.02
24,576
8,192
gemma
16
28
16
0.000001
10,000
null
null
bfloat16
4.43.3
false
256,000
false
0
256
null
null
null
t2ance/slim_full_pretrain_inner_5000
null
null
"2024-11-10T09:25:31Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
khairi/Codex-148M
null
null
"2024-11-10T09:49:01Z"
null
null
27
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
0
0
silu
576
0.041667
1,536
8,192
llama
9
30
3
0.00001
100,000
null
true
float32
4.45.2
true
72,846
false
0
64
false
1
null
second-state/Qwen2.5-Coder-0.5B-Instruct-GGUF
null
null
"2024-11-10T10:00:16Z"
null
null
200
null
null
null
null
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "en", "base_model:Qwen/Qwen2.5-Coder-0.5B-Instruct", "base_model:quantized:Qwen/Qwen2.5-Coder-0.5B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
32,768
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null
t2ance/slim_full_pretrain_inner_10000
null
null
"2024-11-10T10:04:49Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
ihughes15234/phi35_tictactoe_dpo6epoch_v2
null
null
"2024-11-10T10:05:41Z"
null
null
51
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:ihughes15234/phi_3_5_mini_tictactoe1200", "base_model:finetune:ihughes15234/phi_3_5_mini_tictactoe1200", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
131,072
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.46.2
true
32,064
false
0
96
false
1
null
gaianet/Qwen2.5-Coder-0.5B-Instruct-GGUF
null
null
"2024-11-10T10:22:47Z"
null
null
104
null
null
null
null
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "en", "base_model:Qwen/Qwen2.5-Coder-0.5B-Instruct", "base_model:quantized:Qwen/Qwen2.5-Coder-0.5B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
32,768
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null
cnfusion/Rombos-LLM-V2.5-Qwen-32b-Q8-mlx
null
null
"2024-11-10T10:30:43Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mlx", "conversational", "base_model:rombodawg/Rombos-LLM-V2.5-Qwen-32b", "base_model:quantized:rombodawg/Rombos-LLM-V2.5-Qwen-32b", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
t2ance/slim_full_pretrain_inner_15000
null
null
"2024-11-10T10:43:05Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
DavidAU/MN-Dark-Horror-The-Cliffhanger-18.5B
null
null
"2024-11-10T10:47:58Z"
null
null
7
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
63
8
0.00001
1,000,000
null
false
bfloat16
4.46.0
true
131,074
null
0
128
null
null
null
seinare/para-gpt2
null
null
"2024-11-10T11:02:38Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float16
4.45.1
true
50,260
null
null
null
null
null
null
jaeyong2/Qwen2.5-3B-Instruct-Hi-SFT
null
null
"2024-11-10T11:16:15Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "hi", "en", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float32
4.40.0.dev0
false
151,936
null
0
null
null
null
null
jaeyong2/Qwen2.5-3B-Instruct-Id-SFT
null
null
"2024-11-10T11:19:53Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "hi", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float32
4.40.0.dev0
false
151,936
null
0
null
null
null
null
t2ance/slim_full_pretrain_inner_20000
null
null
"2024-11-10T11:21:08Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
amanpreetsingh459/gemma-2-2b-punjabi-finetuned
null
null
"2024-11-10T11:26:29Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "trl", "sft", "pa", "dataset:HydraIndicLM/punjabi_alpaca_52K", "arxiv:1910.09700", "base_model:google/gemma-2-2b", "base_model:finetune:google/gemma-2-2b", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
2,304
0.02
9,216
8,192
gemma2
8
26
4
0.000001
10,000
4,096
null
bfloat16
4.46.2
true
256,000
false
0
256
null
null
null
ihughes15234/phi35_tictactoe_dpo11epoch_v2
null
null
"2024-11-10T11:29:50Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:ihughes15234/phi35_tictactoe_dpo6epoch_v2", "base_model:finetune:ihughes15234/phi35_tictactoe_dpo6epoch_v2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
131,072
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.46.2
true
32,064
false
0
96
false
1
null
roger33303/mistral-7b-Instruct-Finetune-website-QnA
null
null
"2024-11-10T11:42:35Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "en", "base_model:unsloth/mistral-7b-v0.3-bnb-4bit", "base_model:finetune:unsloth/mistral-7b-v0.3-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.46.2
true
32,768
null
0
128
null
null
null
Donfab31/Mistral-7B-Instruct-4bit
null
null
"2024-11-10T11:51:16Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
float16
4.47.0.dev0
true
32,000
null
0
128
null
null
null
Kendamarron/Misaki-1.1B-v0.1
null
null
"2024-11-10T11:58:41Z"
null
null
4
null
null
null
null
0
null
[ "safetensors", "llama", "text-generation", "ja", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
7,168
4,096
llama
16
12
16
0.00001
10,000
null
false
bfloat16
4.44.2
true
99,584
false
0
128
false
1
null
t2ance/slim_full_pretrain_inner_25000
null
null
"2024-11-10T11:59:24Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
makeAmericaGreatAgain/TrumpWall
null
null
"2024-11-10T12:07:44Z"
null
null
49
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
896
0.02
4,864
32,768
qwen2
14
24
2
0.000001
1,000,000
null
true
float32
4.47.0.dev0
true
151,666
null
0
null
null
null
null
minhnguyent546/Alpaca-Llama-2-7b-chat
null
null
"2024-11-10T12:27:25Z"
null
null
74
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.42.3
true
32,000
false
0
null
false
1
null
KR-X-AI/krx-qwen2.5-7b-instruct-v2-m
null
null
"2024-11-10T12:34:56Z"
null
null
16
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:KR-X-AI/krx-qwen2.5-7b-instruct-v2", "base_model:finetune:KR-X-AI/krx-qwen2.5-7b-instruct-v2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
t2ance/slim_full_pretrain_inner_30000
null
null
"2024-11-10T12:37:23Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
AlexeyRyzhikov/mistral7b_text_to_json_v2.1
null
null
"2024-11-10T12:50:29Z"
null
null
37
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
10,000
4,096
false
bfloat16
4.46.2
true
32,000
null
0
128
null
null
null
appvoid/arco-mini-beta
null
null
"2024-11-10T13:14:24Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
float16
4.46.2
true
49,152
false
0
64
false
1
null
t2ance/slim_full_pretrain_inner_35000
null
null
"2024-11-10T13:15:20Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
oodeh/ods-ci-mistral-r64-a16-epoch-19-merged-model
null
null
"2024-11-10T13:29:02Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "base_model:unsloth/mistral-7b-instruct-v0.3-bnb-4bit", "base_model:quantized:unsloth/mistral-7b-instruct-v0.3-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.46.1
true
32,768
null
0
128
null
null
null
iyubondyrev/llama_nlp_2024
null
null
"2024-11-10T13:33:31Z"
null
null
39
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "custom_code", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LLAMA" ]
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
float32
4.46.2
null
32,000
null
null
null
null
null
null
KR-X-AI/krx-qwen2.5-7b-instruct-v2-m-e
null
null
"2024-11-10T13:43:55Z"
null
null
23
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "conversational", "en", "base_model:KR-X-AI/krx-qwen2.5-7b-instruct-v2-m", "base_model:finetune:KR-X-AI/krx-qwen2.5-7b-instruct-v2-m", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
t2ance/slim_full_pretrain_inner_40000
null
null
"2024-11-10T13:53:27Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
second-state/Qwen2.5-Coder-3B-Instruct-GGUF
null
null
"2024-11-10T14:00:34Z"
null
null
106
null
null
null
null
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "en", "base_model:Qwen/Qwen2.5-Coder-3B-Instruct", "base_model:quantized:Qwen/Qwen2.5-Coder-3B-Instruct", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
32,768
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null
gaianet/Qwen2.5-Coder-3B-Instruct-GGUF
null
null
"2024-11-10T14:00:52Z"
null
null
104
null
null
null
null
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "en", "base_model:Qwen/Qwen2.5-Coder-3B-Instruct", "base_model:quantized:Qwen/Qwen2.5-Coder-3B-Instruct", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
32,768
true
bfloat16
4.43.1
true
151,936
null
0
null
null
null
null