id
stringlengths 7
117
| author
stringclasses 6
values | sha
null | created_at
unknown | last_modified
null | disabled
null | downloads
int64 0
18.6M
| downloads_all_time
null | gated
bool 1
class | gguf
null | inference
null | likes
int64 0
4.77k
| library_name
stringclasses 36
values | tags
sequencelengths 1
430
| pipeline_tag
stringclasses 32
values | mask_token
null | model_index
null | trending_score
int64 0
132
| architectures
sequencelengths 1
5
⌀ | bos_token_id
int64 -1
256k
⌀ | eos_token_id
int64 -1
256k
⌀ | hidden_act
stringclasses 15
values | hidden_size
int64 1
20.5k
⌀ | initializer_range
float64 0
1
⌀ | intermediate_size
int64 1
98.3k
⌀ | max_position_embeddings
int64 8
1.05M
⌀ | model_type
stringclasses 530
values | num_attention_heads
int64 1
5k
⌀ | num_hidden_layers
int64 -1
8.93k
⌀ | num_key_value_heads
int64 1
160
⌀ | rms_norm_eps
float64 0
7
⌀ | rope_theta
float64 1k
1,000B
⌀ | sliding_window
int64 0
262k
⌀ | tie_word_embeddings
bool 2
classes | torch_dtype
stringclasses 8
values | transformers_version
stringclasses 207
values | use_cache
bool 2
classes | vocab_size
int64 -1
5.03M
⌀ | attention_bias
bool 2
classes | attention_dropout
float64 0
0.5
⌀ | head_dim
int64 2
256
⌀ | mlp_bias
bool 2
classes | pretraining_tp
int64 0
8
⌀ | rope_scaling
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Takvmi/model_pmc_gamma_0.1_noise0.2_0.2_epoch0 | null | null | "2024-11-09T18:16:29Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_kl_0.1_noise0.2_0.2_epoch0 | null | null | "2024-11-09T18:17:17Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_beta_0.1_noise0.2_0.2_epoch0 | null | null | "2024-11-09T18:18:21Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_beta_0.1_noise0.2_0.2_epoch1 | null | null | "2024-11-09T18:21:55Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
AndreyRzhaksinskiy/CDS-starcoder2-Ins-7b-E2E-20241109 | null | null | "2024-11-09T18:24:02Z" | null | null | 27 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"starcoder2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Starcoder2ForCausalLM"
] | 0 | 0 | gelu_pytorch_tanh | 4,608 | 0.018042 | 18,432 | 16,384 | starcoder2 | 36 | 32 | 4 | null | 1,000,000 | 4,096 | null | float16 | 4.46.0 | true | 49,152 | null | 0.1 | null | null | null | null |
VLKVLK/media-file-recognizer | null | null | "2024-11-09T18:25:28Z" | null | null | 12 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"llama-factory",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,536 | 0.02 | 8,960 | 32,768 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.44.2 | true | 151,936 | null | 0 | null | null | null | null |
Takvmi/model_pmc_kl_0.1_noise0.2_0.2_epoch1 | null | null | "2024-11-09T18:26:46Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_gamma_0.1_noise0.2_0.2_epoch1 | null | null | "2024-11-09T18:27:43Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Hachipo/chico-1.3b-4epoch_inst | null | null | "2024-11-09T18:33:56Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 7 | 7 | silu | 1,536 | 0.02 | 8,192 | 2,048 | llama | 24 | 24 | 24 | 0.000001 | 10,000 | null | false | float32 | 4.46.2 | true | 50,570 | false | 0 | 64 | false | 1 | null |
Uynaity/AutoTrain-Qwen-Rui-SHLR | null | null | "2024-11-09T19:06:46Z" | null | null | 22 | null | null | null | null | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"qwen2",
"text-generation",
"autotrain",
"text-generation-inference",
"peft",
"conversational",
"dataset:Uynaity/Rui-Pro",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-3B-Instruct",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | true | 151,665 | null | 0 | null | null | null | null |
kanishka/opt-babylm2-rewritten-clean-spacy-32k-earlystop_seed-42_1e-3 | null | null | "2024-11-09T19:15:11Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/babylm2-rewritten-clean-spacy",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"OPTForCausalLM"
] | 1 | 1 | null | 768 | null | null | 256 | opt | 12 | 12 | null | null | null | null | null | float32 | 4.45.1 | true | 32,768 | null | 0 | null | null | null | null |
pavl0/LLaMA5M | null | null | "2024-11-09T19:15:18Z" | null | null | 48 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"pavl0_llama",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMA"
] | null | null | null | null | null | null | null | pavl0_llama | null | null | null | null | null | null | null | float32 | 4.46.2 | null | 32,000 | null | null | null | null | null | null |
KaraKaraWitch/HiTZ-GoLLIE-13B-AsSafeTensors | null | null | "2024-11-09T19:44:39Z" | null | null | 20 | null | null | null | null | 0 | null | [
"safetensors",
"llama",
"code",
"text-generation-inference",
"Information Extraction",
"IE",
"Named Entity Recogniton",
"Event Extraction",
"Relation Extraction",
"LLaMA",
"text-generation",
"custom_code",
"en",
"dataset:ACE05",
"dataset:bc5cdr",
"dataset:conll2003",
"dataset:ncbi_disease",
"dataset:conll2012_ontonotesv5",
"dataset:rams",
"dataset:tacred",
"dataset:wnut_17",
"arxiv:2310.03668",
"license:llama2",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 13,824 | 16,384 | llama | 40 | 40 | 40 | 0.00001 | 1,000,000 | null | false | float32 | 4.44.2 | true | 32,016 | false | 0 | null | false | 1 | null |
pavl0/LLaMA5M_config | null | null | "2024-11-09T19:59:04Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"pavl0_llama",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMA"
] | null | null | null | null | null | null | null | pavl0_llama | null | null | null | null | null | null | null | float32 | 4.46.2 | null | 32,000 | null | null | null | null | null | null |
RyanYr/self-reflect_ministral8Bit_math_gsm8k_dpo_iter1 | null | null | "2024-11-09T20:18:14Z" | null | null | 33 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"arxiv:2305.18290",
"base_model:mistralai/Ministral-8B-Instruct-2410",
"base_model:finetune:mistralai/Ministral-8B-Instruct-2410",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 12,288 | 32,768 | mistral | 32 | 36 | 8 | 0.00001 | 100,000,000 | 32,768 | false | bfloat16 | 4.45.2 | false | 131,073 | null | 0 | 128 | null | null | null |
ThatsGroes/gemma-2-27b-it-SkoleGPT | null | null | "2024-11-09T21:08:51Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"base_model:unsloth/gemma-2-27b-it",
"base_model:finetune:unsloth/gemma-2-27b-it",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 4,608 | 0.02 | 36,864 | 8,192 | gemma2 | 32 | 46 | 16 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.47.0.dev0 | true | 256,000 | false | 0 | 128 | null | null | null |
Limby/qwenfloat16 | null | null | "2024-11-09T21:27:08Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | true | 151,936 | null | 0 | null | null | null | null |
yacioo/yassio-17-ben-belgacem | null | null | "2024-11-09T21:56:56Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"phi",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"PhiForCausalLM"
] | 50,256 | 50,256 | gelu_new | 2,560 | 0.02 | 10,240 | 2,048 | phi | 32 | 32 | 32 | null | 10,000 | null | false | float16 | 4.46.2 | true | 51,200 | null | 0 | null | null | null | null |
Takvmi/model_pmc_kl_0.1_noise0.1_0.1_epoch0 | null | null | "2024-11-09T22:09:38Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_kl_0.1_noise0.2_0.1_epoch0 | null | null | "2024-11-09T22:11:05Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
Takvmi/model_pmc_beta_0.05_noise0.2_0.1_epoch0 | null | null | "2024-11-09T22:13:35Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,258 | null | null | null | null | null | null |
tanquangduong/Qwen2.5-3B-Instruct-TinyStories | null | null | "2024-11-09T22:17:58Z" | null | null | 52 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"en",
"base_model:unsloth/Qwen2.5-3B",
"base_model:finetune:unsloth/Qwen2.5-3B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.46.2 | true | 151,936 | null | 0 | null | null | null | null |
TDN-M/tknv1 | null | null | "2024-11-09T22:26:37Z" | null | null | 30 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-3B-Instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.44.2 | true | 151,936 | null | 0 | null | null | null | null |
yakazimir/simpo-exps | null | null | "2024-11-09T22:50:09Z" | null | null | 14 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"alignment-handbook",
"trl",
"simpo",
"generated_from_trainer",
"conversational",
"dataset:princeton-nlp/llama3-ultrafeedback-armorm",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.2 | true | 128,256 | false | 0 | null | false | 1 | null |
thirdeyeai/SmolLM2-1.7B-Instruct-Uncensored | null | null | "2024-11-09T23:01:51Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"base_model:finetune:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 8,192 | 8,192 | llama | 32 | 24 | 32 | 0.00001 | 130,000 | null | true | bfloat16 | 4.44.2 | true | 49,152 | false | 0 | null | false | 1 | null |
henryz2004/Qwen-1_8B-Chat-lens | null | null | "2024-11-09T23:43:48Z" | null | null | 2 | null | null | null | null | 0 | null | [
"safetensors",
"qwen",
"text-generation",
"custom_code",
"zh",
"en",
"arxiv:2309.16609",
"arxiv:2305.08322",
"arxiv:2009.03300",
"region:us"
] | text-generation | null | null | 0 | [
"QWenLMHeadModel"
] | null | null | null | 2,048 | 0.02 | 11,008 | 8,192 | qwen | 16 | 24 | null | null | null | null | false | null | 4.32.0 | true | 151,936 | null | null | null | null | null | null |
harishnair04/Gemma-medtr-2b-sft-v2 | null | null | "2024-11-09T23:54:50Z" | null | null | 12 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 256,000 | 256,001 | gelu_pytorch_tanh | 2,304 | 0.02 | 9,216 | 8,192 | gemma2 | 8 | 26 | 4 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.47.0.dev0 | true | 256,002 | false | 0 | 256 | null | null | null |
Zekunli/qwen2.5-7b-alpaca-discrim | null | null | "2024-11-10T00:10:41Z" | null | null | 164 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | float32 | 4.45.0.dev0 | true | 152,064 | null | 0 | null | null | null | null |
rawsh/mirrorqwen2.5-0.5b-SimPO | null | null | "2024-11-10T00:10:51Z" | null | null | 19 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"trl",
"cpo",
"unsloth",
"arxiv:2401.08417",
"base_model:rawsh/mirrorqwen2.5-0.5b-SFT",
"base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | false | 151,936 | null | 0 | null | null | null | null |
Treza12/Florence3 | null | null | "2024-11-10T00:25:19Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"florence2",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Florence2ForConditionalGeneration"
] | 2 | 1 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 51,289 | null | null | null | null | null | null |
ahmed-tammam/telecom-gemma-2-finetuned-model | null | null | "2024-11-10T00:29:52Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 2,048 | 0.02 | 16,384 | 8,192 | gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | float32 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
SolomonMartin/llama-2-7b-miniplatypus | null | null | "2024-11-10T01:13:54Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.46.2 | true | 32,000 | false | 0 | 128 | false | 1 | null |
Zekunli/gemma2-9b-alpaca-discrim | null | null | "2024-11-10T01:24:59Z" | null | null | 53 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | float32 | 4.45.0.dev0 | true | 256,000 | false | 0 | 256 | null | null | null |
Clark88/Tinyllama-QNA-V1.4-40epochcolab | null | null | "2024-11-10T01:34:03Z" | null | null | 64 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 5,632 | 2,048 | llama | 32 | 22 | 4 | 0.00001 | 10,000 | null | false | float16 | 4.46.2 | true | 32,000 | false | 0 | 64 | false | 1 | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft | null | null | "2024-11-10T01:59:11Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float32 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
Treza12/Florence4 | null | null | "2024-11-10T02:15:21Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"florence2",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Florence2ForConditionalGeneration"
] | 2 | 1 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 51,289 | null | null | null | null | null | null |
KSU-HW-SEC/LLama3_8B_assertion_ablation_cwe | null | null | "2024-11-10T02:16:18Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-factory",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.1 | true | 128,256 | false | 0 | 128 | false | 1 | null |
itorgov/Qwen2-0.5B-Instruct-MLX | null | null | "2024-11-10T02:27:27Z" | null | null | 17 | null | null | null | null | 0 | null | [
"safetensors",
"qwen2",
"chat",
"text-generation",
"conversational",
"en",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.40.1 | true | 151,936 | null | 0 | null | null | null | null |
namannn/short-llama | null | null | "2024-11-10T02:34:07Z" | null | null | 18 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
Treza12/Florence2 | null | null | "2024-11-10T02:37:34Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"florence2",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Florence2ForConditionalGeneration"
] | 2 | 1 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 51,289 | null | null | null | null | null | null |
aalleavitch/HobieLLMFull | null | null | "2024-11-10T02:50:59Z" | null | null | 42 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 6,144 | 0.02 | 16,384 | 131,072 | mistral | 48 | 56 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.2 | true | 32,768 | null | 0 | 128 | null | null | null |
Treza12/Florence1 | null | null | "2024-11-10T03:29:29Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"florence2",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Florence2ForConditionalGeneration"
] | 2 | 1 | null | null | null | null | null | florence2 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 51,289 | null | null | null | null | null | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_lora_positive_dataset_v3 | null | null | "2024-11-10T03:51:10Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
cgus/Qwen2.5-14B-Instruct-abliterated-v2-exl2 | null | null | "2024-11-10T04:01:25Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"qwen2",
"text-generation",
"chat",
"abliterated",
"uncensored",
"conversational",
"en",
"base_model:huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
"base_model:quantized:huihui-ai/Qwen2.5-14B-Instruct-abliterated-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"exl2",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 13,824 | 32,768 | qwen2 | 40 | 48 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
rawsh/simpo-math-model | null | null | "2024-11-10T05:07:15Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"trl",
"cpo",
"unsloth",
"arxiv:2401.08417",
"base_model:rawsh/mirrorqwen2.5-0.5b-SFT",
"base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | false | 151,936 | null | 0 | null | null | null | null |
rawsh/mirrorqwen2.5-0.5b-SimPO-0 | null | null | "2024-11-10T05:12:52Z" | null | null | 20 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"trl",
"cpo",
"unsloth",
"arxiv:2401.08417",
"base_model:rawsh/mirrorqwen2.5-0.5b-SFT",
"base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SFT",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | false | 151,936 | null | 0 | null | null | null | null |
pwork7/gemma7b_meta_math_1epoch | null | null | "2024-11-10T05:23:17Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,072 | 0.02 | 24,576 | 8,192 | gemma | 16 | 28 | 16 | 0.000001 | 10,000 | null | null | bfloat16 | 4.43.3 | false | 256,000 | false | 0 | 256 | null | null | null |
yiran-wang3/ds_coder_reflct_adamw_iter5 | null | null | "2024-11-10T06:07:38Z" | null | null | 76 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"dataset:self-generate/ds_coder_reflct_sppo_hard_new_cn_mining_oj_iter4-binarized-reflection-scored",
"base_model:yiran-wang3/ds_coder_reflct_adamw_iter4",
"base_model:finetune:yiran-wang3/ds_coder_reflct_adamw_iter4",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 100,000 | 100,015 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 30 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.45.0 | true | 102,400 | false | 0 | 128 | false | 1 | null |
Rich-J/subnet29_upload_c02_N10_0 | null | null | "2024-11-10T06:23:07Z" | null | null | 106 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Phi3ForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | phi3 | 32 | 48 | 32 | 0.00001 | 10,000 | 262,144 | false | bfloat16 | 4.44.0 | false | 32,064 | false | 0 | null | null | null | null |
kanishka/opt-babylm2-rewritten-clean-spacy-32k-earlystop_seed-42_3e-4 | null | null | "2024-11-10T06:42:25Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/babylm2-rewritten-clean-spacy",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"OPTForCausalLM"
] | 1 | 1 | null | 768 | null | null | 256 | opt | 12 | 12 | null | null | null | null | null | float32 | 4.45.1 | true | 32,768 | null | 0 | null | null | null | null |
t2ance/pretrain_inner_500 | null | null | "2024-11-10T06:49:52Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
NeerjaK/Urdu_Model | null | null | "2024-11-10T06:50:03Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 256 | 0.02 | 688 | 64 | llama | 4 | 12 | 2 | 0.000001 | 10,000 | null | false | float32 | 4.44.2 | true | 32,769 | false | 0 | null | false | 1 | null |
teka38/Qwen0.5-finetuned_v2 | null | null | "2024-11-10T06:55:05Z" | null | null | 18 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 1,024 | 0.02 | 2,816 | 32,768 | qwen2 | 16 | 24 | 16 | 0.000001 | 1,000,000 | null | true | float32 | 4.44.2 | true | 151,936 | null | 0 | null | null | null | null |
t2ance/pretrain_inner_1000 | null | null | "2024-11-10T07:00:25Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
c01zaut/gemma-2-27b-rk3588-1.1.2 | null | null | "2024-11-10T07:05:09Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"arxiv:2009.03300",
"arxiv:1905.07830",
"arxiv:1911.11641",
"arxiv:1904.09728",
"arxiv:1905.10044",
"arxiv:1907.10641",
"arxiv:1811.00937",
"arxiv:1809.02789",
"arxiv:1911.01547",
"arxiv:1705.03551",
"arxiv:2107.03374",
"arxiv:2108.07732",
"arxiv:2110.14168",
"arxiv:2009.11462",
"arxiv:2101.11718",
"arxiv:2110.08193",
"arxiv:1804.09301",
"arxiv:2109.07958",
"arxiv:1804.06876",
"arxiv:2103.03874",
"arxiv:2304.06364",
"arxiv:2206.04615",
"arxiv:2203.09509",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 4,608 | 0.02 | 36,864 | 8,192 | gemma2 | 32 | 46 | 16 | 0.000001 | 10,000 | 4,096 | null | float32 | 4.42.0.dev0 | true | 256,000 | false | 0 | 128 | null | null | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_gdr_lora_positive_dataset_v3 | null | null | "2024-11-10T07:10:09Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
openerotica/Mistral-Small-Instruct-2409-abliterated-4.5bpw-exl2 | null | null | "2024-11-10T07:52:50Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"base_model:mistralai/Mistral-Small-Instruct-2409",
"base_model:quantized:mistralai/Mistral-Small-Instruct-2409",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"exl2",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 6,144 | 0.02 | 16,384 | 32,768 | mistral | 48 | 56 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 32,768 | null | 0 | 128 | null | null | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_lora_positive_dataset_v1 | null | null | "2024-11-10T08:04:25Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
DavidAU/MN-DARKEST-UNIVERSE-29B | null | null | "2024-11-10T08:07:50Z" | null | null | 11 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 102 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.0 | true | 131,074 | null | 0 | 128 | null | null | null |
moonloon/summary_gemma_ko | null | null | "2024-11-10T08:08:06Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu | 2,048 | 0.02 | 16,384 | 8,192 | gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | float32 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_ga_klr_lora_positive_dataset_v4 | null | null | "2024-11-10T08:09:44Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_gdr_lora_positive_dataset_v4 | null | null | "2024-11-10T08:09:55Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
KR-X-AI/krx-qwen2.5-7b-instruct-v2 | null | null | "2024-11-10T08:30:40Z" | null | null | 57 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
pwork7/gemma7b_meta_math_2epoch | null | null | "2024-11-10T08:36:24Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,072 | 0.02 | 24,576 | 8,192 | gemma | 16 | 28 | 16 | 0.000001 | 10,000 | null | null | bfloat16 | 4.43.3 | false | 256,000 | false | 0 | 256 | null | null | null |
concept-unlearning/Llama-2-7b-hf_ft_lora_all_novels_v1_ft_npo_klr_lora_positive_dataset_v4 | null | null | "2024-11-10T08:46:11Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.41.2 | false | 32,000 | false | 0 | null | false | 1 | null |
Aurora-Gem/Opt_lora16_qwen2.5_14B_model_25k-1109 | null | null | "2024-11-10T08:56:37Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"unsloth",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 13,824 | 131,072 | qwen2 | 40 | 48 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
yiran-wang3/ds_coder_reflct_adamw_iter6 | null | null | "2024-11-10T08:59:29Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"dataset:self-generate/ds_coder_reflct_sppo_hard_new_cn_mining_oj_iter5-binarized-reflection-scored",
"base_model:yiran-wang3/ds_coder_reflct_adamw_iter5",
"base_model:finetune:yiran-wang3/ds_coder_reflct_adamw_iter5",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 100,000 | 100,015 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 30 | 32 | 0.000001 | 10,000 | null | false | bfloat16 | 4.45.0 | true | 102,400 | false | 0 | 128 | false | 1 | null |
pwork7/gemma7b_meta_math_3epoch | null | null | "2024-11-10T09:01:28Z" | null | null | 11 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,072 | 0.02 | 24,576 | 8,192 | gemma | 16 | 28 | 16 | 0.000001 | 10,000 | null | null | bfloat16 | 4.43.3 | false | 256,000 | false | 0 | 256 | null | null | null |
t2ance/slim_full_pretrain_inner_5000 | null | null | "2024-11-10T09:25:31Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
khairi/Codex-148M | null | null | "2024-11-10T09:49:01Z" | null | null | 27 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 0 | 0 | silu | 576 | 0.041667 | 1,536 | 8,192 | llama | 9 | 30 | 3 | 0.00001 | 100,000 | null | true | float32 | 4.45.2 | true | 72,846 | false | 0 | 64 | false | 1 | null |
second-state/Qwen2.5-Coder-0.5B-Instruct-GGUF | null | null | "2024-11-10T10:00:16Z" | null | null | 200 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-0.5B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-0.5B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
t2ance/slim_full_pretrain_inner_10000 | null | null | "2024-11-10T10:04:49Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
ihughes15234/phi35_tictactoe_dpo6epoch_v2 | null | null | "2024-11-10T10:05:41Z" | null | null | 51 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:ihughes15234/phi_3_5_mini_tictactoe1200",
"base_model:finetune:ihughes15234/phi_3_5_mini_tictactoe1200",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 32,064 | false | 0 | 96 | false | 1 | null |
gaianet/Qwen2.5-Coder-0.5B-Instruct-GGUF | null | null | "2024-11-10T10:22:47Z" | null | null | 104 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-0.5B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-0.5B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
cnfusion/Rombos-LLM-V2.5-Qwen-32b-Q8-mlx | null | null | "2024-11-10T10:30:43Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mlx",
"conversational",
"base_model:rombodawg/Rombos-LLM-V2.5-Qwen-32b",
"base_model:quantized:rombodawg/Rombos-LLM-V2.5-Qwen-32b",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"8-bit",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
t2ance/slim_full_pretrain_inner_15000 | null | null | "2024-11-10T10:43:05Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
DavidAU/MN-Dark-Horror-The-Cliffhanger-18.5B | null | null | "2024-11-10T10:47:58Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 63 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.0 | true | 131,074 | null | 0 | 128 | null | null | null |
seinare/para-gpt2 | null | null | "2024-11-10T11:02:38Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float16 | 4.45.1 | true | 50,260 | null | null | null | null | null | null |
jaeyong2/Qwen2.5-3B-Instruct-Hi-SFT | null | null | "2024-11-10T11:16:15Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"hi",
"en",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-3B-Instruct",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.40.0.dev0 | false | 151,936 | null | 0 | null | null | null | null |
jaeyong2/Qwen2.5-3B-Instruct-Id-SFT | null | null | "2024-11-10T11:19:53Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"hi",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-3B-Instruct",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.40.0.dev0 | false | 151,936 | null | 0 | null | null | null | null |
t2ance/slim_full_pretrain_inner_20000 | null | null | "2024-11-10T11:21:08Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
amanpreetsingh459/gemma-2-2b-punjabi-finetuned | null | null | "2024-11-10T11:26:29Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"trl",
"sft",
"pa",
"dataset:HydraIndicLM/punjabi_alpaca_52K",
"arxiv:1910.09700",
"base_model:google/gemma-2-2b",
"base_model:finetune:google/gemma-2-2b",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 2,304 | 0.02 | 9,216 | 8,192 | gemma2 | 8 | 26 | 4 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
ihughes15234/phi35_tictactoe_dpo11epoch_v2 | null | null | "2024-11-10T11:29:50Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:ihughes15234/phi35_tictactoe_dpo6epoch_v2",
"base_model:finetune:ihughes15234/phi35_tictactoe_dpo6epoch_v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 32,064 | false | 0 | 96 | false | 1 | null |
roger33303/mistral-7b-Instruct-Finetune-website-QnA | null | null | "2024-11-10T11:42:35Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"en",
"base_model:unsloth/mistral-7b-v0.3-bnb-4bit",
"base_model:finetune:unsloth/mistral-7b-v0.3-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 32,768 | null | 0 | 128 | null | null | null |
Donfab31/Mistral-7B-Instruct-4bit | null | null | "2024-11-10T11:51:16Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.47.0.dev0 | true | 32,000 | null | 0 | 128 | null | null | null |
Kendamarron/Misaki-1.1B-v0.1 | null | null | "2024-11-10T11:58:41Z" | null | null | 4 | null | null | null | null | 0 | null | [
"safetensors",
"llama",
"text-generation",
"ja",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 7,168 | 4,096 | llama | 16 | 12 | 16 | 0.00001 | 10,000 | null | false | bfloat16 | 4.44.2 | true | 99,584 | false | 0 | 128 | false | 1 | null |
t2ance/slim_full_pretrain_inner_25000 | null | null | "2024-11-10T11:59:24Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
makeAmericaGreatAgain/TrumpWall | null | null | "2024-11-10T12:07:44Z" | null | null | 49 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.47.0.dev0 | true | 151,666 | null | 0 | null | null | null | null |
minhnguyent546/Alpaca-Llama-2-7b-chat | null | null | "2024-11-10T12:27:25Z" | null | null | 74 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.42.3 | true | 32,000 | false | 0 | null | false | 1 | null |
KR-X-AI/krx-qwen2.5-7b-instruct-v2-m | null | null | "2024-11-10T12:34:56Z" | null | null | 16 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:KR-X-AI/krx-qwen2.5-7b-instruct-v2",
"base_model:finetune:KR-X-AI/krx-qwen2.5-7b-instruct-v2",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
t2ance/slim_full_pretrain_inner_30000 | null | null | "2024-11-10T12:37:23Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
AlexeyRyzhikov/mistral7b_text_to_json_v2.1 | null | null | "2024-11-10T12:50:29Z" | null | null | 37 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | bfloat16 | 4.46.2 | true | 32,000 | null | 0 | 128 | null | null | null |
appvoid/arco-mini-beta | null | null | "2024-11-10T13:14:24Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 960 | 0.02 | 2,560 | 8,192 | llama | 15 | 32 | 5 | 0.00001 | 100,000 | null | true | float16 | 4.46.2 | true | 49,152 | false | 0 | 64 | false | 1 | null |
t2ance/slim_full_pretrain_inner_35000 | null | null | "2024-11-10T13:15:20Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
oodeh/ods-ci-mistral-r64-a16-epoch-19-merged-model | null | null | "2024-11-10T13:29:02Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:unsloth/mistral-7b-instruct-v0.3-bnb-4bit",
"base_model:quantized:unsloth/mistral-7b-instruct-v0.3-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.1 | true | 32,768 | null | 0 | 128 | null | null | null |
iyubondyrev/llama_nlp_2024 | null | null | "2024-11-10T13:33:31Z" | null | null | 39 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLAMA"
] | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | float32 | 4.46.2 | null | 32,000 | null | null | null | null | null | null |
KR-X-AI/krx-qwen2.5-7b-instruct-v2-m-e | null | null | "2024-11-10T13:43:55Z" | null | null | 23 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"krx",
"conversational",
"en",
"base_model:KR-X-AI/krx-qwen2.5-7b-instruct-v2-m",
"base_model:finetune:KR-X-AI/krx-qwen2.5-7b-instruct-v2-m",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
t2ance/slim_full_pretrain_inner_40000 | null | null | "2024-11-10T13:53:27Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
second-state/Qwen2.5-Coder-3B-Instruct-GGUF | null | null | "2024-11-10T14:00:34Z" | null | null | 106 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-3B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-3B-Instruct",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |
gaianet/Qwen2.5-Coder-3B-Instruct-GGUF | null | null | "2024-11-10T14:00:52Z" | null | null | 104 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-3B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-3B-Instruct",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | 32,768 | true | bfloat16 | 4.43.1 | true | 151,936 | null | 0 | null | null | null | null |