id
stringlengths
7
117
author
stringclasses
6 values
sha
null
created_at
unknown
last_modified
null
disabled
null
downloads
int64
0
18.6M
downloads_all_time
null
gated
bool
1 class
gguf
null
inference
null
likes
int64
0
4.77k
library_name
stringclasses
36 values
tags
sequencelengths
1
430
pipeline_tag
stringclasses
32 values
mask_token
null
model_index
null
trending_score
int64
0
132
architectures
sequencelengths
1
5
bos_token_id
int64
-1
256k
eos_token_id
int64
-1
256k
hidden_act
stringclasses
15 values
hidden_size
int64
1
20.5k
initializer_range
float64
0
1
intermediate_size
int64
1
98.3k
max_position_embeddings
int64
8
1.05M
model_type
stringclasses
530 values
num_attention_heads
int64
1
5k
num_hidden_layers
int64
-1
8.93k
num_key_value_heads
int64
1
160
rms_norm_eps
float64
0
7
rope_theta
float64
1k
1,000B
sliding_window
int64
0
262k
tie_word_embeddings
bool
2 classes
torch_dtype
stringclasses
8 values
transformers_version
stringclasses
207 values
use_cache
bool
2 classes
vocab_size
int64
-1
5.03M
attention_bias
bool
2 classes
attention_dropout
float64
0
0.5
head_dim
int64
2
256
mlp_bias
bool
2 classes
pretraining_tp
int64
0
8
rope_scaling
dict
jhllm/kospi_report_model_1107
null
null
"2024-11-07T04:32:34Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float32
4.41.0
false
128,256
false
0
null
false
1
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-07T04:33:08Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
llmflow/krx_Qwen2_7B_Instruct_v6
null
null
"2024-11-07T04:33:46Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
Leejy0-0/log_Power
null
null
"2024-11-07T04:38:44Z"
null
null
33
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu
2,048
0.02
16,384
8,192
gemma
8
18
1
0.000001
10,000
null
null
float32
4.38.2
true
256,000
false
0
256
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-07T04:45:55Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
motexture/SmolLCoder-360M-Instruct
null
null
"2024-11-07T04:52:56Z"
null
null
24
null
null
null
null
0
null
[ "safetensors", "llama", "smoll", "coding", "coder", "model", "small", "text-generation", "conversational", "en", "dataset:motexture/cData", "base_model:HuggingFaceTB/SmolLM2-360M-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-360M-Instruct", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
float16
4.46.1
true
49,152
false
0
64
false
1
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-07T04:58:47Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-07T05:09:33Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
llmflow/krx_Qwen2_7B_Instruct_v6_1
null
null
"2024-11-07T05:10:59Z"
null
null
13
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.44.2
true
152,064
null
0
null
null
null
null
homeb82784/Qwen2-7B-Instruct-FinanceSFT
null
null
"2024-11-07T05:17:48Z"
null
null
43
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.45.2
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-07T05:22:30Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
tnwjd68317/suv7_qwen2_lora
null
null
"2024-11-07T05:26:02Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "krx", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
float32
4.46.1
true
151,647
null
0
null
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-07T05:36:20Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-07T05:47:10Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1.4b-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-07T05:59:58Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
AlSamCur123/Mistral-Nemo-InstructContinuedFine
null
null
"2024-11-07T06:01:42Z"
null
null
137
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit", "base_model:finetune:unsloth/Mistral-Nemo-Instruct-2407-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.46.2
true
131,072
null
0
128
null
null
null
c01zaut/Qwen2.5-Coder-7B-Instruct-rk3588-1.1.2
null
null
"2024-11-07T06:09:17Z"
null
null
17
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "code", "codeqwen", "chat", "qwen", "qwen-coder", "conversational", "en", "arxiv:2409.12186", "arxiv:2309.00071", "arxiv:2407.10671", "base_model:Qwen/Qwen2.5-Coder-7B", "base_model:finetune:Qwen/Qwen2.5-Coder-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
131,072
false
bfloat16
4.43.1
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step1-GPTQ-wikitext2
null
null
"2024-11-07T06:12:14Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
SparseLLM/sparsing-law-0.2b-relu
null
null
"2024-11-07T06:15:39Z"
null
null
8
null
null
null
null
0
null
[ "pytorch", "text-generation", "custom_code", "en", "zh", "arxiv:2411.02335", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "MiniCPMForCausalLM" ]
1
2
relu
1,024
0.1
2,560
4,096
null
8
20
8
0.00001
null
null
true
bfloat16
4.36.0
true
122,753
null
null
null
null
null
null
kik41/4bit-lora-formality-formal-llama-3-8b
null
null
"2024-11-07T06:18:15Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
float32
4.45.2
true
128,257
false
0
128
false
1
null
Xu-Ouyang/pythia-1b-deduped-int8-step2-GPTQ-wikitext2
null
null
"2024-11-07T06:22:35Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-07T06:32:55Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
yoohj58072/krx_qwen2_7b_it_v4
null
null
"2024-11-07T06:36:50Z"
null
null
16
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "conversational", "en", "base_model:unsloth/Qwen2-7B-Instruct-bnb-4bit", "base_model:finetune:unsloth/Qwen2-7B-Instruct-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.1
true
152,064
null
0
null
null
null
null
IDEA-FinAI/TouchstoneGPT-7B-Instruct
null
null
"2024-11-07T06:37:52Z"
null
null
17
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "finance", "text-generation-inference", "conversational", "en", "zh", "dataset:IDEA-FinAI/Golden-Touchstone", "arxiv:2411.06272", "base_model:Qwen/Qwen2-7B-Instruct", "base_model:finetune:Qwen/Qwen2-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.43.4
false
152,064
null
0
null
null
null
null
voidful/SmolLM2-360M-Instruct-ASR
null
null
"2024-11-07T06:37:55Z"
null
null
44
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
float32
4.45.2
true
51,870
false
0
64
false
1
null
Xu-Ouyang/pythia-1b-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-07T06:43:14Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
AsphoxK/model_2024_10_29_q32_e1
null
null
"2024-11-07T06:47:54Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "starcoder2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Starcoder2ForCausalLM" ]
0
0
gelu_pytorch_tanh
3,072
0.018042
12,288
16,384
starcoder2
24
30
2
null
999,999.442036
4,096
null
float32
4.44.2
true
49,152
null
0.1
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-07T06:51:19Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
yfarm01/sn29_dec07_c0
null
null
"2024-11-07T07:00:46Z"
null
null
94
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
100,257
100,257
silu
4,096
0.02
18,752
4,096
llama
32
30
8
0.00001
500,000
null
false
bfloat16
4.44.0
false
100,288
false
0
128
false
1
null
Xu-Ouyang/pythia-1b-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-07T07:01:29Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-07T07:09:46Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-07T07:20:04Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-07T07:28:14Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-07T07:38:27Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
SparseLLM/sparsing-law-0.8b-relu
null
null
"2024-11-07T07:45:03Z"
null
null
10
null
null
null
null
0
null
[ "pytorch", "text-generation", "custom_code", "en", "zh", "arxiv:2411.02335", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "MiniCPMForCausalLM" ]
1
2
relu
1,536
0.1
3,840
4,096
null
12
28
12
0.00001
null
null
true
bfloat16
4.36.0
true
122,753
null
null
null
null
null
null
Xu-Ouyang/pythia-1b-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-07T07:46:39Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
2,048
0.02
8,192
2,048
gpt_neox
8
16
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step1-GPTQ-wikitext2
null
null
"2024-11-07T07:56:41Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
broalantap/GPT2-large-4-44000steps
null
null
"2024-11-07T07:59:34Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "pytorch", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
bfloat16
4.41.2
true
50,321
null
null
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step2-GPTQ-wikitext2
null
null
"2024-11-07T08:04:06Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
ethicsadvisorproject/Llama-2-7b-ethical-chat-finetune
null
null
"2024-11-07T08:05:07Z"
null
null
6
null
null
null
null
0
null
[ "pytorch", "llama", "Data", "text-generation", "en", "dataset:MasahiroKaneko/eagle", "base_model:NousResearch/Llama-2-7b-chat-hf", "base_model:finetune:NousResearch/Llama-2-7b-chat-hf", "license:mit", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
null
null
false
float16
4.31.0
true
32,000
null
null
null
null
1
null
Xu-Ouyang/pythia-410m-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-07T08:10:36Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
rewicks/monolingual_de_8k-shared_ep20
null
null
"2024-11-07T08:10:36Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "marian", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MarianForCausalLM" ]
1
2
null
null
null
null
256
marian
null
12
null
null
null
null
null
float32
4.39.3
false
8,000
null
0
null
null
null
null
Liedichi/Monstral-123B_8.0bpw_EXL2
null
null
"2024-11-07T08:16:44Z"
null
null
32
null
null
null
null
0
null
[ "safetensors", "mistral", "chat", "text-generation", "conversational", "en", "base_model:MarsupialAI/Monstral-123B", "base_model:quantized:MarsupialAI/Monstral-123B", "license:other", "8-bit", "exl2", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
12,288
0.02
28,672
131,072
mistral
96
88
8
0.00001
1,000,000
null
false
float16
4.44.2
true
32,768
null
0
128
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-07T08:18:03Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-07T08:24:32Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
KR-X-AI/krx-qwen2-7b-instruct-v4_mx
null
null
"2024-11-07T08:29:52Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "conversational", "en", "base_model:KR-X-AI/krx-qwen2-7b-instruct-v2", "base_model:finetune:KR-X-AI/krx-qwen2-7b-instruct-v2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.47.0.dev0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-07T08:31:02Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-07T08:37:34Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-07T08:44:04Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
kim400/qwen2.5_7b_it_v9
null
null
"2024-11-07T08:46:52Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "unsloth", "trl", "sft", "krx", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
GiKAGraphy/ProductLlama_V5_gguf
null
null
"2024-11-07T08:49:12Z"
null
null
116
null
null
null
null
0
transformers
[ "transformers", "gguf", "llama", "unsloth", "text-generation-inference", "text-generation", "en", "base_model:Sayantan54321/model_ff", "base_model:quantized:Sayantan54321/model_ff", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
llama
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-07T08:50:33Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-07T08:57:56Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-410m-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-07T09:04:25Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
1,024
0.02
4,096
2,048
gpt_neox
16
24
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
SparseLLM/sparsing-law-0.4b-relu
null
null
"2024-11-07T09:07:05Z"
null
null
2
null
null
null
null
0
null
[ "pytorch", "text-generation", "custom_code", "en", "zh", "arxiv:2411.02335", "license:apache-2.0", "region:us" ]
text-generation
null
null
0
[ "MiniCPMForCausalLM" ]
1
2
relu
1,280
0.1
3,200
4,096
null
10
24
10
0.00001
null
null
true
bfloat16
4.36.0
true
122,753
null
null
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step1-GPTQ-wikitext2
null
null
"2024-11-07T09:07:59Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step2-GPTQ-wikitext2
null
null
"2024-11-07T09:10:38Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
seong67360/Qwen2.5-7B-Instruct_v4
null
null
"2024-11-07T09:11:51Z"
null
null
24
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "krx", "conversational", "dataset:gbharti/finance-alpaca", "dataset:sujet-ai/Sujet-Finance-Instruct-177k", "base_model:Qwen/Qwen2.5-7B-Instruct", "base_model:finetune:Qwen/Qwen2.5-7B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
float32
4.47.0.dev0
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step4-GPTQ-wikitext2
null
null
"2024-11-07T09:13:45Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Uynaity/AutoTrain-Qwen-Rui-HLR
null
null
"2024-11-07T09:14:08Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "qwen2", "text-generation", "autotrain", "text-generation-inference", "peft", "conversational", "dataset:Uynaity/Rui-Pro", "base_model:Qwen/Qwen2.5-3B-Instruct", "base_model:finetune:Qwen/Qwen2.5-3B-Instruct", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
2,048
0.02
11,008
32,768
qwen2
16
36
2
0.000001
1,000,000
null
true
float16
4.45.0
true
151,669
null
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step8-GPTQ-wikitext2
null
null
"2024-11-07T09:16:27Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step16-GPTQ-wikitext2
null
null
"2024-11-07T09:19:31Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
allknowingroger/QwenTask1-32B
null
null
"2024-11-07T09:21:06Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2212.04089", "base_model:EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2", "base_model:merge:EVA-UNIT-01/EVA-Qwen2.5-32B-v0.2", "base_model:Qwen/Qwen2.5-32B", "base_model:merge:Qwen/Qwen2.5-32B", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:merge:Qwen/Qwen2.5-32B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
5,120
0.02
27,648
131,072
qwen2
40
64
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
true
152,064
null
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step32-GPTQ-wikitext2
null
null
"2024-11-07T09:22:35Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step64-GPTQ-wikitext2
null
null
"2024-11-07T09:25:15Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step128-GPTQ-wikitext2
null
null
"2024-11-07T09:28:22Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step256-GPTQ-wikitext2
null
null
"2024-11-07T09:31:04Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step512-GPTQ-wikitext2
null
null
"2024-11-07T09:33:45Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
Xu-Ouyang/pythia-160m-deduped-int8-step1000-GPTQ-wikitext2
null
null
"2024-11-07T09:36:53Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "8-bit", "gptq", "region:us" ]
text-generation
null
null
0
[ "GPTNeoXForCausalLM" ]
0
0
gelu
768
0.02
3,072
2,048
gpt_neox
12
12
null
null
null
null
false
float16
4.41.2
true
50,304
true
0
null
null
null
null
cfstar188/gemma2-Code-Instruct-Finetune-test
null
null
"2024-11-07T09:44:52Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
float16
4.46.2
true
256,000
false
0
256
null
null
null
allknowingroger/QwenTask2-32B
null
null
"2024-11-07T09:47:17Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2212.04089", "base_model:Qwen/Qwen2.5-32B-Instruct", "base_model:merge:Qwen/Qwen2.5-32B-Instruct", "base_model:gctian/qwen2.5-32B-roleplay-zh", "base_model:merge:gctian/qwen2.5-32B-roleplay-zh", "base_model:zetasepic/Qwen2.5-32B-Instruct-abliterated-pass2", "base_model:merge:zetasepic/Qwen2.5-32B-Instruct-abliterated-pass2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
5,120
0.02
27,648
32,768
qwen2
40
64
8
0.000001
1,000,000
null
false
bfloat16
4.45.1
true
152,064
null
0
null
null
null
null
IlyaGusev/saiga_nemo_12b_sft_m9_d16_slerp
null
null
"2024-11-07T09:57:40Z"
null
null
103
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
1,024,000
mistral
32
40
8
0.00001
1,000,000
null
false
bfloat16
4.45.1
true
131,072
null
0
128
null
null
null
mashleburneded/Mobiusv0.2
null
null
"2024-11-07T10:01:40Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:mistralai/Codestral-22B-v0.1", "base_model:merge:mistralai/Codestral-22B-v0.1", "base_model:mistralai/Mathstral-7B-v0.1", "base_model:merge:mistralai/Mathstral-7B-v0.1", "base_model:mistralai/Mistral-Large-Instruct-2407", "base_model:merge:mistralai/Mistral-Large-Instruct-2407", "base_model:mistralai/Mistral-Nemo-Base-2407", "base_model:merge:mistralai/Mistral-Nemo-Base-2407", "base_model:nvidia/Mistral-NeMo-Minitron-8B-Instruct", "base_model:merge:nvidia/Mistral-NeMo-Minitron-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
6,144
0.02
16,384
32,768
mistral
48
40
8
0.00001
1,000,000
null
false
float16
4.46.2
true
32,768
null
0
128
null
null
null
Diplomkaazvposlednimsemestru/Textovy_model
null
null
"2024-11-07T10:24:02Z"
null
null
2
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.46.2
true
50,257
null
null
null
null
null
null
Haaaaarsh/gemma_ac_model
null
null
"2024-11-07T10:34:33Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:unsloth/gemma-2b-it-bnb-4bit", "base_model:finetune:unsloth/gemma-2b-it-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GemmaForCausalLM" ]
2
1
gelu
2,048
0.02
16,384
8,192
gemma
8
18
1
0.000001
10,000
null
null
float16
4.46.2
true
256,000
false
0
256
null
null
null
phongtintruong/Mistrava-7B-Instruct-v0.30-noft
null
null
"2024-11-07T10:49:41Z"
null
null
23
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
bfloat16
4.41.2
true
37,904
null
0
null
null
null
null
lazydok/Qwen2-7B-it-finDc-v0.2
null
null
"2024-11-07T10:52:43Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "base_model:unsloth/Qwen2-7B-Instruct-bnb-4bit", "base_model:finetune:unsloth/Qwen2-7B-Instruct-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.0
true
152,064
null
0
null
null
null
null
MaziyarPanahi/Mistral-Crab-SFT-GGUF
null
null
"2024-11-07T10:59:51Z"
null
null
76
null
null
null
null
0
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:THU-KEG/Mistral-Crab-SFT", "base_model:quantized:THU-KEG/Mistral-Crab-SFT", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
joyjimin3070/krx-qwen2.5-7b-instruction-v3
null
null
"2024-11-07T11:10:19Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "pytorch", "qwen2", "text-generation", "unsloth", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
float16
4.46.2
true
152,064
null
0
null
null
null
null
homeb82784/Qwen2-7B-Instruct-it-v1.0-v3.0
null
null
"2024-11-07T11:19:25Z"
null
null
35
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "unsloth", "trl", "sft", "krx", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.45.2
false
152,064
null
0
null
null
null
null
overfit-brothers/Qwen2-7B-smarter-Instruct
null
null
"2024-11-07T11:30:37Z"
null
null
16
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "krx", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
mashleburneded/Mobiusv0.3
null
null
"2024-11-07T11:32:30Z"
null
null
15
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "base_model:mistralai/Codestral-22B-v0.1", "base_model:merge:mistralai/Codestral-22B-v0.1", "base_model:mistralai/Mathstral-7B-v0.1", "base_model:merge:mistralai/Mathstral-7B-v0.1", "base_model:mistralai/Mistral-Large-Instruct-2407", "base_model:merge:mistralai/Mistral-Large-Instruct-2407", "base_model:mistralai/Mistral-Nemo-Base-2407", "base_model:merge:mistralai/Mistral-Nemo-Base-2407", "base_model:nvidia/Mistral-NeMo-Minitron-8B-Instruct", "base_model:merge:nvidia/Mistral-NeMo-Minitron-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
5,120
0.02
14,336
131,072
mistral
32
40
8
0.00001
1,000,000
null
false
float16
4.46.2
true
131,072
null
0
128
null
null
null
hemanth955/Legal-Mistral
null
null
"2024-11-07T11:42:37Z"
null
null
25
null
null
null
null
0
transformers
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
2
silu
4,096
0.02
14,336
32,768
mistral
32
32
8
0.00001
1,000,000
null
false
float16
4.45.1
true
32,768
null
0
128
null
null
null
MaziyarPanahi/ReWiz-Nemo-12B-Instruct-GGUF
null
null
"2024-11-07T11:54:33Z"
null
null
113
null
null
null
null
0
null
[ "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:theprint/ReWiz-Nemo-12B-Instruct", "base_model:quantized:theprint/ReWiz-Nemo-12B-Instruct", "region:us", "imatrix", "conversational" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
twodigit/hades_9b3-95000
null
null
"2024-11-07T12:05:55Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Gemma2ForCausalLM" ]
2
1
gelu_pytorch_tanh
3,584
0.02
14,336
8,192
gemma2
16
42
8
0.000001
10,000
4,096
null
bfloat16
4.43.4
true
256,000
false
0
256
null
null
null
lsw0570168/krx-q25-7b-it-v8
null
null
"2024-11-07T12:06:12Z"
null
null
15
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "krx", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
krytonguard/gpt2-wikitext2
null
null
"2024-11-07T12:06:46Z"
null
null
17
null
null
null
null
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "GPT2LMHeadModel" ]
50,256
50,256
null
null
0.02
null
null
gpt2
null
null
null
null
null
null
null
float32
4.44.2
true
50,257
null
null
null
null
null
null
C10X/LongWriter-Qwen2.5-7B-Instruct
null
null
"2024-11-07T12:35:10Z"
null
null
3
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "Long Context", "qwen2.5", "conversational", "en", "zh", "arxiv:2408.07055", "arxiv:2410.10210", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.44.2
false
152,064
null
0
null
null
null
null
AIFunOver/SmolLM2-1.7B-Instruct-openvino-8bit
null
null
"2024-11-07T12:41:33Z"
null
null
8
null
null
null
null
0
transformers
[ "transformers", "openvino", "llama", "text-generation", "nncf", "8-bit", "conversational", "en", "base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct", "base_model:quantized:HuggingFaceTB/SmolLM2-1.7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
8,192
8,192
llama
32
24
32
0.00001
130,000
null
true
bfloat16
4.45.2
true
49,152
false
0
64
false
1
null
AIFunOver/SmolLM2-1.7B-Instruct-openvino-fp16
null
null
"2024-11-07T12:47:43Z"
null
null
14
null
null
null
null
0
transformers
[ "transformers", "openvino", "llama", "text-generation", "nncf", "fp16", "conversational", "en", "base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM2-1.7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
8,192
8,192
llama
32
24
32
0.00001
130,000
null
true
bfloat16
4.45.2
true
49,152
false
0
64
false
1
null
Sebuzdugan/llama-2-7b-miniplatypus
null
null
"2024-11-07T12:50:28Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
float16
4.46.2
true
32,000
false
0
128
false
1
null
jhllm/elecmachine_model_1107
null
null
"2024-11-07T12:53:40Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
128,000
128,001
silu
4,096
0.02
14,336
8,192
llama
32
32
8
0.00001
500,000
null
false
bfloat16
4.41.0
false
128,256
false
0
null
false
1
null
Nisk36/FT_tokyotech-llm_Swallow-7b-hf
null
null
"2024-11-07T12:56:17Z"
null
null
18
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
4,096
0.02
11,008
4,096
llama
32
32
32
0.00001
10,000
null
false
bfloat16
4.45.2
true
43,176
false
0
128
false
1
null
BigShort/krx_qwen2.5_7B_Instruct_it_v6.0
null
null
"2024-11-07T12:56:36Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,645
silu
3,584
0.02
18,944
32,768
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.2
true
152,064
null
0
null
null
null
null
RTX3090x2/qwen2.5-7B-Instruct-v0.2.2
null
null
"2024-11-07T13:02:57Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "krx", "en", "base_model:unsloth/Qwen2.5-7B", "base_model:finetune:unsloth/Qwen2.5-7B", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "Qwen2ForCausalLM" ]
151,643
151,643
silu
3,584
0.02
18,944
131,072
qwen2
28
28
4
0.000001
1,000,000
null
false
bfloat16
4.46.1
true
152,064
null
0
null
null
null
null
MaziyarPanahi/Starcannon-Unleashed-12B-v1.0-GGUF
null
null
"2024-11-07T13:04:27Z"
null
null
124
null
null
null
null
0
null
[ "gguf", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:VongolaChouko/Starcannon-Unleashed-12B-v1.0", "base_model:quantized:VongolaChouko/Starcannon-Unleashed-12B-v1.0", "region:us", "imatrix" ]
text-generation
null
null
0
null
null
null
null
null
null
null
null
mistral
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
null
AIFunOver/SmolLM2-1.7B-Instruct-openvino-4bit
null
null
"2024-11-07T13:11:21Z"
null
null
12
null
null
null
null
0
transformers
[ "transformers", "openvino", "llama", "text-generation", "nncf", "4-bit", "conversational", "en", "base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct", "base_model:quantized:HuggingFaceTB/SmolLM2-1.7B-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
8,192
8,192
llama
32
24
32
0.00001
130,000
null
true
bfloat16
4.45.2
true
49,152
false
0
64
false
1
null
natriistorm/Phi-3-BABI-SFT
null
null
"2024-11-07T13:12:55Z"
null
null
6
null
null
null
null
0
transformers
[ "transformers", "pytorch", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "MistralForCausalLM" ]
1
32,000
silu
3,072
0.02
8,192
4,096
mistral
32
32
32
0.00001
10,000
2,048
false
float16
4.46.2
true
32,064
null
0
96
null
null
null
NESPED-GEN/TinyLlama-1.1B-GerarSQL-v2-identacao-1071-steps
null
null
"2024-11-07T13:13:01Z"
null
null
4
null
null
null
null
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
2,048
0.02
5,632
2,048
llama
32
22
4
0.00001
10,000
null
false
float16
4.44.2
true
32,000
false
0
null
false
1
null
AIFunOver/SmolLM2-360M-Instruct-openvino-8bit
null
null
"2024-11-07T13:16:15Z"
null
null
10
null
null
null
null
0
transformers
[ "transformers", "openvino", "llama", "text-generation", "nncf", "8-bit", "conversational", "en", "base_model:HuggingFaceTB/SmolLM2-360M-Instruct", "base_model:quantized:HuggingFaceTB/SmolLM2-360M-Instruct", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
null
null
0
[ "LlamaForCausalLM" ]
1
2
silu
960
0.02
2,560
8,192
llama
15
32
5
0.00001
100,000
null
true
bfloat16
4.45.2
true
49,152
false
0
64
false
1
null