Text Generation
GGUF
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
role play
128k context
llama3.2
Inference Endpoints
imatrix
conversational
Update README.md
Browse files
README.md
CHANGED
@@ -40,6 +40,8 @@ pipeline_tag: text-generation
|
|
40 |
|
41 |
<h2>Llama-3.2-1B-Instruct-NEO-SI-FI-GGUF</h2>
|
42 |
|
|
|
|
|
43 |
It is the new "Llama-3.2-1B-Instruct", max context of 131,000 (128k) with the NEO IMATRIX Science Fiction and Story dataset.
|
44 |
|
45 |
The power in this 1B (for its size) is frankly jaw dropping... and at 190 tokens per second + on a GPU.
|
|
|
40 |
|
41 |
<h2>Llama-3.2-1B-Instruct-NEO-SI-FI-GGUF</h2>
|
42 |
|
43 |
+
<img src="llama-si-fi.jpg" style="float:right; width:300px; height:300px; padding:10px;">
|
44 |
+
|
45 |
It is the new "Llama-3.2-1B-Instruct", max context of 131,000 (128k) with the NEO IMATRIX Science Fiction and Story dataset.
|
46 |
|
47 |
The power in this 1B (for its size) is frankly jaw dropping... and at 190 tokens per second + on a GPU.
|