Update README.md
Browse files
README.md
CHANGED
@@ -2,26 +2,22 @@
|
|
2 |
license: cc-by-sa-4.0
|
3 |
---
|
4 |
|
5 |
-
# SLIM-
|
6 |
|
7 |
<!-- Provide a quick summary of what the model is/does. -->
|
8 |
|
9 |
|
10 |
-
**slim-
|
11 |
|
12 |
-
This model
|
13 |
|
14 |
-
The intent of SLIMs is to forge a middle-ground between traditional encoder-based classifiers and open-ended API-based LLMs.
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
[**slim-sa-ner-3b**](https://huggingface.co/llmware/slim-sa-ner-3b) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
|
20 |
|
21 |
To pull the model via API:
|
22 |
|
23 |
from huggingface_hub import snapshot_download
|
24 |
-
snapshot_download("llmware/slim-
|
25 |
|
26 |
|
27 |
Load in your favorite GGUF inference engine, or try with llmware as follows:
|
@@ -29,14 +25,14 @@ Load in your favorite GGUF inference engine, or try with llmware as follows:
|
|
29 |
from llmware.models import ModelCatalog
|
30 |
|
31 |
# to load the model and make a basic inference
|
32 |
-
model = ModelCatalog().load_model("slim-
|
33 |
response = model.function_call(text_sample)
|
34 |
|
35 |
# this one line will download the model and run a series of tests
|
36 |
-
ModelCatalog().tool_test_run("slim-
|
37 |
|
38 |
|
39 |
-
Note: please review [**config.json**](https://huggingface.co/llmware/slim-
|
40 |
|
41 |
|
42 |
## Model Card Contact
|
|
|
2 |
license: cc-by-sa-4.0
|
3 |
---
|
4 |
|
5 |
+
# SLIM-BOOLEAN-TOOL
|
6 |
|
7 |
<!-- Provide a quick summary of what the model is/does. -->
|
8 |
|
9 |
|
10 |
+
**slim-boolean-tool** is a 4_K_M quantized GGUF version of slim-boolean, providing a small, fast inference implementation, optimized for multi-model concurrent deployment.
|
11 |
|
12 |
+
This is an experimental model that takes as input a context passage, a yes-no question, and an optional (explain) parameter, and generates a response consisting of a python dictionary with two keys- 'answer' consisting of the 'yes/no' classification, and 'explanation' which provides a text explanation, derived from the source passage that explains the boolean classification assesment. All of the details on the prompt template as provided in the config.json file in this model repo, along with several examples.
|
13 |
|
|
|
14 |
|
15 |
+
[**slim-boolean**](https://huggingface.co/llmware/slim-boolean) is part of the SLIM ("**S**tructured **L**anguage **I**nstruction **M**odel") series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.
|
|
|
|
|
|
|
16 |
|
17 |
To pull the model via API:
|
18 |
|
19 |
from huggingface_hub import snapshot_download
|
20 |
+
snapshot_download("llmware/slim-boolean-tool", local_dir="/path/on/your/machine/", local_dir_use_symlinks=False)
|
21 |
|
22 |
|
23 |
Load in your favorite GGUF inference engine, or try with llmware as follows:
|
|
|
25 |
from llmware.models import ModelCatalog
|
26 |
|
27 |
# to load the model and make a basic inference
|
28 |
+
model = ModelCatalog().load_model("slim-boolean-tool")
|
29 |
response = model.function_call(text_sample)
|
30 |
|
31 |
# this one line will download the model and run a series of tests
|
32 |
+
ModelCatalog().tool_test_run("slim-boolean-tool", verbose=True)
|
33 |
|
34 |
|
35 |
+
Note: please review [**config.json**](https://huggingface.co/llmware/slim-boolean-tool/blob/main/config.json) in the repository for prompt wrapping information, details on the model, and full test set.
|
36 |
|
37 |
|
38 |
## Model Card Contact
|