Update README.md
Browse files
README.md
CHANGED
@@ -53,7 +53,7 @@ This version underwent a subset of DPO, but is fairly censored. For a less cens
|
|
53 |
3) Once inside the VM, open the terminal and run `conda activate text-generation-inference`
|
54 |
4) Then `cd Desktop/text-generation-inference/`
|
55 |
5) Run `volume=$PWD/data`
|
56 |
-
6) Run`model=jondurbin/
|
57 |
7) `sudo docker run --gpus '"device=0,1"' --shm-size 1g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.3 --model-id $model`
|
58 |
8) The model will take some time to load...
|
59 |
9) Once loaded the model will be available on port 8080
|
@@ -62,7 +62,7 @@ Sample command within the VM
|
|
62 |
```
|
63 |
curl 0.0.0.0:8080/generate \
|
64 |
-X POST \
|
65 |
-
-d '{"inputs":"
|
66 |
-H 'Content-Type: application/json'
|
67 |
```
|
68 |
|
@@ -70,7 +70,7 @@ You can also access the model from outside the VM
|
|
70 |
```
|
71 |
curl IP_ADDRESS_PROVIDED_BY_MASSED_COMPUTE_VM:8080/generate \
|
72 |
-X POST \
|
73 |
-
-d '{"inputs":"
|
74 |
-H 'Content-Type: application/json
|
75 |
```
|
76 |
|
|
|
53 |
3) Once inside the VM, open the terminal and run `conda activate text-generation-inference`
|
54 |
4) Then `cd Desktop/text-generation-inference/`
|
55 |
5) Run `volume=$PWD/data`
|
56 |
+
6) Run`model=jondurbin/nontoxicbagel-34b-v0.2`
|
57 |
7) `sudo docker run --gpus '"device=0,1"' --shm-size 1g -p 8080:80 -v $volume:/data ghcr.io/huggingface/text-generation-inference:1.3 --model-id $model`
|
58 |
8) The model will take some time to load...
|
59 |
9) Once loaded the model will be available on port 8080
|
|
|
62 |
```
|
63 |
curl 0.0.0.0:8080/generate \
|
64 |
-X POST \
|
65 |
+
-d '{"inputs":"[INST] <</SYS>>\nYou are a friendly chatbot.\n<</SYS>>\n\nWhat type of model are you? [/INST]","parameters":{"do_sample": true, "max_new_tokens": 100, "repetition_penalty": 1.15, "temperature": 0.7, "top_k": 20, "top_p": 0.9, "best_of": 1}}'\
|
66 |
-H 'Content-Type: application/json'
|
67 |
```
|
68 |
|
|
|
70 |
```
|
71 |
curl IP_ADDRESS_PROVIDED_BY_MASSED_COMPUTE_VM:8080/generate \
|
72 |
-X POST \
|
73 |
+
-d '{"inputs":"[INST] <</SYS>>\nYou are a friendly chatbot.\n<</SYS>>\n\nWhat type of model are you? [/INST]","parameters":{"do_sample": true, "max_new_tokens": 100, "repetition_penalty": 1.15, "temperature": 0.7, "top_k": 20, "top_p": 0.9, "best_of": 1}}'\
|
74 |
-H 'Content-Type: application/json
|
75 |
```
|
76 |
|