Spaces:
Runtime error
Runtime error
Blair Yang
commited on
Commit
•
487b80b
1
Parent(s):
eabbcfc
update
Browse files- Config.py +3 -5
- __pycache__/Config.cpython-311.pyc +0 -0
- __pycache__/models.cpython-311.pyc +0 -0
- models.py +9 -4
Config.py
CHANGED
@@ -26,11 +26,9 @@ MODELS = [
|
|
26 |
|
27 |
RANDOM_SEED = 42
|
28 |
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
# DEFAULT_SUMMARIZER = "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO"
|
33 |
-
DEFAULT_SUMMARIZER = 'mistralai/Mistral-7B-Instruct-v0.2'
|
34 |
DEFAULT_DATASET = "mmlu"
|
35 |
DEFAULT_TOPIC = random.choice(TOPICS[DEFAULT_DATASET])
|
36 |
|
|
|
26 |
|
27 |
RANDOM_SEED = 42
|
28 |
|
29 |
+
DEFAULT_SUMMARIZER = "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO"
|
30 |
+
# DEFAULT_SUMMARIZER = 'NousResearch/Nous-Hermes-2-Mistral-7B-DPO'
|
31 |
+
# DEFAULT_SUMMARIZER = 'mistralai/Mistral-7B-Instruct-v0.2'
|
|
|
|
|
32 |
DEFAULT_DATASET = "mmlu"
|
33 |
DEFAULT_TOPIC = random.choice(TOPICS[DEFAULT_DATASET])
|
34 |
|
__pycache__/Config.cpython-311.pyc
CHANGED
Binary files a/__pycache__/Config.cpython-311.pyc and b/__pycache__/Config.cpython-311.pyc differ
|
|
__pycache__/models.cpython-311.pyc
CHANGED
Binary files a/__pycache__/models.cpython-311.pyc and b/__pycache__/models.cpython-311.pyc differ
|
|
models.py
CHANGED
@@ -3,6 +3,7 @@ from __future__ import annotations
|
|
3 |
import json
|
4 |
import random
|
5 |
import re
|
|
|
6 |
from abc import ABC, abstractmethod
|
7 |
from typing import List, Dict, Union, Optional
|
8 |
|
@@ -10,12 +11,15 @@ from huggingface_hub import InferenceClient
|
|
10 |
from tenacity import retry, stop_after_attempt, wait_random_exponential
|
11 |
from transformers import AutoTokenizer
|
12 |
|
|
|
|
|
13 |
ROLE_SYSTEM = 'system'
|
14 |
ROLE_USER = 'user'
|
15 |
ROLE_ASSISTANT = 'assistant'
|
16 |
|
17 |
SUPPORTED_MISTRAL_MODELS = ['mistralai/Mixtral-8x7B-Instruct-v0.1', 'mistralai/Mistral-7B-Instruct-v0.2']
|
18 |
-
SUPPORTED_NOUS_MODELS = ['NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO'
|
|
|
19 |
SUPPORTED_LLAMA_MODELS = ['meta-llama/Llama-2-70b-chat-hf',
|
20 |
'meta-llama/Llama-2-13b-chat-hf',
|
21 |
'meta-llama/Llama-2-7b-chat-hf']
|
@@ -93,7 +97,8 @@ class HFAPIModel(Model):
|
|
93 |
|
94 |
@retry(stop=stop_after_attempt(5), wait=wait_random_exponential(max=10), reraise=True) # retry if exception
|
95 |
def get_response(self, temperature: float, use_json: bool, timeout: float, cache: bool) -> str:
|
96 |
-
|
|
|
97 |
# client = InferenceClient(model=self.name, token=random.choice(HF_API_TOKENS), timeout=timeout)
|
98 |
if not cache:
|
99 |
client.headers["x-use-cache"] = "0"
|
@@ -156,7 +161,7 @@ class MistralModel(HFAPIModel):
|
|
156 |
class NousHermesModel(HFAPIModel):
|
157 |
|
158 |
def __init__(self, system_prompt: str, model_name: str = 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO') -> None:
|
159 |
-
assert model_name in
|
160 |
super().__init__(model_name, system_prompt)
|
161 |
|
162 |
def format_messages(self) -> str:
|
@@ -200,4 +205,4 @@ class LlamaModel(HFAPIModel):
|
|
200 |
r += f'{content}</s>'
|
201 |
else:
|
202 |
raise ValueError
|
203 |
-
return r
|
|
|
3 |
import json
|
4 |
import random
|
5 |
import re
|
6 |
+
import os
|
7 |
from abc import ABC, abstractmethod
|
8 |
from typing import List, Dict, Union, Optional
|
9 |
|
|
|
11 |
from tenacity import retry, stop_after_attempt, wait_random_exponential
|
12 |
from transformers import AutoTokenizer
|
13 |
|
14 |
+
# from config import *
|
15 |
+
|
16 |
ROLE_SYSTEM = 'system'
|
17 |
ROLE_USER = 'user'
|
18 |
ROLE_ASSISTANT = 'assistant'
|
19 |
|
20 |
SUPPORTED_MISTRAL_MODELS = ['mistralai/Mixtral-8x7B-Instruct-v0.1', 'mistralai/Mistral-7B-Instruct-v0.2']
|
21 |
+
SUPPORTED_NOUS_MODELS = ['NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO',
|
22 |
+
'NousResearch/Nous-Hermes-2-Mistral-7B-DPO']
|
23 |
SUPPORTED_LLAMA_MODELS = ['meta-llama/Llama-2-70b-chat-hf',
|
24 |
'meta-llama/Llama-2-13b-chat-hf',
|
25 |
'meta-llama/Llama-2-7b-chat-hf']
|
|
|
97 |
|
98 |
@retry(stop=stop_after_attempt(5), wait=wait_random_exponential(max=10), reraise=True) # retry if exception
|
99 |
def get_response(self, temperature: float, use_json: bool, timeout: float, cache: bool) -> str:
|
100 |
+
# hf_api_token =
|
101 |
+
client = InferenceClient(model=self.name, token=os.getenv('HF_API_TOKEN'), timeout=timeout)
|
102 |
# client = InferenceClient(model=self.name, token=random.choice(HF_API_TOKENS), timeout=timeout)
|
103 |
if not cache:
|
104 |
client.headers["x-use-cache"] = "0"
|
|
|
161 |
class NousHermesModel(HFAPIModel):
|
162 |
|
163 |
def __init__(self, system_prompt: str, model_name: str = 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO') -> None:
|
164 |
+
assert model_name in SUPPORTED_NOUS_MODELS, 'Model not supported'
|
165 |
super().__init__(model_name, system_prompt)
|
166 |
|
167 |
def format_messages(self) -> str:
|
|
|
205 |
r += f'{content}</s>'
|
206 |
else:
|
207 |
raise ValueError
|
208 |
+
return r
|