Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -133,7 +133,7 @@ class PromptRefiner:
|
|
133 |
messages = [
|
134 |
{
|
135 |
"role": "system",
|
136 |
-
"content": "You are a helpful assistant. Answer in stylized version with latex format or markdown if relevant.
|
137 |
},
|
138 |
#"content": "You are a helpful assistant. Answer in stylized version with latex format or markdown if relevant. Separate your answer into logical sections using level 2 headers (##) for sections and bolding (**) for subsections. Incorporate a variety of lists, headers, and text to make the answer visually appealing"
|
139 |
{
|
@@ -141,7 +141,7 @@ class PromptRefiner:
|
|
141 |
"content": prompt
|
142 |
}
|
143 |
]
|
144 |
-
messages = [
|
145 |
{
|
146 |
"role": "system",
|
147 |
"content": """You are a professional markdown formatting expert. Transform any content into well-structured documentation following these precise rules:
|
@@ -192,7 +192,7 @@ class PromptRefiner:
|
|
192 |
"role": "user",
|
193 |
"content": prompt
|
194 |
}
|
195 |
-
]
|
196 |
response = self.client.chat_completion(
|
197 |
model=model,
|
198 |
messages=messages,
|
|
|
133 |
messages = [
|
134 |
{
|
135 |
"role": "system",
|
136 |
+
"content": "You are a helpful assistant. Answer in stylized version with latex format or markdown if relevant. Separated paragraph and text to make the answer visually appealing"
|
137 |
},
|
138 |
#"content": "You are a helpful assistant. Answer in stylized version with latex format or markdown if relevant. Separate your answer into logical sections using level 2 headers (##) for sections and bolding (**) for subsections. Incorporate a variety of lists, headers, and text to make the answer visually appealing"
|
139 |
{
|
|
|
141 |
"content": prompt
|
142 |
}
|
143 |
]
|
144 |
+
''' messages = [
|
145 |
{
|
146 |
"role": "system",
|
147 |
"content": """You are a professional markdown formatting expert. Transform any content into well-structured documentation following these precise rules:
|
|
|
192 |
"role": "user",
|
193 |
"content": prompt
|
194 |
}
|
195 |
+
]'''
|
196 |
response = self.client.chat_completion(
|
197 |
model=model,
|
198 |
messages=messages,
|