Spaces:
Runtime error
Runtime error
NewBreaker
commited on
Commit
•
00b02de
1
Parent(s):
3126b53
add app_local
Browse files- .idea/.gitignore +8 -0
- .idea/chatglm-6b-int4.iml +12 -0
- .idea/inspectionProfiles/Project_Default.xml +5 -0
- .idea/inspectionProfiles/profiles_settings.xml +6 -0
- .idea/misc.xml +7 -0
- .idea/modules.xml +8 -0
- .idea/vcs.xml +6 -0
- app.py +4 -4
- app_local.py +35 -0
.idea/.gitignore
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Default ignored files
|
2 |
+
/shelf/
|
3 |
+
/workspace.xml
|
4 |
+
# Editor-based HTTP Client requests
|
5 |
+
/httpRequests/
|
6 |
+
# Datasource local storage ignored files
|
7 |
+
/dataSources/
|
8 |
+
/dataSources.local.xml
|
.idea/chatglm-6b-int4.iml
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<module type="PYTHON_MODULE" version="4">
|
3 |
+
<component name="NewModuleRootManager">
|
4 |
+
<content url="file://$MODULE_DIR$" />
|
5 |
+
<orderEntry type="inheritedJdk" />
|
6 |
+
<orderEntry type="sourceFolder" forTests="false" />
|
7 |
+
</component>
|
8 |
+
<component name="PyDocumentationSettings">
|
9 |
+
<option name="format" value="PLAIN" />
|
10 |
+
<option name="myDocStringFormat" value="Plain" />
|
11 |
+
</component>
|
12 |
+
</module>
|
.idea/inspectionProfiles/Project_Default.xml
ADDED
@@ -0,0 +1,5 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<component name="InspectionProjectProfileManager">
|
2 |
+
<profile version="1.0">
|
3 |
+
<option name="myName" value="Project Default" />
|
4 |
+
</profile>
|
5 |
+
</component>
|
.idea/inspectionProfiles/profiles_settings.xml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<component name="InspectionProjectProfileManager">
|
2 |
+
<settings>
|
3 |
+
<option name="USE_PROJECT_PROFILE" value="false" />
|
4 |
+
<version value="1.0" />
|
5 |
+
</settings>
|
6 |
+
</component>
|
.idea/misc.xml
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="MarkdownSettingsMigration">
|
4 |
+
<option name="stateVersion" value="1" />
|
5 |
+
</component>
|
6 |
+
<component name="ProjectRootManager" version="2" project-jdk-name="Python 3.8" project-jdk-type="Python SDK" />
|
7 |
+
</project>
|
.idea/modules.xml
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="ProjectModuleManager">
|
4 |
+
<modules>
|
5 |
+
<module fileurl="file://$PROJECT_DIR$/.idea/chatglm-6b-int4.iml" filepath="$PROJECT_DIR$/.idea/chatglm-6b-int4.iml" />
|
6 |
+
</modules>
|
7 |
+
</component>
|
8 |
+
</project>
|
.idea/vcs.xml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="VcsDirectoryMappings">
|
4 |
+
<mapping directory="" vcs="Git" />
|
5 |
+
</component>
|
6 |
+
</project>
|
app.py
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
from transformers import AutoTokenizer, AutoModel
|
2 |
import gradio as gr
|
3 |
|
4 |
-
|
5 |
-
|
6 |
|
7 |
|
8 |
# from transformers import AutoTokenizer, AutoModel
|
@@ -12,9 +12,9 @@ import gradio as gr
|
|
12 |
|
13 |
|
14 |
# kernel_file = "./models/chatglm-6b-int4/quantization_kernels.so"
|
15 |
-
tokenizer = AutoTokenizer.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="")
|
16 |
# model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").half().cuda()
|
17 |
-
model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").float()
|
18 |
|
19 |
|
20 |
|
|
|
1 |
from transformers import AutoTokenizer, AutoModel
|
2 |
import gradio as gr
|
3 |
|
4 |
+
tokenizer = AutoTokenizer.from_pretrained(".\\models\\chatglm-6b-int4", trust_remote_code=True, revision="")
|
5 |
+
model = AutoModel.from_pretrained(".\\models\\chatglm-6b-int4", trust_remote_code=True, revision="").half().cuda()
|
6 |
|
7 |
|
8 |
# from transformers import AutoTokenizer, AutoModel
|
|
|
12 |
|
13 |
|
14 |
# kernel_file = "./models/chatglm-6b-int4/quantization_kernels.so"
|
15 |
+
# tokenizer = AutoTokenizer.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="")
|
16 |
# model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").half().cuda()
|
17 |
+
# model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").float()
|
18 |
|
19 |
|
20 |
|
app_local.py
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from transformers import AutoTokenizer, AutoModel
|
2 |
+
import gradio as gr
|
3 |
+
|
4 |
+
tokenizer = AutoTokenizer.from_pretrained(".\\models\\chatglm-6b-int4", trust_remote_code=True, revision="")
|
5 |
+
model = AutoModel.from_pretrained(".\\models\\chatglm-6b-int4", trust_remote_code=True, revision="").half().cuda()
|
6 |
+
|
7 |
+
|
8 |
+
# from transformers import AutoTokenizer, AutoModel
|
9 |
+
# tokenizer = AutoTokenizer.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True)
|
10 |
+
# model = AutoModel.from_pretrained("THUDM/chatglm-6b", trust_remote_code=True)
|
11 |
+
# model = model.eval()
|
12 |
+
|
13 |
+
|
14 |
+
# kernel_file = "./models/chatglm-6b-int4/quantization_kernels.so"
|
15 |
+
# tokenizer = AutoTokenizer.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="")
|
16 |
+
# model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").half().cuda()
|
17 |
+
# model = AutoModel.from_pretrained("./models/chatglm-6b-int4", trust_remote_code=True, revision="").float()
|
18 |
+
|
19 |
+
|
20 |
+
|
21 |
+
# model = model.quantize(bits=model_args.quantization_bit, kernel_file=kernel_file)
|
22 |
+
|
23 |
+
model = model.eval()
|
24 |
+
|
25 |
+
|
26 |
+
|
27 |
+
def chat(msg):
|
28 |
+
history = []
|
29 |
+
response, history = model.chat(tokenizer, msg, history=history)
|
30 |
+
print("response:", response)
|
31 |
+
return response
|
32 |
+
|
33 |
+
|
34 |
+
iface = gr.Interface(fn=chat, inputs="text", outputs="text")
|
35 |
+
iface.launch()
|