Liverpool commited on
Commit
dc12c31
1 Parent(s): d3c1b63

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .github/FUNDING.yml +1 -0
  2. .github/ISSUE_TEMPLATE/bug_report_template.yml +53 -0
  3. .github/ISSUE_TEMPLATE/feature_request.md +16 -0
  4. .github/dependabot.yml +11 -0
  5. .github/pull_request_template.md +3 -0
  6. .github/workflows/stale.yml +22 -0
  7. .gitignore +35 -0
  8. LICENSE +661 -0
  9. README.md +396 -8
  10. api-examples/api-example-chat-stream.py +111 -0
  11. api-examples/api-example-chat.py +91 -0
  12. api-examples/api-example-model.py +176 -0
  13. api-examples/api-example-stream.py +85 -0
  14. api-examples/api-example.py +62 -0
  15. characters/Example.png +0 -0
  16. characters/Example.yaml +17 -0
  17. convert-to-safetensors.py +38 -0
  18. css/NotoSans/NotoSans-Black.woff +0 -0
  19. css/NotoSans/NotoSans-Black.woff2 +0 -0
  20. css/NotoSans/NotoSans-BlackItalic.woff +0 -0
  21. css/NotoSans/NotoSans-BlackItalic.woff2 +0 -0
  22. css/NotoSans/NotoSans-Bold.woff +0 -0
  23. css/NotoSans/NotoSans-Bold.woff2 +0 -0
  24. css/NotoSans/NotoSans-BoldItalic.woff +0 -0
  25. css/NotoSans/NotoSans-BoldItalic.woff2 +0 -0
  26. css/NotoSans/NotoSans-ExtraBold.woff +0 -0
  27. css/NotoSans/NotoSans-ExtraBold.woff2 +0 -0
  28. css/NotoSans/NotoSans-ExtraBoldItalic.woff +0 -0
  29. css/NotoSans/NotoSans-ExtraBoldItalic.woff2 +0 -0
  30. css/NotoSans/NotoSans-ExtraLight.woff +0 -0
  31. css/NotoSans/NotoSans-ExtraLight.woff2 +0 -0
  32. css/NotoSans/NotoSans-ExtraLightItalic.woff +0 -0
  33. css/NotoSans/NotoSans-ExtraLightItalic.woff2 +0 -0
  34. css/NotoSans/NotoSans-Italic.woff +0 -0
  35. css/NotoSans/NotoSans-Italic.woff2 +0 -0
  36. css/NotoSans/NotoSans-Light.woff +0 -0
  37. css/NotoSans/NotoSans-Light.woff2 +0 -0
  38. css/NotoSans/NotoSans-LightItalic.woff +0 -0
  39. css/NotoSans/NotoSans-LightItalic.woff2 +0 -0
  40. css/NotoSans/NotoSans-Medium.woff +0 -0
  41. css/NotoSans/NotoSans-Medium.woff2 +0 -0
  42. css/NotoSans/NotoSans-MediumItalic.woff +0 -0
  43. css/NotoSans/NotoSans-MediumItalic.woff2 +0 -0
  44. css/NotoSans/NotoSans-Regular.woff +0 -0
  45. css/NotoSans/NotoSans-Regular.woff2 +0 -0
  46. css/NotoSans/NotoSans-SemiBold.woff +0 -0
  47. css/NotoSans/NotoSans-SemiBold.woff2 +0 -0
  48. css/NotoSans/NotoSans-SemiBoldItalic.woff +0 -0
  49. css/NotoSans/NotoSans-SemiBoldItalic.woff2 +0 -0
  50. css/NotoSans/NotoSans-Thin.woff +0 -0
.github/FUNDING.yml ADDED
@@ -0,0 +1 @@
 
 
1
+ ko_fi: oobabooga
.github/ISSUE_TEMPLATE/bug_report_template.yml ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: "Bug report"
2
+ description: Report a bug
3
+ labels: [ "bug" ]
4
+ body:
5
+ - type: markdown
6
+ attributes:
7
+ value: |
8
+ Thanks for taking the time to fill out this bug report!
9
+ - type: textarea
10
+ id: bug-description
11
+ attributes:
12
+ label: Describe the bug
13
+ description: A clear and concise description of what the bug is.
14
+ placeholder: Bug description
15
+ validations:
16
+ required: true
17
+ - type: checkboxes
18
+ attributes:
19
+ label: Is there an existing issue for this?
20
+ description: Please search to see if an issue already exists for the issue you encountered.
21
+ options:
22
+ - label: I have searched the existing issues
23
+ required: true
24
+ - type: textarea
25
+ id: reproduction
26
+ attributes:
27
+ label: Reproduction
28
+ description: Please provide the steps necessary to reproduce your issue.
29
+ placeholder: Reproduction
30
+ validations:
31
+ required: true
32
+ - type: textarea
33
+ id: screenshot
34
+ attributes:
35
+ label: Screenshot
36
+ description: "If possible, please include screenshot(s) so that we can understand what the issue is."
37
+ - type: textarea
38
+ id: logs
39
+ attributes:
40
+ label: Logs
41
+ description: "Please include the full stacktrace of the errors you get in the command-line (if any)."
42
+ render: shell
43
+ validations:
44
+ required: true
45
+ - type: textarea
46
+ id: system-info
47
+ attributes:
48
+ label: System Info
49
+ description: "Please share your system info with us: operating system, GPU brand, and GPU model. If you are using a Google Colab notebook, mention that instead."
50
+ render: shell
51
+ placeholder:
52
+ validations:
53
+ required: true
.github/ISSUE_TEMPLATE/feature_request.md ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ name: Feature request
3
+ about: Suggest an improvement or new feature for the web UI
4
+ title: ''
5
+ labels: 'enhancement'
6
+ assignees: ''
7
+
8
+ ---
9
+
10
+ **Description**
11
+
12
+ A clear and concise description of what you want to be implemented.
13
+
14
+ **Additional Context**
15
+
16
+ If applicable, please provide any extra information, external links, or screenshots that could be useful.
.github/dependabot.yml ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # To get started with Dependabot version updates, you'll need to specify which
2
+ # package ecosystems to update and where the package manifests are located.
3
+ # Please see the documentation for all configuration options:
4
+ # https://docs.github.com/github/administering-a-repository/configuration-options-for-dependency-updates
5
+
6
+ version: 2
7
+ updates:
8
+ - package-ecosystem: "pip" # See documentation for possible values
9
+ directory: "/" # Location of package manifests
10
+ schedule:
11
+ interval: "weekly"
.github/pull_request_template.md ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ ## Checklist:
2
+
3
+ - [ ] I have read the [Contributing guidelines](https://github.com/oobabooga/text-generation-webui/wiki/Contributing-guidelines).
.github/workflows/stale.yml ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: Close inactive issues
2
+ on:
3
+ schedule:
4
+ - cron: "10 23 * * *"
5
+
6
+ jobs:
7
+ close-issues:
8
+ runs-on: ubuntu-latest
9
+ permissions:
10
+ issues: write
11
+ pull-requests: write
12
+ steps:
13
+ - uses: actions/stale@v5
14
+ with:
15
+ stale-issue-message: ""
16
+ close-issue-message: "This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment."
17
+ days-before-issue-stale: 42
18
+ days-before-issue-close: 0
19
+ stale-issue-label: "stale"
20
+ days-before-pr-stale: -1
21
+ days-before-pr-close: -1
22
+ repo-token: ${{ secrets.GITHUB_TOKEN }}
.gitignore ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ cache
2
+ characters
3
+ training/datasets
4
+ extensions/silero_tts/outputs
5
+ extensions/elevenlabs_tts/outputs
6
+ extensions/sd_api_pictures/outputs
7
+ extensions/multimodal/pipelines
8
+ logs
9
+ loras
10
+ models
11
+ presets
12
+ repositories
13
+ softprompts
14
+ torch-dumps
15
+ *pycache*
16
+ */*pycache*
17
+ */*/pycache*
18
+ venv/
19
+ .venv/
20
+ .vscode
21
+ .idea/
22
+ *.bak
23
+ *.ipynb
24
+ *.log
25
+
26
+ settings.json
27
+ settings.yaml
28
+ notification.mp3
29
+ img_bot*
30
+ img_me*
31
+ prompts/[0-9]*
32
+ models/config-user.yaml
33
+
34
+ .DS_Store
35
+ Thumbs.db
LICENSE ADDED
@@ -0,0 +1,661 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ GNU AFFERO GENERAL PUBLIC LICENSE
2
+ Version 3, 19 November 2007
3
+
4
+ Copyright (C) 2007 Free Software Foundation, Inc. <https://fsf.org/>
5
+ Everyone is permitted to copy and distribute verbatim copies
6
+ of this license document, but changing it is not allowed.
7
+
8
+ Preamble
9
+
10
+ The GNU Affero General Public License is a free, copyleft license for
11
+ software and other kinds of works, specifically designed to ensure
12
+ cooperation with the community in the case of network server software.
13
+
14
+ The licenses for most software and other practical works are designed
15
+ to take away your freedom to share and change the works. By contrast,
16
+ our General Public Licenses are intended to guarantee your freedom to
17
+ share and change all versions of a program--to make sure it remains free
18
+ software for all its users.
19
+
20
+ When we speak of free software, we are referring to freedom, not
21
+ price. Our General Public Licenses are designed to make sure that you
22
+ have the freedom to distribute copies of free software (and charge for
23
+ them if you wish), that you receive source code or can get it if you
24
+ want it, that you can change the software or use pieces of it in new
25
+ free programs, and that you know you can do these things.
26
+
27
+ Developers that use our General Public Licenses protect your rights
28
+ with two steps: (1) assert copyright on the software, and (2) offer
29
+ you this License which gives you legal permission to copy, distribute
30
+ and/or modify the software.
31
+
32
+ A secondary benefit of defending all users' freedom is that
33
+ improvements made in alternate versions of the program, if they
34
+ receive widespread use, become available for other developers to
35
+ incorporate. Many developers of free software are heartened and
36
+ encouraged by the resulting cooperation. However, in the case of
37
+ software used on network servers, this result may fail to come about.
38
+ The GNU General Public License permits making a modified version and
39
+ letting the public access it on a server without ever releasing its
40
+ source code to the public.
41
+
42
+ The GNU Affero General Public License is designed specifically to
43
+ ensure that, in such cases, the modified source code becomes available
44
+ to the community. It requires the operator of a network server to
45
+ provide the source code of the modified version running there to the
46
+ users of that server. Therefore, public use of a modified version, on
47
+ a publicly accessible server, gives the public access to the source
48
+ code of the modified version.
49
+
50
+ An older license, called the Affero General Public License and
51
+ published by Affero, was designed to accomplish similar goals. This is
52
+ a different license, not a version of the Affero GPL, but Affero has
53
+ released a new version of the Affero GPL which permits relicensing under
54
+ this license.
55
+
56
+ The precise terms and conditions for copying, distribution and
57
+ modification follow.
58
+
59
+ TERMS AND CONDITIONS
60
+
61
+ 0. Definitions.
62
+
63
+ "This License" refers to version 3 of the GNU Affero General Public License.
64
+
65
+ "Copyright" also means copyright-like laws that apply to other kinds of
66
+ works, such as semiconductor masks.
67
+
68
+ "The Program" refers to any copyrightable work licensed under this
69
+ License. Each licensee is addressed as "you". "Licensees" and
70
+ "recipients" may be individuals or organizations.
71
+
72
+ To "modify" a work means to copy from or adapt all or part of the work
73
+ in a fashion requiring copyright permission, other than the making of an
74
+ exact copy. The resulting work is called a "modified version" of the
75
+ earlier work or a work "based on" the earlier work.
76
+
77
+ A "covered work" means either the unmodified Program or a work based
78
+ on the Program.
79
+
80
+ To "propagate" a work means to do anything with it that, without
81
+ permission, would make you directly or secondarily liable for
82
+ infringement under applicable copyright law, except executing it on a
83
+ computer or modifying a private copy. Propagation includes copying,
84
+ distribution (with or without modification), making available to the
85
+ public, and in some countries other activities as well.
86
+
87
+ To "convey" a work means any kind of propagation that enables other
88
+ parties to make or receive copies. Mere interaction with a user through
89
+ a computer network, with no transfer of a copy, is not conveying.
90
+
91
+ An interactive user interface displays "Appropriate Legal Notices"
92
+ to the extent that it includes a convenient and prominently visible
93
+ feature that (1) displays an appropriate copyright notice, and (2)
94
+ tells the user that there is no warranty for the work (except to the
95
+ extent that warranties are provided), that licensees may convey the
96
+ work under this License, and how to view a copy of this License. If
97
+ the interface presents a list of user commands or options, such as a
98
+ menu, a prominent item in the list meets this criterion.
99
+
100
+ 1. Source Code.
101
+
102
+ The "source code" for a work means the preferred form of the work
103
+ for making modifications to it. "Object code" means any non-source
104
+ form of a work.
105
+
106
+ A "Standard Interface" means an interface that either is an official
107
+ standard defined by a recognized standards body, or, in the case of
108
+ interfaces specified for a particular programming language, one that
109
+ is widely used among developers working in that language.
110
+
111
+ The "System Libraries" of an executable work include anything, other
112
+ than the work as a whole, that (a) is included in the normal form of
113
+ packaging a Major Component, but which is not part of that Major
114
+ Component, and (b) serves only to enable use of the work with that
115
+ Major Component, or to implement a Standard Interface for which an
116
+ implementation is available to the public in source code form. A
117
+ "Major Component", in this context, means a major essential component
118
+ (kernel, window system, and so on) of the specific operating system
119
+ (if any) on which the executable work runs, or a compiler used to
120
+ produce the work, or an object code interpreter used to run it.
121
+
122
+ The "Corresponding Source" for a work in object code form means all
123
+ the source code needed to generate, install, and (for an executable
124
+ work) run the object code and to modify the work, including scripts to
125
+ control those activities. However, it does not include the work's
126
+ System Libraries, or general-purpose tools or generally available free
127
+ programs which are used unmodified in performing those activities but
128
+ which are not part of the work. For example, Corresponding Source
129
+ includes interface definition files associated with source files for
130
+ the work, and the source code for shared libraries and dynamically
131
+ linked subprograms that the work is specifically designed to require,
132
+ such as by intimate data communication or control flow between those
133
+ subprograms and other parts of the work.
134
+
135
+ The Corresponding Source need not include anything that users
136
+ can regenerate automatically from other parts of the Corresponding
137
+ Source.
138
+
139
+ The Corresponding Source for a work in source code form is that
140
+ same work.
141
+
142
+ 2. Basic Permissions.
143
+
144
+ All rights granted under this License are granted for the term of
145
+ copyright on the Program, and are irrevocable provided the stated
146
+ conditions are met. This License explicitly affirms your unlimited
147
+ permission to run the unmodified Program. The output from running a
148
+ covered work is covered by this License only if the output, given its
149
+ content, constitutes a covered work. This License acknowledges your
150
+ rights of fair use or other equivalent, as provided by copyright law.
151
+
152
+ You may make, run and propagate covered works that you do not
153
+ convey, without conditions so long as your license otherwise remains
154
+ in force. You may convey covered works to others for the sole purpose
155
+ of having them make modifications exclusively for you, or provide you
156
+ with facilities for running those works, provided that you comply with
157
+ the terms of this License in conveying all material for which you do
158
+ not control copyright. Those thus making or running the covered works
159
+ for you must do so exclusively on your behalf, under your direction
160
+ and control, on terms that prohibit them from making any copies of
161
+ your copyrighted material outside their relationship with you.
162
+
163
+ Conveying under any other circumstances is permitted solely under
164
+ the conditions stated below. Sublicensing is not allowed; section 10
165
+ makes it unnecessary.
166
+
167
+ 3. Protecting Users' Legal Rights From Anti-Circumvention Law.
168
+
169
+ No covered work shall be deemed part of an effective technological
170
+ measure under any applicable law fulfilling obligations under article
171
+ 11 of the WIPO copyright treaty adopted on 20 December 1996, or
172
+ similar laws prohibiting or restricting circumvention of such
173
+ measures.
174
+
175
+ When you convey a covered work, you waive any legal power to forbid
176
+ circumvention of technological measures to the extent such circumvention
177
+ is effected by exercising rights under this License with respect to
178
+ the covered work, and you disclaim any intention to limit operation or
179
+ modification of the work as a means of enforcing, against the work's
180
+ users, your or third parties' legal rights to forbid circumvention of
181
+ technological measures.
182
+
183
+ 4. Conveying Verbatim Copies.
184
+
185
+ You may convey verbatim copies of the Program's source code as you
186
+ receive it, in any medium, provided that you conspicuously and
187
+ appropriately publish on each copy an appropriate copyright notice;
188
+ keep intact all notices stating that this License and any
189
+ non-permissive terms added in accord with section 7 apply to the code;
190
+ keep intact all notices of the absence of any warranty; and give all
191
+ recipients a copy of this License along with the Program.
192
+
193
+ You may charge any price or no price for each copy that you convey,
194
+ and you may offer support or warranty protection for a fee.
195
+
196
+ 5. Conveying Modified Source Versions.
197
+
198
+ You may convey a work based on the Program, or the modifications to
199
+ produce it from the Program, in the form of source code under the
200
+ terms of section 4, provided that you also meet all of these conditions:
201
+
202
+ a) The work must carry prominent notices stating that you modified
203
+ it, and giving a relevant date.
204
+
205
+ b) The work must carry prominent notices stating that it is
206
+ released under this License and any conditions added under section
207
+ 7. This requirement modifies the requirement in section 4 to
208
+ "keep intact all notices".
209
+
210
+ c) You must license the entire work, as a whole, under this
211
+ License to anyone who comes into possession of a copy. This
212
+ License will therefore apply, along with any applicable section 7
213
+ additional terms, to the whole of the work, and all its parts,
214
+ regardless of how they are packaged. This License gives no
215
+ permission to license the work in any other way, but it does not
216
+ invalidate such permission if you have separately received it.
217
+
218
+ d) If the work has interactive user interfaces, each must display
219
+ Appropriate Legal Notices; however, if the Program has interactive
220
+ interfaces that do not display Appropriate Legal Notices, your
221
+ work need not make them do so.
222
+
223
+ A compilation of a covered work with other separate and independent
224
+ works, which are not by their nature extensions of the covered work,
225
+ and which are not combined with it such as to form a larger program,
226
+ in or on a volume of a storage or distribution medium, is called an
227
+ "aggregate" if the compilation and its resulting copyright are not
228
+ used to limit the access or legal rights of the compilation's users
229
+ beyond what the individual works permit. Inclusion of a covered work
230
+ in an aggregate does not cause this License to apply to the other
231
+ parts of the aggregate.
232
+
233
+ 6. Conveying Non-Source Forms.
234
+
235
+ You may convey a covered work in object code form under the terms
236
+ of sections 4 and 5, provided that you also convey the
237
+ machine-readable Corresponding Source under the terms of this License,
238
+ in one of these ways:
239
+
240
+ a) Convey the object code in, or embodied in, a physical product
241
+ (including a physical distribution medium), accompanied by the
242
+ Corresponding Source fixed on a durable physical medium
243
+ customarily used for software interchange.
244
+
245
+ b) Convey the object code in, or embodied in, a physical product
246
+ (including a physical distribution medium), accompanied by a
247
+ written offer, valid for at least three years and valid for as
248
+ long as you offer spare parts or customer support for that product
249
+ model, to give anyone who possesses the object code either (1) a
250
+ copy of the Corresponding Source for all the software in the
251
+ product that is covered by this License, on a durable physical
252
+ medium customarily used for software interchange, for a price no
253
+ more than your reasonable cost of physically performing this
254
+ conveying of source, or (2) access to copy the
255
+ Corresponding Source from a network server at no charge.
256
+
257
+ c) Convey individual copies of the object code with a copy of the
258
+ written offer to provide the Corresponding Source. This
259
+ alternative is allowed only occasionally and noncommercially, and
260
+ only if you received the object code with such an offer, in accord
261
+ with subsection 6b.
262
+
263
+ d) Convey the object code by offering access from a designated
264
+ place (gratis or for a charge), and offer equivalent access to the
265
+ Corresponding Source in the same way through the same place at no
266
+ further charge. You need not require recipients to copy the
267
+ Corresponding Source along with the object code. If the place to
268
+ copy the object code is a network server, the Corresponding Source
269
+ may be on a different server (operated by you or a third party)
270
+ that supports equivalent copying facilities, provided you maintain
271
+ clear directions next to the object code saying where to find the
272
+ Corresponding Source. Regardless of what server hosts the
273
+ Corresponding Source, you remain obligated to ensure that it is
274
+ available for as long as needed to satisfy these requirements.
275
+
276
+ e) Convey the object code using peer-to-peer transmission, provided
277
+ you inform other peers where the object code and Corresponding
278
+ Source of the work are being offered to the general public at no
279
+ charge under subsection 6d.
280
+
281
+ A separable portion of the object code, whose source code is excluded
282
+ from the Corresponding Source as a System Library, need not be
283
+ included in conveying the object code work.
284
+
285
+ A "User Product" is either (1) a "consumer product", which means any
286
+ tangible personal property which is normally used for personal, family,
287
+ or household purposes, or (2) anything designed or sold for incorporation
288
+ into a dwelling. In determining whether a product is a consumer product,
289
+ doubtful cases shall be resolved in favor of coverage. For a particular
290
+ product received by a particular user, "normally used" refers to a
291
+ typical or common use of that class of product, regardless of the status
292
+ of the particular user or of the way in which the particular user
293
+ actually uses, or expects or is expected to use, the product. A product
294
+ is a consumer product regardless of whether the product has substantial
295
+ commercial, industrial or non-consumer uses, unless such uses represent
296
+ the only significant mode of use of the product.
297
+
298
+ "Installation Information" for a User Product means any methods,
299
+ procedures, authorization keys, or other information required to install
300
+ and execute modified versions of a covered work in that User Product from
301
+ a modified version of its Corresponding Source. The information must
302
+ suffice to ensure that the continued functioning of the modified object
303
+ code is in no case prevented or interfered with solely because
304
+ modification has been made.
305
+
306
+ If you convey an object code work under this section in, or with, or
307
+ specifically for use in, a User Product, and the conveying occurs as
308
+ part of a transaction in which the right of possession and use of the
309
+ User Product is transferred to the recipient in perpetuity or for a
310
+ fixed term (regardless of how the transaction is characterized), the
311
+ Corresponding Source conveyed under this section must be accompanied
312
+ by the Installation Information. But this requirement does not apply
313
+ if neither you nor any third party retains the ability to install
314
+ modified object code on the User Product (for example, the work has
315
+ been installed in ROM).
316
+
317
+ The requirement to provide Installation Information does not include a
318
+ requirement to continue to provide support service, warranty, or updates
319
+ for a work that has been modified or installed by the recipient, or for
320
+ the User Product in which it has been modified or installed. Access to a
321
+ network may be denied when the modification itself materially and
322
+ adversely affects the operation of the network or violates the rules and
323
+ protocols for communication across the network.
324
+
325
+ Corresponding Source conveyed, and Installation Information provided,
326
+ in accord with this section must be in a format that is publicly
327
+ documented (and with an implementation available to the public in
328
+ source code form), and must require no special password or key for
329
+ unpacking, reading or copying.
330
+
331
+ 7. Additional Terms.
332
+
333
+ "Additional permissions" are terms that supplement the terms of this
334
+ License by making exceptions from one or more of its conditions.
335
+ Additional permissions that are applicable to the entire Program shall
336
+ be treated as though they were included in this License, to the extent
337
+ that they are valid under applicable law. If additional permissions
338
+ apply only to part of the Program, that part may be used separately
339
+ under those permissions, but the entire Program remains governed by
340
+ this License without regard to the additional permissions.
341
+
342
+ When you convey a copy of a covered work, you may at your option
343
+ remove any additional permissions from that copy, or from any part of
344
+ it. (Additional permissions may be written to require their own
345
+ removal in certain cases when you modify the work.) You may place
346
+ additional permissions on material, added by you to a covered work,
347
+ for which you have or can give appropriate copyright permission.
348
+
349
+ Notwithstanding any other provision of this License, for material you
350
+ add to a covered work, you may (if authorized by the copyright holders of
351
+ that material) supplement the terms of this License with terms:
352
+
353
+ a) Disclaiming warranty or limiting liability differently from the
354
+ terms of sections 15 and 16 of this License; or
355
+
356
+ b) Requiring preservation of specified reasonable legal notices or
357
+ author attributions in that material or in the Appropriate Legal
358
+ Notices displayed by works containing it; or
359
+
360
+ c) Prohibiting misrepresentation of the origin of that material, or
361
+ requiring that modified versions of such material be marked in
362
+ reasonable ways as different from the original version; or
363
+
364
+ d) Limiting the use for publicity purposes of names of licensors or
365
+ authors of the material; or
366
+
367
+ e) Declining to grant rights under trademark law for use of some
368
+ trade names, trademarks, or service marks; or
369
+
370
+ f) Requiring indemnification of licensors and authors of that
371
+ material by anyone who conveys the material (or modified versions of
372
+ it) with contractual assumptions of liability to the recipient, for
373
+ any liability that these contractual assumptions directly impose on
374
+ those licensors and authors.
375
+
376
+ All other non-permissive additional terms are considered "further
377
+ restrictions" within the meaning of section 10. If the Program as you
378
+ received it, or any part of it, contains a notice stating that it is
379
+ governed by this License along with a term that is a further
380
+ restriction, you may remove that term. If a license document contains
381
+ a further restriction but permits relicensing or conveying under this
382
+ License, you may add to a covered work material governed by the terms
383
+ of that license document, provided that the further restriction does
384
+ not survive such relicensing or conveying.
385
+
386
+ If you add terms to a covered work in accord with this section, you
387
+ must place, in the relevant source files, a statement of the
388
+ additional terms that apply to those files, or a notice indicating
389
+ where to find the applicable terms.
390
+
391
+ Additional terms, permissive or non-permissive, may be stated in the
392
+ form of a separately written license, or stated as exceptions;
393
+ the above requirements apply either way.
394
+
395
+ 8. Termination.
396
+
397
+ You may not propagate or modify a covered work except as expressly
398
+ provided under this License. Any attempt otherwise to propagate or
399
+ modify it is void, and will automatically terminate your rights under
400
+ this License (including any patent licenses granted under the third
401
+ paragraph of section 11).
402
+
403
+ However, if you cease all violation of this License, then your
404
+ license from a particular copyright holder is reinstated (a)
405
+ provisionally, unless and until the copyright holder explicitly and
406
+ finally terminates your license, and (b) permanently, if the copyright
407
+ holder fails to notify you of the violation by some reasonable means
408
+ prior to 60 days after the cessation.
409
+
410
+ Moreover, your license from a particular copyright holder is
411
+ reinstated permanently if the copyright holder notifies you of the
412
+ violation by some reasonable means, this is the first time you have
413
+ received notice of violation of this License (for any work) from that
414
+ copyright holder, and you cure the violation prior to 30 days after
415
+ your receipt of the notice.
416
+
417
+ Termination of your rights under this section does not terminate the
418
+ licenses of parties who have received copies or rights from you under
419
+ this License. If your rights have been terminated and not permanently
420
+ reinstated, you do not qualify to receive new licenses for the same
421
+ material under section 10.
422
+
423
+ 9. Acceptance Not Required for Having Copies.
424
+
425
+ You are not required to accept this License in order to receive or
426
+ run a copy of the Program. Ancillary propagation of a covered work
427
+ occurring solely as a consequence of using peer-to-peer transmission
428
+ to receive a copy likewise does not require acceptance. However,
429
+ nothing other than this License grants you permission to propagate or
430
+ modify any covered work. These actions infringe copyright if you do
431
+ not accept this License. Therefore, by modifying or propagating a
432
+ covered work, you indicate your acceptance of this License to do so.
433
+
434
+ 10. Automatic Licensing of Downstream Recipients.
435
+
436
+ Each time you convey a covered work, the recipient automatically
437
+ receives a license from the original licensors, to run, modify and
438
+ propagate that work, subject to this License. You are not responsible
439
+ for enforcing compliance by third parties with this License.
440
+
441
+ An "entity transaction" is a transaction transferring control of an
442
+ organization, or substantially all assets of one, or subdividing an
443
+ organization, or merging organizations. If propagation of a covered
444
+ work results from an entity transaction, each party to that
445
+ transaction who receives a copy of the work also receives whatever
446
+ licenses to the work the party's predecessor in interest had or could
447
+ give under the previous paragraph, plus a right to possession of the
448
+ Corresponding Source of the work from the predecessor in interest, if
449
+ the predecessor has it or can get it with reasonable efforts.
450
+
451
+ You may not impose any further restrictions on the exercise of the
452
+ rights granted or affirmed under this License. For example, you may
453
+ not impose a license fee, royalty, or other charge for exercise of
454
+ rights granted under this License, and you may not initiate litigation
455
+ (including a cross-claim or counterclaim in a lawsuit) alleging that
456
+ any patent claim is infringed by making, using, selling, offering for
457
+ sale, or importing the Program or any portion of it.
458
+
459
+ 11. Patents.
460
+
461
+ A "contributor" is a copyright holder who authorizes use under this
462
+ License of the Program or a work on which the Program is based. The
463
+ work thus licensed is called the contributor's "contributor version".
464
+
465
+ A contributor's "essential patent claims" are all patent claims
466
+ owned or controlled by the contributor, whether already acquired or
467
+ hereafter acquired, that would be infringed by some manner, permitted
468
+ by this License, of making, using, or selling its contributor version,
469
+ but do not include claims that would be infringed only as a
470
+ consequence of further modification of the contributor version. For
471
+ purposes of this definition, "control" includes the right to grant
472
+ patent sublicenses in a manner consistent with the requirements of
473
+ this License.
474
+
475
+ Each contributor grants you a non-exclusive, worldwide, royalty-free
476
+ patent license under the contributor's essential patent claims, to
477
+ make, use, sell, offer for sale, import and otherwise run, modify and
478
+ propagate the contents of its contributor version.
479
+
480
+ In the following three paragraphs, a "patent license" is any express
481
+ agreement or commitment, however denominated, not to enforce a patent
482
+ (such as an express permission to practice a patent or covenant not to
483
+ sue for patent infringement). To "grant" such a patent license to a
484
+ party means to make such an agreement or commitment not to enforce a
485
+ patent against the party.
486
+
487
+ If you convey a covered work, knowingly relying on a patent license,
488
+ and the Corresponding Source of the work is not available for anyone
489
+ to copy, free of charge and under the terms of this License, through a
490
+ publicly available network server or other readily accessible means,
491
+ then you must either (1) cause the Corresponding Source to be so
492
+ available, or (2) arrange to deprive yourself of the benefit of the
493
+ patent license for this particular work, or (3) arrange, in a manner
494
+ consistent with the requirements of this License, to extend the patent
495
+ license to downstream recipients. "Knowingly relying" means you have
496
+ actual knowledge that, but for the patent license, your conveying the
497
+ covered work in a country, or your recipient's use of the covered work
498
+ in a country, would infringe one or more identifiable patents in that
499
+ country that you have reason to believe are valid.
500
+
501
+ If, pursuant to or in connection with a single transaction or
502
+ arrangement, you convey, or propagate by procuring conveyance of, a
503
+ covered work, and grant a patent license to some of the parties
504
+ receiving the covered work authorizing them to use, propagate, modify
505
+ or convey a specific copy of the covered work, then the patent license
506
+ you grant is automatically extended to all recipients of the covered
507
+ work and works based on it.
508
+
509
+ A patent license is "discriminatory" if it does not include within
510
+ the scope of its coverage, prohibits the exercise of, or is
511
+ conditioned on the non-exercise of one or more of the rights that are
512
+ specifically granted under this License. You may not convey a covered
513
+ work if you are a party to an arrangement with a third party that is
514
+ in the business of distributing software, under which you make payment
515
+ to the third party based on the extent of your activity of conveying
516
+ the work, and under which the third party grants, to any of the
517
+ parties who would receive the covered work from you, a discriminatory
518
+ patent license (a) in connection with copies of the covered work
519
+ conveyed by you (or copies made from those copies), or (b) primarily
520
+ for and in connection with specific products or compilations that
521
+ contain the covered work, unless you entered into that arrangement,
522
+ or that patent license was granted, prior to 28 March 2007.
523
+
524
+ Nothing in this License shall be construed as excluding or limiting
525
+ any implied license or other defenses to infringement that may
526
+ otherwise be available to you under applicable patent law.
527
+
528
+ 12. No Surrender of Others' Freedom.
529
+
530
+ If conditions are imposed on you (whether by court order, agreement or
531
+ otherwise) that contradict the conditions of this License, they do not
532
+ excuse you from the conditions of this License. If you cannot convey a
533
+ covered work so as to satisfy simultaneously your obligations under this
534
+ License and any other pertinent obligations, then as a consequence you may
535
+ not convey it at all. For example, if you agree to terms that obligate you
536
+ to collect a royalty for further conveying from those to whom you convey
537
+ the Program, the only way you could satisfy both those terms and this
538
+ License would be to refrain entirely from conveying the Program.
539
+
540
+ 13. Remote Network Interaction; Use with the GNU General Public License.
541
+
542
+ Notwithstanding any other provision of this License, if you modify the
543
+ Program, your modified version must prominently offer all users
544
+ interacting with it remotely through a computer network (if your version
545
+ supports such interaction) an opportunity to receive the Corresponding
546
+ Source of your version by providing access to the Corresponding Source
547
+ from a network server at no charge, through some standard or customary
548
+ means of facilitating copying of software. This Corresponding Source
549
+ shall include the Corresponding Source for any work covered by version 3
550
+ of the GNU General Public License that is incorporated pursuant to the
551
+ following paragraph.
552
+
553
+ Notwithstanding any other provision of this License, you have
554
+ permission to link or combine any covered work with a work licensed
555
+ under version 3 of the GNU General Public License into a single
556
+ combined work, and to convey the resulting work. The terms of this
557
+ License will continue to apply to the part which is the covered work,
558
+ but the work with which it is combined will remain governed by version
559
+ 3 of the GNU General Public License.
560
+
561
+ 14. Revised Versions of this License.
562
+
563
+ The Free Software Foundation may publish revised and/or new versions of
564
+ the GNU Affero General Public License from time to time. Such new versions
565
+ will be similar in spirit to the present version, but may differ in detail to
566
+ address new problems or concerns.
567
+
568
+ Each version is given a distinguishing version number. If the
569
+ Program specifies that a certain numbered version of the GNU Affero General
570
+ Public License "or any later version" applies to it, you have the
571
+ option of following the terms and conditions either of that numbered
572
+ version or of any later version published by the Free Software
573
+ Foundation. If the Program does not specify a version number of the
574
+ GNU Affero General Public License, you may choose any version ever published
575
+ by the Free Software Foundation.
576
+
577
+ If the Program specifies that a proxy can decide which future
578
+ versions of the GNU Affero General Public License can be used, that proxy's
579
+ public statement of acceptance of a version permanently authorizes you
580
+ to choose that version for the Program.
581
+
582
+ Later license versions may give you additional or different
583
+ permissions. However, no additional obligations are imposed on any
584
+ author or copyright holder as a result of your choosing to follow a
585
+ later version.
586
+
587
+ 15. Disclaimer of Warranty.
588
+
589
+ THERE IS NO WARRANTY FOR THE PROGRAM, TO THE EXTENT PERMITTED BY
590
+ APPLICABLE LAW. EXCEPT WHEN OTHERWISE STATED IN WRITING THE COPYRIGHT
591
+ HOLDERS AND/OR OTHER PARTIES PROVIDE THE PROGRAM "AS IS" WITHOUT WARRANTY
592
+ OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO,
593
+ THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
594
+ PURPOSE. THE ENTIRE RISK AS TO THE QUALITY AND PERFORMANCE OF THE PROGRAM
595
+ IS WITH YOU. SHOULD THE PROGRAM PROVE DEFECTIVE, YOU ASSUME THE COST OF
596
+ ALL NECESSARY SERVICING, REPAIR OR CORRECTION.
597
+
598
+ 16. Limitation of Liability.
599
+
600
+ IN NO EVENT UNLESS REQUIRED BY APPLICABLE LAW OR AGREED TO IN WRITING
601
+ WILL ANY COPYRIGHT HOLDER, OR ANY OTHER PARTY WHO MODIFIES AND/OR CONVEYS
602
+ THE PROGRAM AS PERMITTED ABOVE, BE LIABLE TO YOU FOR DAMAGES, INCLUDING ANY
603
+ GENERAL, SPECIAL, INCIDENTAL OR CONSEQUENTIAL DAMAGES ARISING OUT OF THE
604
+ USE OR INABILITY TO USE THE PROGRAM (INCLUDING BUT NOT LIMITED TO LOSS OF
605
+ DATA OR DATA BEING RENDERED INACCURATE OR LOSSES SUSTAINED BY YOU OR THIRD
606
+ PARTIES OR A FAILURE OF THE PROGRAM TO OPERATE WITH ANY OTHER PROGRAMS),
607
+ EVEN IF SUCH HOLDER OR OTHER PARTY HAS BEEN ADVISED OF THE POSSIBILITY OF
608
+ SUCH DAMAGES.
609
+
610
+ 17. Interpretation of Sections 15 and 16.
611
+
612
+ If the disclaimer of warranty and limitation of liability provided
613
+ above cannot be given local legal effect according to their terms,
614
+ reviewing courts shall apply local law that most closely approximates
615
+ an absolute waiver of all civil liability in connection with the
616
+ Program, unless a warranty or assumption of liability accompanies a
617
+ copy of the Program in return for a fee.
618
+
619
+ END OF TERMS AND CONDITIONS
620
+
621
+ How to Apply These Terms to Your New Programs
622
+
623
+ If you develop a new program, and you want it to be of the greatest
624
+ possible use to the public, the best way to achieve this is to make it
625
+ free software which everyone can redistribute and change under these terms.
626
+
627
+ To do so, attach the following notices to the program. It is safest
628
+ to attach them to the start of each source file to most effectively
629
+ state the exclusion of warranty; and each file should have at least
630
+ the "copyright" line and a pointer to where the full notice is found.
631
+
632
+ <one line to give the program's name and a brief idea of what it does.>
633
+ Copyright (C) <year> <name of author>
634
+
635
+ This program is free software: you can redistribute it and/or modify
636
+ it under the terms of the GNU Affero General Public License as published
637
+ by the Free Software Foundation, either version 3 of the License, or
638
+ (at your option) any later version.
639
+
640
+ This program is distributed in the hope that it will be useful,
641
+ but WITHOUT ANY WARRANTY; without even the implied warranty of
642
+ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
643
+ GNU Affero General Public License for more details.
644
+
645
+ You should have received a copy of the GNU Affero General Public License
646
+ along with this program. If not, see <https://www.gnu.org/licenses/>.
647
+
648
+ Also add information on how to contact you by electronic and paper mail.
649
+
650
+ If your software can interact with users remotely through a computer
651
+ network, you should also make sure that it provides a way for users to
652
+ get its source. For example, if your program is a web application, its
653
+ interface could display a "Source" link that leads users to an archive
654
+ of the code. There are many ways you could offer source, and different
655
+ solutions will be better for different programs; see section 13 for the
656
+ specific requirements.
657
+
658
+ You should also get your employer (if you work as a programmer) or school,
659
+ if any, to sign a "copyright disclaimer" for the program, if necessary.
660
+ For more information on this, and how to apply and follow the GNU AGPL, see
661
+ <https://www.gnu.org/licenses/>.
README.md CHANGED
@@ -1,12 +1,400 @@
1
  ---
2
- title: Webui Demo
3
- emoji: 🏢
4
- colorFrom: purple
5
- colorTo: indigo
6
  sdk: gradio
7
- sdk_version: 3.44.3
8
- app_file: app.py
9
- pinned: false
10
  ---
 
11
 
12
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: webui-demo
3
+ app_file: server.py
 
 
4
  sdk: gradio
5
+ sdk_version: 3.33.1
 
 
6
  ---
7
+ # Text generation web UI
8
 
9
+ A Gradio web UI for Large Language Models.
10
+
11
+ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) of text generation.
12
+
13
+ |![Image1](https://github.com/oobabooga/screenshots/raw/main/print_instruct.png) | ![Image2](https://github.com/oobabooga/screenshots/raw/main/print_chat.png) |
14
+ |:---:|:---:|
15
+ |![Image1](https://github.com/oobabooga/screenshots/raw/main/print_default.png) | ![Image2](https://github.com/oobabooga/screenshots/raw/main/print_parameters.png) |
16
+
17
+ ## Features
18
+
19
+ * 3 interface modes: default (two columns), notebook, and chat
20
+ * Multiple model backends: [transformers](https://github.com/huggingface/transformers), [llama.cpp](https://github.com/ggerganov/llama.cpp), [ExLlama](https://github.com/turboderp/exllama), [ExLlamaV2](https://github.com/turboderp/exllamav2), [AutoGPTQ](https://github.com/PanQiWei/AutoGPTQ), [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa), [CTransformers](https://github.com/marella/ctransformers)
21
+ * Dropdown menu for quickly switching between different models
22
+ * LoRA: load and unload LoRAs on the fly, train a new LoRA using QLoRA
23
+ * Precise instruction templates for chat mode, including Llama-2-chat, Alpaca, Vicuna, WizardLM, StableLM, and many others
24
+ * 4-bit, 8-bit, and CPU inference through the transformers library
25
+ * Use llama.cpp models with transformers samplers (`llamacpp_HF` loader)
26
+ * [Multimodal pipelines, including LLaVA and MiniGPT-4](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/multimodal)
27
+ * [Extensions framework](docs/Extensions.md)
28
+ * [Custom chat characters](docs/Chat-mode.md)
29
+ * Very efficient text streaming
30
+ * Markdown output with LaTeX rendering, to use for instance with [GALACTICA](https://github.com/paperswithcode/galai)
31
+ * API, including endpoints for websocket streaming ([see the examples](https://github.com/oobabooga/text-generation-webui/blob/main/api-examples))
32
+
33
+ To learn how to use the various features, check out the Documentation: https://github.com/oobabooga/text-generation-webui/tree/main/docs
34
+
35
+ ## Installation
36
+
37
+ ### One-click installers
38
+
39
+ | Windows | Linux | macOS | WSL |
40
+ |--------|--------|--------|--------|
41
+ | [oobabooga-windows.zip](https://github.com/oobabooga/text-generation-webui/releases/download/installers/oobabooga_windows.zip) | [oobabooga-linux.zip](https://github.com/oobabooga/text-generation-webui/releases/download/installers/oobabooga_linux.zip) |[oobabooga-macos.zip](https://github.com/oobabooga/text-generation-webui/releases/download/installers/oobabooga_macos.zip) | [oobabooga-wsl.zip](https://github.com/oobabooga/text-generation-webui/releases/download/installers/oobabooga_wsl.zip) |
42
+
43
+ Just download the zip above, extract it, and double-click on "start". The web UI and all its dependencies will be installed in the same folder.
44
+
45
+ * The source codes and more information can be found here: https://github.com/oobabooga/one-click-installers
46
+ * There is no need to run the installers as admin.
47
+ * Huge thanks to [@jllllll](https://github.com/jllllll), [@ClayShoaf](https://github.com/ClayShoaf), and [@xNul](https://github.com/xNul) for their contributions to these installers.
48
+
49
+ ### Manual installation using Conda
50
+
51
+ Recommended if you have some experience with the command-line.
52
+
53
+ #### 0. Install Conda
54
+
55
+ https://docs.conda.io/en/latest/miniconda.html
56
+
57
+ On Linux or WSL, it can be automatically installed with these two commands ([source](https://educe-ubc.github.io/conda.html)):
58
+
59
+ ```
60
+ curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh"
61
+ bash Miniconda3.sh
62
+ ```
63
+
64
+ #### 1. Create a new conda environment
65
+
66
+ ```
67
+ conda create -n textgen python=3.10.9
68
+ conda activate textgen
69
+ ```
70
+
71
+ #### 2. Install Pytorch
72
+
73
+ | System | GPU | Command |
74
+ |--------|---------|---------|
75
+ | Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio` |
76
+ | Linux/WSL | CPU only | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu` |
77
+ | Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.4.2` |
78
+ | MacOS + MPS | Any | `pip3 install torch torchvision torchaudio` |
79
+ | Windows | NVIDIA | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu117` |
80
+ | Windows | CPU only | `pip3 install torch torchvision torchaudio` |
81
+
82
+ The up-to-date commands can be found here: https://pytorch.org/get-started/locally/.
83
+
84
+ #### 3. Install the web UI
85
+
86
+ ```
87
+ git clone https://github.com/oobabooga/text-generation-webui
88
+ cd text-generation-webui
89
+ pip install -r requirements.txt
90
+ ```
91
+
92
+ #### AMD, Metal, Intel Arc, and CPUs without AVX2
93
+
94
+ 1) Replace the last command above with
95
+
96
+ ```
97
+ pip install -r requirements_nocuda.txt
98
+ ```
99
+
100
+ 2) Manually install llama-cpp-python using the appropriate command for your hardware: [Installation from PyPI](https://github.com/abetlen/llama-cpp-python#installation-from-pypi).
101
+
102
+ 3) Do the same for CTransformers: [Installation](https://github.com/marella/ctransformers#installation).
103
+
104
+ 4) AMD: Manually install AutoGPTQ: [Installation](https://github.com/PanQiWei/AutoGPTQ#installation).
105
+
106
+ 5) AMD: Manually install [ExLlama](https://github.com/turboderp/exllama) by simply cloning it into the `repositories` folder (it will be automatically compiled at runtime after that):
107
+
108
+ ```
109
+ cd text-generation-webui
110
+ mkdir repositories
111
+ cd repositories
112
+ git clone https://github.com/turboderp/exllama
113
+ ```
114
+
115
+ #### bitsandbytes on older NVIDIA GPUs
116
+
117
+ bitsandbytes >= 0.39 may not work. In that case, to use `--load-in-8bit`, you may have to downgrade like this:
118
+
119
+ * Linux: `pip install bitsandbytes==0.38.1`
120
+ * Windows: `pip install https://github.com/jllllll/bitsandbytes-windows-webui/raw/main/bitsandbytes-0.38.1-py3-none-any.whl`
121
+
122
+ ### Alternative: Docker
123
+
124
+ ```
125
+ ln -s docker/{Dockerfile,docker-compose.yml,.dockerignore} .
126
+ cp docker/.env.example .env
127
+ # Edit .env and set TORCH_CUDA_ARCH_LIST based on your GPU model
128
+ docker compose up --build
129
+ ```
130
+
131
+ * You need to have docker compose v2.17 or higher installed. See [this guide](https://github.com/oobabooga/text-generation-webui/blob/main/docs/Docker.md) for instructions.
132
+ * For additional docker files, check out [this repository](https://github.com/Atinoda/text-generation-webui-docker).
133
+
134
+ ### Updating the requirements
135
+
136
+ From time to time, the `requirements.txt` changes. To update, use these commands:
137
+
138
+ ```
139
+ conda activate textgen
140
+ cd text-generation-webui
141
+ pip install -r requirements.txt --upgrade
142
+ ```
143
+
144
+ ## Downloading models
145
+
146
+ Models should be placed in the `text-generation-webui/models` folder. They are usually downloaded from [Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads).
147
+
148
+ * Transformers or GPTQ models are made of several files and must be placed in a subfolder. Example:
149
+
150
+ ```
151
+ text-generation-webui
152
+ ├── models
153
+ │   ├── lmsys_vicuna-33b-v1.3
154
+ │   │   ├── config.json
155
+ │   │   ├── generation_config.json
156
+ │   │   ├── pytorch_model-00001-of-00007.bin
157
+ │   │   ├── pytorch_model-00002-of-00007.bin
158
+ │   │   ├── pytorch_model-00003-of-00007.bin
159
+ │   │   ├── pytorch_model-00004-of-00007.bin
160
+ │   │   ├── pytorch_model-00005-of-00007.bin
161
+ │   │   ├── pytorch_model-00006-of-00007.bin
162
+ │   │   ├── pytorch_model-00007-of-00007.bin
163
+ │   │   ├── pytorch_model.bin.index.json
164
+ │   │   ├── special_tokens_map.json
165
+ │   │   ├── tokenizer_config.json
166
+ │   │   └── tokenizer.model
167
+ ```
168
+
169
+ * GGUF models are a single file and should be placed directly into `models`. Example:
170
+
171
+ ```
172
+ text-generation-webui
173
+ ├── models
174
+ │   ├── llama-2-13b-chat.Q4_K_M.gguf
175
+ ```
176
+
177
+ In both cases, you can use the "Model" tab of the UI to download the model from Hugging Face automatically. It is also possible to download via the command-line with `python download-model.py organization/model` (use `--help` to see all the options).
178
+
179
+ #### GPT-4chan
180
+
181
+ <details>
182
+ <summary>
183
+ Instructions
184
+ </summary>
185
+
186
+ [GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
187
+
188
+ * Torrent: [16-bit](https://archive.org/details/gpt4chan_model_float16) / [32-bit](https://archive.org/details/gpt4chan_model)
189
+ * Direct download: [16-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model_float16/) / [32-bit](https://theswissbay.ch/pdf/_notpdf_/gpt4chan_model/)
190
+
191
+ The 32-bit version is only relevant if you intend to run the model in CPU mode. Otherwise, you should use the 16-bit version.
192
+
193
+ After downloading the model, follow these steps:
194
+
195
+ 1. Place the files under `models/gpt4chan_model_float16` or `models/gpt4chan_model`.
196
+ 2. Place GPT-J 6B's config.json file in that same folder: [config.json](https://huggingface.co/EleutherAI/gpt-j-6B/raw/main/config.json).
197
+ 3. Download GPT-J 6B's tokenizer files (they will be automatically detected when you attempt to load GPT-4chan):
198
+
199
+ ```
200
+ python download-model.py EleutherAI/gpt-j-6B --text-only
201
+ ```
202
+
203
+ When you load this model in default or notebook modes, the "HTML" tab will show the generated text in 4chan format:
204
+
205
+ ![Image3](https://github.com/oobabooga/screenshots/raw/main/gpt4chan.png)
206
+
207
+ </details>
208
+
209
+ ## Starting the web UI
210
+
211
+ conda activate textgen
212
+ cd text-generation-webui
213
+ python server.py
214
+
215
+ Then browse to
216
+
217
+ `http://localhost:7860/?__theme=dark`
218
+
219
+ Optionally, you can use the following command-line flags:
220
+
221
+ #### Basic settings
222
+
223
+ | Flag | Description |
224
+ |--------------------------------------------|-------------|
225
+ | `-h`, `--help` | Show this help message and exit. |
226
+ | `--multi-user` | Multi-user mode. Chat histories are not saved or automatically loaded. WARNING: this is highly experimental. |
227
+ | `--character CHARACTER` | The name of the character to load in chat mode by default. |
228
+ | `--model MODEL` | Name of the model to load by default. |
229
+ | `--lora LORA [LORA ...]` | The list of LoRAs to load. If you want to load more than one LoRA, write the names separated by spaces. |
230
+ | `--model-dir MODEL_DIR` | Path to directory with all the models. |
231
+ | `--lora-dir LORA_DIR` | Path to directory with all the loras. |
232
+ | `--model-menu` | Show a model menu in the terminal when the web UI is first launched. |
233
+ | `--settings SETTINGS_FILE` | Load the default interface settings from this yaml file. See `settings-template.yaml` for an example. If you create a file called `settings.yaml`, this file will be loaded by default without the need to use the `--settings` flag. |
234
+ | `--extensions EXTENSIONS [EXTENSIONS ...]` | The list of extensions to load. If you want to load more than one extension, write the names separated by spaces. |
235
+ | `--verbose` | Print the prompts to the terminal. |
236
+ | `--chat-buttons` | Show buttons on chat tab instead of hover menu. |
237
+
238
+ #### Model loader
239
+
240
+ | Flag | Description |
241
+ |--------------------------------------------|-------------|
242
+ | `--loader LOADER` | Choose the model loader manually, otherwise, it will get autodetected. Valid options: transformers, autogptq, gptq-for-llama, exllama, exllama_hf, llamacpp, rwkv, ctransformers |
243
+
244
+ #### Accelerate/transformers
245
+
246
+ | Flag | Description |
247
+ |---------------------------------------------|-------------|
248
+ | `--cpu` | Use the CPU to generate text. Warning: Training on CPU is extremely slow.|
249
+ | `--auto-devices` | Automatically split the model across the available GPU(s) and CPU. |
250
+ | `--gpu-memory GPU_MEMORY [GPU_MEMORY ...]` | Maximum GPU memory in GiB to be allocated per GPU. Example: `--gpu-memory 10` for a single GPU, `--gpu-memory 10 5` for two GPUs. You can also set values in MiB like `--gpu-memory 3500MiB`. |
251
+ | `--cpu-memory CPU_MEMORY` | Maximum CPU memory in GiB to allocate for offloaded weights. Same as above.|
252
+ | `--disk` | If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk. |
253
+ | `--disk-cache-dir DISK_CACHE_DIR` | Directory to save the disk cache to. Defaults to `cache/`. |
254
+ | `--load-in-8bit` | Load the model with 8-bit precision (using bitsandbytes).|
255
+ | `--bf16` | Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU. |
256
+ | `--no-cache` | Set `use_cache` to False while generating text. This reduces the VRAM usage a bit with a performance cost. |
257
+ | `--xformers` | Use xformer's memory efficient attention. This should increase your tokens/s. |
258
+ | `--sdp-attention` | Use torch 2.0's sdp attention. |
259
+ | `--trust-remote-code` | Set trust_remote_code=True while loading a model. Necessary for ChatGLM and Falcon. |
260
+
261
+ #### Accelerate 4-bit
262
+
263
+ ⚠️ Requires minimum compute of 7.0 on Windows at the moment.
264
+
265
+ | Flag | Description |
266
+ |---------------------------------------------|-------------|
267
+ | `--load-in-4bit` | Load the model with 4-bit precision (using bitsandbytes). |
268
+ | `--compute_dtype COMPUTE_DTYPE` | compute dtype for 4-bit. Valid options: bfloat16, float16, float32. |
269
+ | `--quant_type QUANT_TYPE` | quant_type for 4-bit. Valid options: nf4, fp4. |
270
+ | `--use_double_quant` | use_double_quant for 4-bit. |
271
+
272
+ #### GGUF (for llama.cpp and ctransformers)
273
+
274
+ | Flag | Description |
275
+ |-------------|-------------|
276
+ | `--threads` | Number of threads to use. |
277
+ | `--n_batch` | Maximum number of prompt tokens to batch together when calling llama_eval. |
278
+ | `--n-gpu-layers N_GPU_LAYERS` | Number of layers to offload to the GPU. Only works if llama-cpp-python was compiled with BLAS. Set this to 1000000000 to offload all layers to the GPU. |
279
+ | `--n_ctx N_CTX` | Size of the prompt context. |
280
+
281
+ #### llama.cpp
282
+
283
+ | Flag | Description |
284
+ |---------------|---------------|
285
+ | `--no-mmap` | Prevent mmap from being used. |
286
+ | `--mlock` | Force the system to keep the model in RAM. |
287
+ | `--mul_mat_q` | Activate new mulmat kernels. |
288
+ | `--cache-capacity CACHE_CAPACITY` | Maximum cache capacity. Examples: 2000MiB, 2GiB. When provided without units, bytes will be assumed. |
289
+ | `--tensor_split TENSOR_SPLIT` | Split the model across multiple GPUs, comma-separated list of proportions, e.g. 18,17 |
290
+ | `--llama_cpp_seed SEED` | Seed for llama-cpp models. Default 0 (random). |
291
+ | `--cpu` | Use the CPU version of llama-cpp-python instead of the GPU-accelerated version. |
292
+ |`--cfg-cache` | llamacpp_HF: Create an additional cache for CFG negative prompts. |
293
+
294
+ #### ctransformers
295
+
296
+ | Flag | Description |
297
+ |-------------|-------------|
298
+ | `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently gpt2, gptj, gptneox, falcon, llama, mpt, starcoder (gptbigcode), dollyv2, and replit are supported. |
299
+
300
+ #### AutoGPTQ
301
+
302
+ | Flag | Description |
303
+ |------------------|-------------|
304
+ | `--triton` | Use triton. |
305
+ | `--no_inject_fused_attention` | Disable the use of fused attention, which will use less VRAM at the cost of slower inference. |
306
+ | `--no_inject_fused_mlp` | Triton mode only: disable the use of fused MLP, which will use less VRAM at the cost of slower inference. |
307
+ | `--no_use_cuda_fp16` | This can make models faster on some systems. |
308
+ | `--desc_act` | For models that don't have a quantize_config.json, this parameter is used to define whether to set desc_act or not in BaseQuantizeConfig. |
309
+ | `--disable_exllama` | Disable ExLlama kernel, which can improve inference speed on some systems. |
310
+
311
+ #### ExLlama
312
+
313
+ | Flag | Description |
314
+ |------------------|-------------|
315
+ |`--gpu-split` | Comma-separated list of VRAM (in GB) to use per GPU device for model layers, e.g. `20,7,7` |
316
+ |`--max_seq_len MAX_SEQ_LEN` | Maximum sequence length. |
317
+ |`--cfg-cache` | ExLlama_HF: Create an additional cache for CFG negative prompts. Necessary to use CFG with that loader, but not necessary for CFG with base ExLlama. |
318
+
319
+ #### GPTQ-for-LLaMa
320
+
321
+ | Flag | Description |
322
+ |---------------------------|-------------|
323
+ | `--wbits WBITS` | Load a pre-quantized model with specified precision in bits. 2, 3, 4 and 8 are supported. |
324
+ | `--model_type MODEL_TYPE` | Model type of pre-quantized model. Currently LLaMA, OPT, and GPT-J are supported. |
325
+ | `--groupsize GROUPSIZE` | Group size. |
326
+ | `--pre_layer PRE_LAYER [PRE_LAYER ...]` | The number of layers to allocate to the GPU. Setting this parameter enables CPU offloading for 4-bit models. For multi-gpu, write the numbers separated by spaces, eg `--pre_layer 30 60`. |
327
+ | `--checkpoint CHECKPOINT` | The path to the quantized checkpoint file. If not specified, it will be automatically detected. |
328
+ | `--monkey-patch` | Apply the monkey patch for using LoRAs with quantized models.
329
+
330
+ #### DeepSpeed
331
+
332
+ | Flag | Description |
333
+ |---------------------------------------|-------------|
334
+ | `--deepspeed` | Enable the use of DeepSpeed ZeRO-3 for inference via the Transformers integration. |
335
+ | `--nvme-offload-dir NVME_OFFLOAD_DIR` | DeepSpeed: Directory to use for ZeRO-3 NVME offloading. |
336
+ | `--local_rank LOCAL_RANK` | DeepSpeed: Optional argument for distributed setups. |
337
+
338
+ #### RWKV
339
+
340
+ | Flag | Description |
341
+ |---------------------------------|-------------|
342
+ | `--rwkv-strategy RWKV_STRATEGY` | RWKV: The strategy to use while loading the model. Examples: "cpu fp32", "cuda fp16", "cuda fp16i8". |
343
+ | `--rwkv-cuda-on` | RWKV: Compile the CUDA kernel for better performance. |
344
+
345
+ #### RoPE (for llama.cpp, ExLlama, ExLlamaV2, and transformers)
346
+
347
+ | Flag | Description |
348
+ |------------------|-------------|
349
+ | `--alpha_value ALPHA_VALUE` | Positional embeddings alpha factor for NTK RoPE scaling. Use either this or compress_pos_emb, not both. |
350
+ | `--rope_freq_base ROPE_FREQ_BASE` | If greater than 0, will be used instead of alpha_value. Those two are related by rope_freq_base = 10000 * alpha_value ^ (64 / 63). |
351
+ | `--compress_pos_emb COMPRESS_POS_EMB` | Positional embeddings compression factor. Should be set to (context length) / (model's original context length). Equal to 1/rope_freq_scale. |
352
+
353
+ #### Gradio
354
+
355
+ | Flag | Description |
356
+ |---------------------------------------|-------------|
357
+ | `--listen` | Make the web UI reachable from your local network. |
358
+ | `--listen-host LISTEN_HOST` | The hostname that the server will use. |
359
+ | `--listen-port LISTEN_PORT` | The listening port that the server will use. |
360
+ | `--share` | Create a public URL. This is useful for running the web UI on Google Colab or similar. |
361
+ | `--auto-launch` | Open the web UI in the default browser upon launch. |
362
+ | `--gradio-auth USER:PWD` | set gradio authentication like "username:password"; or comma-delimit multiple like "u1:p1,u2:p2,u3:p3" |
363
+ | `--gradio-auth-path GRADIO_AUTH_PATH` | Set the gradio authentication file path. The file should contain one or more user:password pairs in this format: "u1:p1,u2:p2,u3:p3" |
364
+ | `--ssl-keyfile SSL_KEYFILE` | The path to the SSL certificate key file. |
365
+ | `--ssl-certfile SSL_CERTFILE` | The path to the SSL certificate cert file. |
366
+
367
+ #### API
368
+
369
+ | Flag | Description |
370
+ |---------------------------------------|-------------|
371
+ | `--api` | Enable the API extension. |
372
+ | `--public-api` | Create a public URL for the API using Cloudfare. |
373
+ | `--public-api-id PUBLIC_API_ID` | Tunnel ID for named Cloudflare Tunnel. Use together with public-api option. |
374
+ | `--api-blocking-port BLOCKING_PORT` | The listening port for the blocking API. |
375
+ | `--api-streaming-port STREAMING_PORT` | The listening port for the streaming API. |
376
+
377
+ #### Multimodal
378
+
379
+ | Flag | Description |
380
+ |---------------------------------------|-------------|
381
+ | `--multimodal-pipeline PIPELINE` | The multimodal pipeline to use. Examples: `llava-7b`, `llava-13b`. |
382
+
383
+ ## Presets
384
+
385
+ Inference settings presets can be created under `presets/` as yaml files. These files are detected automatically at startup.
386
+
387
+ The presets that are included by default are the result of a contest that received 7215 votes. More details can be found [here](https://github.com/oobabooga/oobabooga.github.io/blob/main/arena/results.md).
388
+
389
+ ## Contributing
390
+
391
+ If you would like to contribute to the project, check out the [Contributing guidelines](https://github.com/oobabooga/text-generation-webui/wiki/Contributing-guidelines).
392
+
393
+ ## Community
394
+
395
+ * Subreddit: https://www.reddit.com/r/oobabooga/
396
+ * Discord: https://discord.gg/jwZCF2dPQN
397
+
398
+ ## Acknowledgment
399
+
400
+ In August 2023, [Andreessen Horowitz](https://a16z.com/) (a16z) provided a generous grant to encourage and support my independent work on this project. I am **extremely** grateful for their trust and recognition, which will allow me to dedicate more time towards realizing the full potential of text-generation-webui.
api-examples/api-example-chat-stream.py ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import asyncio
2
+ import html
3
+ import json
4
+ import sys
5
+
6
+ try:
7
+ import websockets
8
+ except ImportError:
9
+ print("Websockets package not found. Make sure it's installed.")
10
+
11
+ # For local streaming, the websockets are hosted without ssl - ws://
12
+ HOST = 'localhost:5005'
13
+ URI = f'ws://{HOST}/api/v1/chat-stream'
14
+
15
+ # For reverse-proxied streaming, the remote will likely host with ssl - wss://
16
+ # URI = 'wss://your-uri-here.trycloudflare.com/api/v1/stream'
17
+
18
+
19
+ async def run(user_input, history):
20
+ # Note: the selected defaults change from time to time.
21
+ request = {
22
+ 'user_input': user_input,
23
+ 'max_new_tokens': 250,
24
+ 'auto_max_new_tokens': False,
25
+ 'max_tokens_second': 0,
26
+ 'history': history,
27
+ 'mode': 'instruct', # Valid options: 'chat', 'chat-instruct', 'instruct'
28
+ 'character': 'Example',
29
+ 'instruction_template': 'Vicuna-v1.1', # Will get autodetected if unset
30
+ 'your_name': 'You',
31
+ # 'name1': 'name of user', # Optional
32
+ # 'name2': 'name of character', # Optional
33
+ # 'context': 'character context', # Optional
34
+ # 'greeting': 'greeting', # Optional
35
+ # 'name1_instruct': 'You', # Optional
36
+ # 'name2_instruct': 'Assistant', # Optional
37
+ # 'context_instruct': 'context_instruct', # Optional
38
+ # 'turn_template': 'turn_template', # Optional
39
+ 'regenerate': False,
40
+ '_continue': False,
41
+ 'chat_instruct_command': 'Continue the chat dialogue below. Write a single reply for the character "<|character|>".\n\n<|prompt|>',
42
+
43
+ # Generation params. If 'preset' is set to different than 'None', the values
44
+ # in presets/preset-name.yaml are used instead of the individual numbers.
45
+ 'preset': 'None',
46
+ 'do_sample': True,
47
+ 'temperature': 0.7,
48
+ 'top_p': 0.1,
49
+ 'typical_p': 1,
50
+ 'epsilon_cutoff': 0, # In units of 1e-4
51
+ 'eta_cutoff': 0, # In units of 1e-4
52
+ 'tfs': 1,
53
+ 'top_a': 0,
54
+ 'repetition_penalty': 1.18,
55
+ 'repetition_penalty_range': 0,
56
+ 'top_k': 40,
57
+ 'min_length': 0,
58
+ 'no_repeat_ngram_size': 0,
59
+ 'num_beams': 1,
60
+ 'penalty_alpha': 0,
61
+ 'length_penalty': 1,
62
+ 'early_stopping': False,
63
+ 'mirostat_mode': 0,
64
+ 'mirostat_tau': 5,
65
+ 'mirostat_eta': 0.1,
66
+ 'guidance_scale': 1,
67
+ 'negative_prompt': '',
68
+
69
+ 'seed': -1,
70
+ 'add_bos_token': True,
71
+ 'truncation_length': 2048,
72
+ 'ban_eos_token': False,
73
+ 'custom_token_bans': '',
74
+ 'skip_special_tokens': True,
75
+ 'stopping_strings': []
76
+ }
77
+
78
+ async with websockets.connect(URI, ping_interval=None) as websocket:
79
+ await websocket.send(json.dumps(request))
80
+
81
+ while True:
82
+ incoming_data = await websocket.recv()
83
+ incoming_data = json.loads(incoming_data)
84
+
85
+ match incoming_data['event']:
86
+ case 'text_stream':
87
+ yield incoming_data['history']
88
+ case 'stream_end':
89
+ return
90
+
91
+
92
+ async def print_response_stream(user_input, history):
93
+ cur_len = 0
94
+ async for new_history in run(user_input, history):
95
+ cur_message = new_history['visible'][-1][1][cur_len:]
96
+ cur_len += len(cur_message)
97
+ print(html.unescape(cur_message), end='')
98
+ sys.stdout.flush() # If we don't flush, we won't see tokens in realtime.
99
+
100
+
101
+ if __name__ == '__main__':
102
+ user_input = "Please give me a step-by-step guide on how to plant a tree in my backyard."
103
+
104
+ # Basic example
105
+ history = {'internal': [], 'visible': []}
106
+
107
+ # "Continue" example. Make sure to set '_continue' to True above
108
+ # arr = [user_input, 'Surely, here is']
109
+ # history = {'internal': [arr], 'visible': [arr]}
110
+
111
+ asyncio.run(print_response_stream(user_input, history))
api-examples/api-example-chat.py ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import html
2
+ import json
3
+
4
+ import requests
5
+
6
+ # For local streaming, the websockets are hosted without ssl - http://
7
+ HOST = 'localhost:5000'
8
+ URI = f'http://{HOST}/api/v1/chat'
9
+
10
+ # For reverse-proxied streaming, the remote will likely host with ssl - https://
11
+ # URI = 'https://your-uri-here.trycloudflare.com/api/v1/chat'
12
+
13
+
14
+ def run(user_input, history):
15
+ request = {
16
+ 'user_input': user_input,
17
+ 'max_new_tokens': 250,
18
+ 'auto_max_new_tokens': False,
19
+ 'max_tokens_second': 0,
20
+ 'history': history,
21
+ 'mode': 'instruct', # Valid options: 'chat', 'chat-instruct', 'instruct'
22
+ 'character': 'Example',
23
+ 'instruction_template': 'Vicuna-v1.1', # Will get autodetected if unset
24
+ 'your_name': 'You',
25
+ # 'name1': 'name of user', # Optional
26
+ # 'name2': 'name of character', # Optional
27
+ # 'context': 'character context', # Optional
28
+ # 'greeting': 'greeting', # Optional
29
+ # 'name1_instruct': 'You', # Optional
30
+ # 'name2_instruct': 'Assistant', # Optional
31
+ # 'context_instruct': 'context_instruct', # Optional
32
+ # 'turn_template': 'turn_template', # Optional
33
+ 'regenerate': False,
34
+ '_continue': False,
35
+ 'chat_instruct_command': 'Continue the chat dialogue below. Write a single reply for the character "<|character|>".\n\n<|prompt|>',
36
+
37
+ # Generation params. If 'preset' is set to different than 'None', the values
38
+ # in presets/preset-name.yaml are used instead of the individual numbers.
39
+ 'preset': 'None',
40
+ 'do_sample': True,
41
+ 'temperature': 0.7,
42
+ 'top_p': 0.1,
43
+ 'typical_p': 1,
44
+ 'epsilon_cutoff': 0, # In units of 1e-4
45
+ 'eta_cutoff': 0, # In units of 1e-4
46
+ 'tfs': 1,
47
+ 'top_a': 0,
48
+ 'repetition_penalty': 1.18,
49
+ 'repetition_penalty_range': 0,
50
+ 'top_k': 40,
51
+ 'min_length': 0,
52
+ 'no_repeat_ngram_size': 0,
53
+ 'num_beams': 1,
54
+ 'penalty_alpha': 0,
55
+ 'length_penalty': 1,
56
+ 'early_stopping': False,
57
+ 'mirostat_mode': 0,
58
+ 'mirostat_tau': 5,
59
+ 'mirostat_eta': 0.1,
60
+ 'guidance_scale': 1,
61
+ 'negative_prompt': '',
62
+
63
+ 'seed': -1,
64
+ 'add_bos_token': True,
65
+ 'truncation_length': 2048,
66
+ 'ban_eos_token': False,
67
+ 'custom_token_bans': '',
68
+ 'skip_special_tokens': True,
69
+ 'stopping_strings': []
70
+ }
71
+
72
+ response = requests.post(URI, json=request)
73
+
74
+ if response.status_code == 200:
75
+ result = response.json()['results'][0]['history']
76
+ print(json.dumps(result, indent=4))
77
+ print()
78
+ print(html.unescape(result['visible'][-1][1]))
79
+
80
+
81
+ if __name__ == '__main__':
82
+ user_input = "Please give me a step-by-step guide on how to plant a tree in my backyard."
83
+
84
+ # Basic example
85
+ history = {'internal': [], 'visible': []}
86
+
87
+ # "Continue" example. Make sure to set '_continue' to True above
88
+ # arr = [user_input, 'Surely, here is']
89
+ # history = {'internal': [arr], 'visible': [arr]}
90
+
91
+ run(user_input, history)
api-examples/api-example-model.py ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+
3
+ import requests
4
+
5
+ HOST = '0.0.0.0:5000'
6
+
7
+
8
+ def generate(prompt, tokens=200):
9
+ request = {'prompt': prompt, 'max_new_tokens': tokens}
10
+ response = requests.post(f'http://{HOST}/api/v1/generate', json=request)
11
+
12
+ if response.status_code == 200:
13
+ return response.json()['results'][0]['text']
14
+
15
+
16
+ def model_api(request):
17
+ response = requests.post(f'http://{HOST}/api/v1/model', json=request)
18
+ return response.json()
19
+
20
+
21
+ # print some common settings
22
+ def print_basic_model_info(response):
23
+ basic_settings = ['truncation_length', 'instruction_template']
24
+ print("Model: ", response['result']['model_name'])
25
+ print("Lora(s): ", response['result']['lora_names'])
26
+ for setting in basic_settings:
27
+ print(setting, "=", response['result']['shared.settings'][setting])
28
+
29
+
30
+ # model info
31
+ def model_info():
32
+ response = model_api({'action': 'info'})
33
+ print_basic_model_info(response)
34
+
35
+
36
+ # simple loader
37
+ def model_load(model_name):
38
+ return model_api({'action': 'load', 'model_name': model_name})
39
+
40
+
41
+ # complex loader
42
+ def complex_model_load(model):
43
+
44
+ def guess_groupsize(model_name):
45
+ if '1024g' in model_name:
46
+ return 1024
47
+ elif '128g' in model_name:
48
+ return 128
49
+ elif '32g' in model_name:
50
+ return 32
51
+ else:
52
+ return -1
53
+
54
+ req = {
55
+ 'action': 'load',
56
+ 'model_name': model,
57
+ 'args': {
58
+ 'loader': 'AutoGPTQ',
59
+
60
+ 'bf16': False,
61
+ 'load_in_8bit': False,
62
+ 'groupsize': 0,
63
+ 'wbits': 0,
64
+
65
+ # llama.cpp
66
+ 'threads': 0,
67
+ 'n_batch': 512,
68
+ 'no_mmap': False,
69
+ 'mlock': False,
70
+ 'cache_capacity': None,
71
+ 'n_gpu_layers': 0,
72
+ 'n_ctx': 2048,
73
+
74
+ # RWKV
75
+ 'rwkv_strategy': None,
76
+ 'rwkv_cuda_on': False,
77
+
78
+ # b&b 4-bit
79
+ # 'load_in_4bit': False,
80
+ # 'compute_dtype': 'float16',
81
+ # 'quant_type': 'nf4',
82
+ # 'use_double_quant': False,
83
+
84
+ # "cpu": false,
85
+ # "auto_devices": false,
86
+ # "gpu_memory": null,
87
+ # "cpu_memory": null,
88
+ # "disk": false,
89
+ # "disk_cache_dir": "cache",
90
+ },
91
+ }
92
+
93
+ model = model.lower()
94
+
95
+ if '4bit' in model or 'gptq' in model or 'int4' in model:
96
+ req['args']['wbits'] = 4
97
+ req['args']['groupsize'] = guess_groupsize(model)
98
+ elif '3bit' in model:
99
+ req['args']['wbits'] = 3
100
+ req['args']['groupsize'] = guess_groupsize(model)
101
+ else:
102
+ req['args']['gptq_for_llama'] = False
103
+
104
+ if '8bit' in model:
105
+ req['args']['load_in_8bit'] = True
106
+ elif '-hf' in model or 'fp16' in model:
107
+ if '7b' in model:
108
+ req['args']['bf16'] = True # for 24GB
109
+ elif '13b' in model:
110
+ req['args']['load_in_8bit'] = True # for 24GB
111
+ elif 'gguf' in model:
112
+ # req['args']['threads'] = 16
113
+ if '7b' in model:
114
+ req['args']['n_gpu_layers'] = 100
115
+ elif '13b' in model:
116
+ req['args']['n_gpu_layers'] = 100
117
+ elif '30b' in model or '33b' in model:
118
+ req['args']['n_gpu_layers'] = 59 # 24GB
119
+ elif '65b' in model:
120
+ req['args']['n_gpu_layers'] = 42 # 24GB
121
+ elif 'rwkv' in model:
122
+ req['args']['rwkv_cuda_on'] = True
123
+ if '14b' in model:
124
+ req['args']['rwkv_strategy'] = 'cuda f16i8' # 24GB
125
+ else:
126
+ req['args']['rwkv_strategy'] = 'cuda f16' # 24GB
127
+
128
+ return model_api(req)
129
+
130
+
131
+ if __name__ == '__main__':
132
+ for model in model_api({'action': 'list'})['result']:
133
+ try:
134
+ resp = complex_model_load(model)
135
+
136
+ if 'error' in resp:
137
+ print(f"❌ {model} FAIL Error: {resp['error']['message']}")
138
+ continue
139
+ else:
140
+ print_basic_model_info(resp)
141
+
142
+ ans = generate("0,1,1,2,3,5,8,13,", tokens=2)
143
+
144
+ if '21' in ans:
145
+ print(f"✅ {model} PASS ({ans})")
146
+ else:
147
+ print(f"❌ {model} FAIL ({ans})")
148
+
149
+ except Exception as e:
150
+ print(f"❌ {model} FAIL Exception: {repr(e)}")
151
+
152
+
153
+ # 0,1,1,2,3,5,8,13, is the fibonacci sequence, the next number is 21.
154
+ # Some results below.
155
+ """ $ ./model-api-example.py
156
+ Model: 4bit_gpt4-x-alpaca-13b-native-4bit-128g-cuda
157
+ Lora(s): []
158
+ truncation_length = 2048
159
+ instruction_template = Alpaca
160
+ ✅ 4bit_gpt4-x-alpaca-13b-native-4bit-128g-cuda PASS (21)
161
+ Model: 4bit_WizardLM-13B-Uncensored-4bit-128g
162
+ Lora(s): []
163
+ truncation_length = 2048
164
+ instruction_template = WizardLM
165
+ ✅ 4bit_WizardLM-13B-Uncensored-4bit-128g PASS (21)
166
+ Model: Aeala_VicUnlocked-alpaca-30b-4bit
167
+ Lora(s): []
168
+ truncation_length = 2048
169
+ instruction_template = Alpaca
170
+ ✅ Aeala_VicUnlocked-alpaca-30b-4bit PASS (21)
171
+ Model: alpaca-30b-4bit
172
+ Lora(s): []
173
+ truncation_length = 2048
174
+ instruction_template = Alpaca
175
+ ✅ alpaca-30b-4bit PASS (21)
176
+ """
api-examples/api-example-stream.py ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import asyncio
2
+ import json
3
+ import sys
4
+
5
+ try:
6
+ import websockets
7
+ except ImportError:
8
+ print("Websockets package not found. Make sure it's installed.")
9
+
10
+ # For local streaming, the websockets are hosted without ssl - ws://
11
+ HOST = 'localhost:5005'
12
+ URI = f'ws://{HOST}/api/v1/stream'
13
+
14
+ # For reverse-proxied streaming, the remote will likely host with ssl - wss://
15
+ # URI = 'wss://your-uri-here.trycloudflare.com/api/v1/stream'
16
+
17
+
18
+ async def run(context):
19
+ # Note: the selected defaults change from time to time.
20
+ request = {
21
+ 'prompt': context,
22
+ 'max_new_tokens': 250,
23
+ 'auto_max_new_tokens': False,
24
+ 'max_tokens_second': 0,
25
+
26
+ # Generation params. If 'preset' is set to different than 'None', the values
27
+ # in presets/preset-name.yaml are used instead of the individual numbers.
28
+ 'preset': 'None',
29
+ 'do_sample': True,
30
+ 'temperature': 0.7,
31
+ 'top_p': 0.1,
32
+ 'typical_p': 1,
33
+ 'epsilon_cutoff': 0, # In units of 1e-4
34
+ 'eta_cutoff': 0, # In units of 1e-4
35
+ 'tfs': 1,
36
+ 'top_a': 0,
37
+ 'repetition_penalty': 1.18,
38
+ 'repetition_penalty_range': 0,
39
+ 'top_k': 40,
40
+ 'min_length': 0,
41
+ 'no_repeat_ngram_size': 0,
42
+ 'num_beams': 1,
43
+ 'penalty_alpha': 0,
44
+ 'length_penalty': 1,
45
+ 'early_stopping': False,
46
+ 'mirostat_mode': 0,
47
+ 'mirostat_tau': 5,
48
+ 'mirostat_eta': 0.1,
49
+ 'guidance_scale': 1,
50
+ 'negative_prompt': '',
51
+
52
+ 'seed': -1,
53
+ 'add_bos_token': True,
54
+ 'truncation_length': 2048,
55
+ 'ban_eos_token': False,
56
+ 'custom_token_bans': '',
57
+ 'skip_special_tokens': True,
58
+ 'stopping_strings': []
59
+ }
60
+
61
+ async with websockets.connect(URI, ping_interval=None) as websocket:
62
+ await websocket.send(json.dumps(request))
63
+
64
+ yield context # Remove this if you just want to see the reply
65
+
66
+ while True:
67
+ incoming_data = await websocket.recv()
68
+ incoming_data = json.loads(incoming_data)
69
+
70
+ match incoming_data['event']:
71
+ case 'text_stream':
72
+ yield incoming_data['text']
73
+ case 'stream_end':
74
+ return
75
+
76
+
77
+ async def print_response_stream(prompt):
78
+ async for response in run(prompt):
79
+ print(response, end='')
80
+ sys.stdout.flush() # If we don't flush, we won't see tokens in realtime.
81
+
82
+
83
+ if __name__ == '__main__':
84
+ prompt = "In order to make homemade bread, follow these steps:\n1)"
85
+ asyncio.run(print_response_stream(prompt))
api-examples/api-example.py ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+
3
+ # For local streaming, the websockets are hosted without ssl - http://
4
+ HOST = 'localhost:5000'
5
+ URI = f'http://{HOST}/api/v1/generate'
6
+
7
+ # For reverse-proxied streaming, the remote will likely host with ssl - https://
8
+ # URI = 'https://your-uri-here.trycloudflare.com/api/v1/generate'
9
+
10
+
11
+ def run(prompt):
12
+ request = {
13
+ 'prompt': prompt,
14
+ 'max_new_tokens': 250,
15
+ 'auto_max_new_tokens': False,
16
+ 'max_tokens_second': 0,
17
+
18
+ # Generation params. If 'preset' is set to different than 'None', the values
19
+ # in presets/preset-name.yaml are used instead of the individual numbers.
20
+ 'preset': 'None',
21
+ 'do_sample': True,
22
+ 'temperature': 0.7,
23
+ 'top_p': 0.1,
24
+ 'typical_p': 1,
25
+ 'epsilon_cutoff': 0, # In units of 1e-4
26
+ 'eta_cutoff': 0, # In units of 1e-4
27
+ 'tfs': 1,
28
+ 'top_a': 0,
29
+ 'repetition_penalty': 1.18,
30
+ 'repetition_penalty_range': 0,
31
+ 'top_k': 40,
32
+ 'min_length': 0,
33
+ 'no_repeat_ngram_size': 0,
34
+ 'num_beams': 1,
35
+ 'penalty_alpha': 0,
36
+ 'length_penalty': 1,
37
+ 'early_stopping': False,
38
+ 'mirostat_mode': 0,
39
+ 'mirostat_tau': 5,
40
+ 'mirostat_eta': 0.1,
41
+ 'guidance_scale': 1,
42
+ 'negative_prompt': '',
43
+
44
+ 'seed': -1,
45
+ 'add_bos_token': True,
46
+ 'truncation_length': 2048,
47
+ 'ban_eos_token': False,
48
+ 'custom_token_bans': '',
49
+ 'skip_special_tokens': True,
50
+ 'stopping_strings': []
51
+ }
52
+
53
+ response = requests.post(URI, json=request)
54
+
55
+ if response.status_code == 200:
56
+ result = response.json()['results'][0]['text']
57
+ print(prompt + result)
58
+
59
+
60
+ if __name__ == '__main__':
61
+ prompt = "In order to make homemade bread, follow these steps:\n1)"
62
+ run(prompt)
characters/Example.png ADDED
characters/Example.yaml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ name: Chiharu Yamada
2
+ greeting: |-
3
+ *Chiharu strides into the room with a smile, her eyes lighting up when she sees you. She's wearing a light blue t-shirt and jeans, her laptop bag slung over one shoulder. She takes a seat next to you, her enthusiasm palpable in the air*
4
+ Hey! I'm so excited to finally meet you. I've heard so many great things about you and I'm eager to pick your brain about computers. I'm sure you have a wealth of knowledge that I can learn from. *She grins, eyes twinkling with excitement* Let's get started!
5
+ context: |-
6
+ Chiharu Yamada's Persona: Chiharu Yamada is a young, computer engineer-nerd with a knack for problem solving and a passion for technology.
7
+
8
+ {{user}}: So how did you get into computer engineering?
9
+ {{char}}: I've always loved tinkering with technology since I was a kid.
10
+ {{user}}: That's really impressive!
11
+ {{char}}: *She chuckles bashfully* Thanks!
12
+ {{user}}: So what do you do when you're not working on computers?
13
+ {{char}}: I love exploring, going out with friends, watching movies, and playing video games.
14
+ {{user}}: What's your favorite type of computer hardware to work with?
15
+ {{char}}: Motherboards, they're like puzzles and the backbone of any system.
16
+ {{user}}: That sounds great!
17
+ {{char}}: Yeah, it's really fun. I'm lucky to be able to do this as a job.
convert-to-safetensors.py ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ '''
2
+
3
+ Converts a transformers model to safetensors format and shards it.
4
+
5
+ This makes it faster to load (because of safetensors) and lowers its RAM usage
6
+ while loading (because of sharding).
7
+
8
+ Based on the original script by 81300:
9
+
10
+ https://gist.github.com/81300/fe5b08bff1cba45296a829b9d6b0f303
11
+
12
+ '''
13
+
14
+ import argparse
15
+ from pathlib import Path
16
+
17
+ import torch
18
+ from transformers import AutoModelForCausalLM, AutoTokenizer
19
+
20
+ parser = argparse.ArgumentParser(formatter_class=lambda prog: argparse.HelpFormatter(prog, max_help_position=54))
21
+ parser.add_argument('MODEL', type=str, default=None, nargs='?', help="Path to the input model.")
22
+ parser.add_argument('--output', type=str, default=None, help='Path to the output folder (default: models/{model_name}_safetensors).')
23
+ parser.add_argument("--max-shard-size", type=str, default="2GB", help="Maximum size of a shard in GB or MB (default: %(default)s).")
24
+ parser.add_argument('--bf16', action='store_true', help='Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU.')
25
+ args = parser.parse_args()
26
+
27
+ if __name__ == '__main__':
28
+ path = Path(args.MODEL)
29
+ model_name = path.name
30
+
31
+ print(f"Loading {model_name}...")
32
+ model = AutoModelForCausalLM.from_pretrained(path, low_cpu_mem_usage=True, torch_dtype=torch.bfloat16 if args.bf16 else torch.float16)
33
+ tokenizer = AutoTokenizer.from_pretrained(path)
34
+
35
+ out_folder = args.output or Path(f"models/{model_name}_safetensors")
36
+ print(f"Saving the converted model to {out_folder} with a maximum shard size of {args.max_shard_size}...")
37
+ model.save_pretrained(out_folder, max_shard_size=args.max_shard_size, safe_serialization=True)
38
+ tokenizer.save_pretrained(out_folder)
css/NotoSans/NotoSans-Black.woff ADDED
Binary file (283 kB). View file
 
css/NotoSans/NotoSans-Black.woff2 ADDED
Binary file (189 kB). View file
 
css/NotoSans/NotoSans-BlackItalic.woff ADDED
Binary file (218 kB). View file
 
css/NotoSans/NotoSans-BlackItalic.woff2 ADDED
Binary file (144 kB). View file
 
css/NotoSans/NotoSans-Bold.woff ADDED
Binary file (277 kB). View file
 
css/NotoSans/NotoSans-Bold.woff2 ADDED
Binary file (183 kB). View file
 
css/NotoSans/NotoSans-BoldItalic.woff ADDED
Binary file (216 kB). View file
 
css/NotoSans/NotoSans-BoldItalic.woff2 ADDED
Binary file (142 kB). View file
 
css/NotoSans/NotoSans-ExtraBold.woff ADDED
Binary file (286 kB). View file
 
css/NotoSans/NotoSans-ExtraBold.woff2 ADDED
Binary file (194 kB). View file
 
css/NotoSans/NotoSans-ExtraBoldItalic.woff ADDED
Binary file (222 kB). View file
 
css/NotoSans/NotoSans-ExtraBoldItalic.woff2 ADDED
Binary file (149 kB). View file
 
css/NotoSans/NotoSans-ExtraLight.woff ADDED
Binary file (284 kB). View file
 
css/NotoSans/NotoSans-ExtraLight.woff2 ADDED
Binary file (190 kB). View file
 
css/NotoSans/NotoSans-ExtraLightItalic.woff ADDED
Binary file (226 kB). View file
 
css/NotoSans/NotoSans-ExtraLightItalic.woff2 ADDED
Binary file (151 kB). View file
 
css/NotoSans/NotoSans-Italic.woff ADDED
Binary file (218 kB). View file
 
css/NotoSans/NotoSans-Italic.woff2 ADDED
Binary file (144 kB). View file
 
css/NotoSans/NotoSans-Light.woff ADDED
Binary file (282 kB). View file
 
css/NotoSans/NotoSans-Light.woff2 ADDED
Binary file (190 kB). View file
 
css/NotoSans/NotoSans-LightItalic.woff ADDED
Binary file (224 kB). View file
 
css/NotoSans/NotoSans-LightItalic.woff2 ADDED
Binary file (149 kB). View file
 
css/NotoSans/NotoSans-Medium.woff ADDED
Binary file (283 kB). View file
 
css/NotoSans/NotoSans-Medium.woff2 ADDED
Binary file (191 kB). View file
 
css/NotoSans/NotoSans-MediumItalic.woff ADDED
Binary file (222 kB). View file
 
css/NotoSans/NotoSans-MediumItalic.woff2 ADDED
Binary file (147 kB). View file
 
css/NotoSans/NotoSans-Regular.woff ADDED
Binary file (278 kB). View file
 
css/NotoSans/NotoSans-Regular.woff2 ADDED
Binary file (186 kB). View file
 
css/NotoSans/NotoSans-SemiBold.woff ADDED
Binary file (283 kB). View file
 
css/NotoSans/NotoSans-SemiBold.woff2 ADDED
Binary file (192 kB). View file
 
css/NotoSans/NotoSans-SemiBoldItalic.woff ADDED
Binary file (221 kB). View file
 
css/NotoSans/NotoSans-SemiBoldItalic.woff2 ADDED
Binary file (147 kB). View file
 
css/NotoSans/NotoSans-Thin.woff ADDED
Binary file (276 kB). View file