Fix tokenizer load error
#6
by
xu-song
- opened
- tokenization_moss.py +4 -3
tokenization_moss.py
CHANGED
@@ -146,6 +146,10 @@ class MossTokenizer(PreTrainedTokenizer):
|
|
146 |
eos_token = AddedToken(eos_token, lstrip=False, rstrip=False) if isinstance(eos_token, str) else eos_token
|
147 |
unk_token = AddedToken(unk_token, lstrip=False, rstrip=False) if isinstance(unk_token, str) else unk_token
|
148 |
pad_token = AddedToken(pad_token, lstrip=False, rstrip=False) if isinstance(pad_token, str) else pad_token
|
|
|
|
|
|
|
|
|
149 |
super().__init__(
|
150 |
errors=errors,
|
151 |
unk_token=unk_token,
|
@@ -156,10 +160,7 @@ class MossTokenizer(PreTrainedTokenizer):
|
|
156 |
add_bos_token=add_bos_token,
|
157 |
**kwargs,
|
158 |
)
|
159 |
-
self.add_bos_token = add_bos_token
|
160 |
|
161 |
-
with open(vocab_file, encoding="utf-8") as vocab_handle:
|
162 |
-
self.encoder = json.load(vocab_handle)
|
163 |
self.decoder = {v: k for k, v in self.encoder.items()}
|
164 |
self.errors = errors # how to handle errors in decoding
|
165 |
self.byte_encoder = bytes_to_unicode()
|
|
|
146 |
eos_token = AddedToken(eos_token, lstrip=False, rstrip=False) if isinstance(eos_token, str) else eos_token
|
147 |
unk_token = AddedToken(unk_token, lstrip=False, rstrip=False) if isinstance(unk_token, str) else unk_token
|
148 |
pad_token = AddedToken(pad_token, lstrip=False, rstrip=False) if isinstance(pad_token, str) else pad_token
|
149 |
+
self.add_bos_token = add_bos_token
|
150 |
+
with open(vocab_file, encoding="utf-8") as vocab_handle:
|
151 |
+
self.encoder = json.load(vocab_handle)
|
152 |
+
|
153 |
super().__init__(
|
154 |
errors=errors,
|
155 |
unk_token=unk_token,
|
|
|
160 |
add_bos_token=add_bos_token,
|
161 |
**kwargs,
|
162 |
)
|
|
|
163 |
|
|
|
|
|
164 |
self.decoder = {v: k for k, v in self.encoder.items()}
|
165 |
self.errors = errors # how to handle errors in decoding
|
166 |
self.byte_encoder = bytes_to_unicode()
|