Why didn't my output show up |startthought|? Such tokens

#3
by yihang0017 - opened

Why didn't my output show up |startthought|? Such tokens
here is my code:

from transformers import pipeline
local_model_path = "/localpath/Mistral-quiet-star-demo"
pipe = pipeline("text-generation", model=local_model_path, tokenizer=local_model_path)
text = '''user: "hi there. id like to build a time machine. how might i go about achieving this?"'''
output = pipe(text, max_length=50)
print(output)

Loading checkpoint shards: 100%
 3/3 [00:06<00:00,  2.15s/it]
Truncation was not explicitly activated but max_length is provided a specific value, please use truncation=True to explicitly truncate examples to max length. Defaulting to 'longest_first' truncation strategy. If you encode pairs of sequences (GLUE-style) with the tokenizer you can select this strategy more precisely by providing a specific strategy to truncation.
Setting pad_token_id to eos_token_id:2 for open-end generation.
[{'generated_text': 'user: "hi there. id like to build a time machine. how might i go about achieving this?"\n response: "you could start by reading up on the principles of relativity and quantum mechanics, then look into current research on time'}]

Sign up or log in to comment