Model Details
Model Description
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
Model type: Mamba
Language(s) (NLP): Japanese
Tokenizer: ku-nlp/gpt2-large-japanese-char
License: Model: Aptach 2.0, Tokenizer: CC-BY-SA
Run the model
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "misdelivery/mamba-char-japanese-790m"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
input_prompt = "夏目漱石について教えてください。"
with torch.no_grad():
input_ids = tokenizer.encode(f"以下は、タスクを説明する指示です。要求を適切に満たす応答を書きなさい。\n\n### 指示:\n{input_prompt}\n\n### 応答:\n", add_special_tokens=False, return_tensors="pt")
output_ids = model.generate(
input_ids.to(model.device),
max_length=512,
do_sample=True,
temperature=0.6,
repetition_penalty=1.2
)
output_ids.tolist()[0]
result = tokenizer.decode(output_ids.tolist()[0], skip_special_tokens=True)
print(result)
【LOCAL AI HACKATHON #001】240時間ハッカソンにおいてGPUをお借りしました。
関係者の方々に深く御礼申し上げます。
メタデータラボ株式会社様 【AI声づくり技術研究会】 サーバー主:やなぎ(Yanagi)様 (@Yanagi_1112) 【ローカルLLMに向き合う会】 サーバー主:saldra(サルドラ)様 (@sald_ra) Witness様 (@i_witnessed_it)
チームメンバー: hayashi, kzms, chatblanc, Ryunosuke Ikeda
- Downloads last month
- 53
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.