LLM-jp-3 Fine-tuned Models Collection Fine-tuned models in the LLM-jp-3 model series β’ 5 items β’ Updated 7 days ago β’ 1
LLM-jp-3 Pre-trained Models Collection Pre-trained models in the LLM-jp-3 model series β’ 5 items β’ Updated 7 days ago β’ 1
view article Article How to generate text: using different decoding methods for language generation with Transformers Mar 1, 2020 β’ 111
PLaMo-100B: A Ground-Up Language Model Designed for Japanese Proficiency Paper β’ 2410.07563 β’ Published Oct 10 β’ 2
gemma-2-baku Collection The baku model series are based on the gemma-2 series and have been continually pre-trained on Japanese-specific corpora. β’ 4 items β’ Updated Oct 3 β’ 3
Gemma 2 JPN Release Collection A Gemma 2 2B model fine-tuned on Japanese text. It supports the Japanese language the same level of performance of EN only queries on Gemma 2. β’ 3 items β’ Updated Oct 3 β’ 25
Japanese SimCSE Collection Tsukagoshi et al., Japanese SimCSE Technical Report, arXiv 2023. https://arxiv.org/abs/2310.19349 β’ 5 items β’ Updated Sep 4 β’ 2
llama-3-youko Collection The youko model series are based on the llama-3 series and have been continually pre-trained on Japanese-specific corpora. β’ 9 items β’ Updated Sep 30 β’ 1
Llama 3.1 GPTQ, AWQ, and BNB Quants Collection Optimised Quants for high-throughput deployments! Compatible with Transformers, TGI & VLLM π€ β’ 9 items β’ Updated Sep 26 β’ 55
Sarashina Collection Large Language Models developed by SB Intuitions β’ 7 items β’ Updated 14 days ago β’ 2
LLM-jp: A Cross-organizational Project for the Research and Development of Fully Open Japanese LLMs Paper β’ 2407.03963 β’ Published Jul 4 β’ 15