학술논문
RakutenAI-7B: Extending Large Language Models for Japanese
Document Type
Working Paper
Author
Rakuten Group; Levine, Aaron; Huang, Connie; Wang, Chenguang; Batista, Eduardo; Szymanska, Ewa; Ding, Hongyi; Chou, Hou Wei; Pessiot, Jean-François; Effendi, Johanes; Chiu, Justin; Ohlhus, Kai Torben; Chopra, Karan; Shinzato, Keiji; Murakami, Koji; Xiong, Lee; Chen, Lei; Kubota, Maki; Tkachenko, Maksim; Lee, Miroku; Takahashi, Naoki; Jwalapuram, Prathyusha; Tatsushima, Ryutaro; Jain, Saurabh; Yadav, Sunil Kumar; Cai, Ting; Chen, Wei-Te; Xia, Yandi; Nakayama, Yuki; Higashiyama, Yutaka
Source
Subject
Language
Abstract
We introduce RakutenAI-7B, a suite of Japanese-oriented large language models that achieve the best performance on the Japanese LM Harness benchmarks among the open 7B models. Along with the foundation model, we release instruction- and chat-tuned models, RakutenAI-7B-instruct and RakutenAI-7B-chat respectively, under the Apache 2.0 license.