Tokyo, December 9, 2024 - Rakuten Group, Inc. today announced that it has been awarded the Information Technology Award (Open Innovation Field)*1 in the Japan Institute of Information Technology’s 42nd IT Awards*2 in 2024 for its development of the Rakuten AI 7B large language model (LLM), optimized for the Japanese language.
- Rakuten Group, Inc.
Rakuten Receives Open Innovation Field IT Award from Japan Institute of Information Technology’s 42nd IT Awards
The IT Awards are presented by the Japan Institute of Information Technology to enterprises, organizations, groups and individuals whose outstanding efforts in driving business innovation have produced results through advanced IT applications. Award recipients have demonstrated success in creating new business opportunities, establishing effective business models and improving productivity across Japan’s industry and government institution operations.
Under its AI-nization initiative, Rakuten is integrating AI into every aspect of its business to augment human creativity, accelerate productivity and drive further growth. Going forward, Rakuten aims to continue leveraging rich data and cutting-edge AI technology to add value and enrich the lives of people around the world.
Overview of the Award-Winning Project:
Development of Rakuten AI 7B – Building a High-Performance, Japanese-Optimized Open LLM
Project details: Rakuten has developed Rakuten AI 7B, a 7 billion parameter Japanese language foundation model*3, optimized specifically for the Japanese language. Rakuten AI 7B Instruct is an instruct model*4 fine-tuned from the Rakuten AI 7B foundation model. Rakuten AI 7B Chat is a chat model*5 based on a version of the instruct model and fine-tuned on chat data for conversational text generation. All models were released as open models on March 21, 2024.
Rakuten AI 7B was developed by continuous training on Mistral-7B-v0.1, an open LLM from France-based AI company Mistral AI. The foundation and instruct models earned top average performance scores among open Japanese LLMs via LM Evaluation Harness*6 and Rakuten AI 7B has been downloaded 74,000 times as of December 2024.
Reasons for award: Rakuten was recognized for its contributions in developing high-quality LLMs by leveraging years of experience and data from e-commerce-specific text processing, utilizing a proprietary tokenizer and refining LLM training datasets. The lightweight architecture of Rakuten AI 7B, with 7 billion parameters, was also noted for addressing recent concerns about the significant energy consumption of AI operations.
Notes:
*1 IT Award (Open Innovation Field): Awarded for leveraging IT for operational standardization, information exchange, functional integration and building business platforms that lead to new business collaboration across multiple industries and businesses notably distinct from transactions and partnerships in the traditional value chain.
*2 Site only available in Japanese.
*3 Foundation models are pre-trained on vast amounts of data and can be fine-tuned for specific tasks or applications.
*4 An instruct model is a version of a foundation model fine-tuned on instruction-style data. This results in the model replying to prompted instructions.
*5 A chat model is a version of a foundation model fine-tuned on chat data, resulting in a model exhibiting chat-style text generation.
*6 The LM Evaluation Harness is a framework for evaluating language models. Evaluation tests were carried out internally by Rakuten from January to March 2024 using the same compute environment for reproducibility and fair comparison. The Japanese language tests were conducted using the v0.3 template of the following evaluation framework:
https://github.com/-AStabilityI/lm-evaluation-harness/tree/0fa86429679f521161d5b81a94c0c385e0a0976d
*The average result for Japanese evaluation is based on the average results of 7 items.
English language evaluation: https://github.com/EleutherAI/lm-evaluation-harness/tree/b281b0921b636bc36ad05c0b0b0763bd6dd43463