Japanese AI scientists release a fully open LLM

2025-01-27
2 min read.
AI scientists in Japan have developed a new LLM with about 172 billion parameters and shared parameters and training data with the public.

Artificial intelligence (AI) scientists at the National Institute of Informatics (NII) in Japan have developed a new large language model (LLM) with about 172 billion parameters. The new LLM goes by the name "llm-jp-3-172b-instruct3."

The scientists trained the model with 2.1 trillion tokens. The NII has shared both the parameters and the training data of the model with the public.

The NII scientists tested this model against others like GPT-3.5 on two benchmarks: "llm-jp-eval" for Japanese language skills and "llm-leaderboard" for general language understanding. The model did better than GPT-3.5 on both. The model was built on a platform called mdx and with help from ABCI, a supercomputer system.

For training, the NII scientists used texts in Japanese, English, Chinese, Korean, and program code. They got Japanese texts from places like Common Crawl, the National Diet Library, Wikipedia, and research project summaries. The model uses the LLaMA-2 architecture.

The NII scientists tuned the model with 13 types of Japanese instructions and some English translations. The results showed it performed well, scoring higher than GPT-3.5 in both evaluations. However, the NII team admit that ensuring complete safety in responses is challenging, and they've done what they can with current technology. They tested for safety with 181 items, and while it mostly passed, there were seven responses that didn't meet safety standards.

Accelerating innovation in AI

The model's data and tools are available online. See the website of the LLM-jp consortium and the llm-jp-3-172b-instruct3 repository at Hugging Face.

The NII team plan to keep developing these models to make them transparent and reliable, focusing on safety as they grow. They've also kept intermediate training data which they might share later. This work is part of a broader effort by NII to advance language model research in Japan.

This development by NII is pushes forward the boundaries of AI language understanding, particularly in Japanese. By making the model's parameters and training data public, it not only fosters transparency but also empowers researchers worldwide to further refine and study language models. This openness can accelerate innovation in AI.

#NeuralNetworks



Related Articles


Comments on this article

Before posting or replying to a comment, please review it carefully to avoid any errors. Reason: you are not able to edit or delete your comment on Mindplex, because every interaction is tied to our reputation system. Thanks!

Mindplex

Mindplex is an AI company, a decentralized media platform, a global brain experiment, and a community dedicated to the rapidly unfolding future. Our platform empowers our community to share and discuss futurist content while showcasing AI and blockchain tools that enhance the media experience. Join us and shape the future of digital media!

ABOUT US

FAQ

CONTACT

Editors

© 2025 MindPlex. All rights reserved