AI ExplorerAI Explorer
ToolsCategoriesSitesLLMsCompareAI QuizAlternativesPremium

—

AI Tools

—

Sites & Blogs

—

LLMs & Models

—

Categories

AI Explorer

Find and compare the best artificial intelligence tools for your projects.

Made within France

Explore

  • All tools
  • Sites & Blogs
  • LLMs & Models
  • Compare
  • Chatbots
  • AI Images
  • Code & Dev

Company

  • Premium
  • About
  • Contact
  • Blog

Legal

  • Legal notice
  • Privacy
  • Terms

© 2026 AI Explorer. All rights reserved.

HomeLLMsopen calm 3b

open calm 3b

by cyberagent

Open source · 213k downloads · 20 likes

1.7
(20 reviews)ChatAPI & Local
About

OpenCALM-3B is a Japanese language model developed by CyberAgent, designed for generating Japanese text. Trained on datasets such as Wikipedia and Common Crawl, it excels in various tasks, including writing, summarization, and question answering. Its optimized architecture makes it particularly well-suited for applications requiring a deep understanding of the Japanese language. The model stands out for its open license (CC BY-SA 4.4.0), which encourages its use and sharing while requiring clear attribution. It is ideal for developers and researchers seeking a high-performance and accessible Japanese language solution.

Documentation

OpenCALM-3B

Model Description

OpenCALM is a suite of decoder-only language models pre-trained on Japanese datasets, developed by CyberAgent, Inc.

Usage

Python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("cyberagent/open-calm-3b", device_map="auto", torch_dtype=torch.float16)
tokenizer = AutoTokenizer.from_pretrained("cyberagent/open-calm-3b")

inputs = tokenizer("AIによって私達の暮らしは、", return_tensors="pt").to(model.device)
with torch.no_grad():
    tokens = model.generate(
        **inputs,
        max_new_tokens=64,
        do_sample=True,
        temperature=0.7,
        top_p=0.9,
        repetition_penalty=1.05,
        pad_token_id=tokenizer.pad_token_id,
    )
    
output = tokenizer.decode(tokens[0], skip_special_tokens=True)
print(output)

Model Details

ModelParamsLayersDimHeadsDev ppl
cyberagent/open-calm-small160M127681219.7
cyberagent/open-calm-medium400M2410241613.8
cyberagent/open-calm-large830M2415361611.3
cyberagent/open-calm-1b1.4B2420481610.3
cyberagent/open-calm-3b2.7B322560329.7
cyberagent/open-calm-7b6.8B324096328.2
  • Developed by: CyberAgent, Inc.
  • Model type: Transformer-based Language Model
  • Language: Japanese
  • Library: GPT-NeoX
  • License: OpenCALM is licensed under the Creative Commons Attribution-ShareAlike 4.0 International License (CC BY-SA 4.0). When using this model, please provide appropriate credit to CyberAgent, Inc.
    • Example (en): This model is a fine-tuned version of OpenCALM-XX developed by CyberAgent, Inc. The original model is released under the CC BY-SA 4.0 license, and this model is also released under the same CC BY-SA 4.0 license. For more information, please visit: https://creativecommons.org/licenses/by-sa/4.0/
    • Example (ja): 本モデルは、株式会社サイバーエージェントによるOpenCALM-XXをファインチューニングしたものです。元のモデルはCC BY-SA 4.0ライセンスのもとで公開されており、本モデルも同じくCC BY-SA 4.0ライセンスで公開します。詳しくはこちらをご覧ください: https://creativecommons.org/licenses/by-sa/4.0/

Training Dataset

  • Wikipedia (ja)
  • Common Crawl (ja)

Author

Ryosuke Ishigami

Citations

Bibtext
@software{gpt-neox-library,
  title = {{GPT-NeoX: Large Scale Autoregressive Language Modeling in PyTorch}},
  author = {Andonian, Alex and Anthony, Quentin and Biderman, Stella and Black, Sid and Gali, Preetham and Gao, Leo and Hallahan, Eric and Levy-Kramer, Josh and Leahy, Connor and Nestler, Lucas and Parker, Kip and Pieler, Michael and Purohit, Shivanshu and Songz, Tri and Phil, Wang and Weinbach, Samuel},
  url = {https://www.github.com/eleutherai/gpt-neox},
  doi = {10.5281/zenodo.5879544},
  month = {8},
  year = {2021},
  version = {0.0.1},
}
Capabilities & Tags
transformerspytorchgpt_neoxtext-generationjapanesecausal-lmjatext-generation-inference
Links & Resources
Specifications
CategoryChat
AccessAPI & Local
LicenseOpen Source
PricingOpen Source
Parameters3B parameters
Rating
1.7

Try open calm 3b

Access the model directly