by mlx-community
Open source · 625k downloads · 34 likes
Kimi K2.5 is an advanced language model designed to understand and generate text with high precision. It excels in conversational tasks, answering complex questions, and synthesizing information, thanks to its ability to process extended contexts and produce nuanced responses. Its primary use cases include conversational assistance, document analysis, creative or technical content generation, and decision-making support. What sets it apart is its balance between performance and efficiency, delivering high-quality results while remaining accessible for local use or deployment across varied infrastructures.
This model mlx-community/Kimi-K2.5 was converted to MLX format from moonshotai/Kimi-K2.5 using mlx-lm version 0.30.5 (slightly modified).
pip install mlx-lm
from mlx_lm import load, generate
model, tokenizer = load("mlx-community/Kimi-K2.5")
prompt = "hello"
if tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, add_generation_prompt=True
)
response = generate(model, tokenizer, prompt=prompt, verbose=True)