🧠
Model

Kimi K2.5 Mxfp4 Autoround

by INCModel hf-model--incmodel--kimi-k2.5-mxfp4-autoround
Nexus Index
39.6 Top 100%
S: Semantic 50
A: Authority 0
P: Popularity 10
R: Recency 98
Q: Quality 65
Tech Context
Vital Performance
125 DL / 30D
0.0%
Audited 39.6 FNI Score
Tiny - Params
- Context
125 Downloads
Restricted OTHER License
Model Information Summary
Entity Passport
Registry ID hf-model--incmodel--kimi-k2.5-mxfp4-autoround
License Other
Provider huggingface
📜

Cite this model

Academic & Research Attribution

BibTeX
@misc{hf_model__incmodel__kimi_k2.5_mxfp4_autoround,
  author = {INCModel},
  title = {Kimi K2.5 Mxfp4 Autoround Model},
  year = {2026},
  howpublished = {\url{https://huggingface.co/incmodel/kimi-k2.5-mxfp4-autoround}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
INCModel. (2026). Kimi K2.5 Mxfp4 Autoround [Model]. Free2AITools. https://huggingface.co/incmodel/kimi-k2.5-mxfp4-autoround

đŸ”ŦTechnical Deep Dive

Full Specifications [+]

Quick Commands

🤗 HF Download
huggingface-cli download incmodel/kimi-k2.5-mxfp4-autoround
đŸ“Ļ Install Lib
pip install -U transformers

âš–ī¸ Nexus Index V2.0

39.6
TOP 100% SYSTEM IMPACT
Semantic (S) 50
Authority (A) 0
Popularity (P) 10
Recency (R) 98
Quality (Q) 65

đŸ’Ŧ Index Insight

FNI V2.0 for Kimi K2.5 Mxfp4 Autoround: Semantic (S:50), Authority (A:0), Popularity (P:10), Recency (R:98), Quality (Q:65).

Free2AITools Nexus Index

Verification Authority

Unbiased Data Node Refresh: VFS Live
---

🚀 What's Next?

Technical Deep Dive

Model Details

This model is a MXFP4 model of moonshotai/Kimi-K2.5 generated by intel/auto-round with RTN mode. Please follow the license of the original model.

vllm Infernece Example

bash
vllm serve INCModel/Kimi-K2.5-MXFP4-AutoRound -tp 8 --mm-encoder-tp-mode data --trust-remote-code --tool-call-parser kimi_k2 --reasoning-parser kimi_k2  --served-model-name kimi  --max-model-len 4096
bash
curl http://localhost:8000/v1/chat/completions -H "Content-Type: application/json" -d ' {
    "model": "kimi",
    "messages": [
      {"role": "system", "content": "You are a helpful assistant."},
      {"role": "user", "content": "Write code to fine-tune an LLM."}
    ],
    "temperature": 1,
    "max_tokens": 2048
  } '

Generate the Model

This pr is required https://github.com/intel/auto-round/pull/1642

RTN version

bash
auto-round /workspace/models/moonshotai/Kimi-K2.5  --iters 0 --disable_opt_rtn --scheme mxfp4  --format llm_compressor  --output_dir /workspace/models/moonshotai/Kimi-K2.5-MXFP4
# the automatic saved preprocessor_config.json doesn't work, copy it.
cp /workspace/models/moonshotai/Kimi-K2.5/preprocessor_config.json /workspace/models/moonshotai/Kimi-K2.5-MXFP4/Kimi-K2.5-mxfp-w4g32/preprocessor_config.json

Ethical Considerations and Limitations

The model can produce factually incorrect output, and should not be relied on to produce factually accurate information. Because of the limitations of the pretrained model and the finetuning datasets, it is possible that this model could generate lewd, biased or otherwise offensive outputs.

Therefore, before deploying any applications of the model, developers should perform safety testing.

Caveats and Recommendations

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model.

Here are a couple of useful links to learn more about Intel's AI software:

Disclaimer

The license on this model does not constitute legal advice. We are not responsible for the actions of third parties who use this model. Please consult an attorney before using this model for commercial purposes.

Cite

@article{cheng2023optimize, title={Optimize weight rounding via signed gradient descent for the quantization of llms}, author={Cheng, Wenhua and Zhang, Weiwei and Shen, Haihao and Cai, Yiyang and He, Xin and Lv, Kaokao and Liu, Yi}, journal={arXiv preprint arXiv:2309.05516}, year={2023} }

arxiv github

âš ī¸ Incomplete Data

Some information about this model is not available. Use with Caution - Verify details from the original source before relying on this data.

View Original Source →

📝 Limitations & Considerations

  • â€ĸ Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • â€ĸ VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • â€ĸ FNI scores are relative rankings and may change as new models are added.
  • ⚠ License Unknown: Verify licensing terms before commercial use.

Social Proof

HuggingFace Hub
125Downloads
🔄 Daily sync (03:00 UTC)

AI Summary: Based on Hugging Face metadata. Not a recommendation.

📊 FNI Methodology 📚 Knowledge Baseâ„šī¸ Verify with original source

đŸ›Ąī¸ Model Transparency Report

Technical metadata sourced from upstream repositories.

Open Metadata

🆔 Identity & Source

id
hf-model--incmodel--kimi-k2.5-mxfp4-autoround
slug
incmodel--kimi-k2.5-mxfp4-autoround
source
huggingface
author
INCModel
license
Other
tags
transformers, safetensors, kimi_k25, feature-extraction, auto-round, image-text-to-text, conversational, custom_code, arxiv:2309.05516, base_model:moonshotai/kimi-k2.5, base_model:quantized:moonshotai/kimi-k2.5, license:other, 8-bit, compressed-tensors, region:us

âš™ī¸ Technical Specs

architecture
null
params billions
null
context length
null
pipeline tag
image-text-to-text

📊 Engagement & Metrics

downloads
125
stars
0
forks
0

Data indexed from public sources. Updated daily.