๐Ÿง 

glm-4.6

by zai-org Model ID: hf-model--zai-org--glm-4.6
FNI 9.7
Top 93%

"๐Ÿ‘‹ Join our Discord community. ๐Ÿ“– Check out the GLM-4.6 technical blog,..."

๐Ÿ”— View Source
Audited 9.7 FNI Score
Massive 356.79B Params
4k Context
Hot 332.6K Downloads
H100+ ~271GB Est. VRAM

โšก Quick Commands

๐Ÿค— HF Download
huggingface-cli download zai-org/glm-4.6
๐Ÿ“ฆ Install Lib
pip install -U transformers
๐Ÿ“Š

Engineering Specs

โšก Hardware

Parameters
356.79B
Architecture
Glm4MoeForCausalLM
Context Length
4K
Model Size
664.6GB

๐Ÿง  Lifecycle

Library
-
Precision
float16
Tokenizer
-

๐ŸŒ Identity

Source
HuggingFace
License
Open Access
๐Ÿ’พ

Est. VRAM Benchmark

~270.1GB

Analyze Hardware

* Technical estimation for FP16/Q4 weights. Does not include OS overhead or long-context batching. For Technical Reference Only.

๐Ÿ•ธ๏ธ Neural Mesh Hub

Interconnecting Research, Data & Ecosystem

๐Ÿ”ฌ

๐Ÿ”ฌ Research & Data

๐Ÿ“ˆ Interest Trend

--

* Real-time activity index across HuggingFace, GitHub and Research citations.

No similar models found.

๐Ÿ”ฌTechnical Deep Dive

Full Specifications [+]
---

๐Ÿš€ What's Next?

โšก Quick Commands

๐Ÿค— HF Download
huggingface-cli download zai-org/glm-4.6
๐Ÿ“ฆ Install Lib
pip install -U transformers
๐Ÿ–ฅ๏ธ

Hardware Compatibility

Multi-Tier Validation Matrix

Live Sync
๐ŸŽฎ Compatible

RTX 3060 / 4060 Ti

Entry 8GB VRAM
๐ŸŽฎ Compatible

RTX 4070 Super

Mid 12GB VRAM
๐Ÿ’ป Compatible

RTX 4080 / Mac M3

High 16GB VRAM
๐Ÿš€ Compatible

RTX 3090 / 4090

Pro 24GB VRAM
๐Ÿ—๏ธ Compatible

RTX 6000 Ada

Workstation 48GB VRAM
๐Ÿญ Compatible

A100 / H100

Datacenter 80GB VRAM
โ„น๏ธ

Pro Tip: Compatibility is estimated for 4-bit quantization (Q4). High-precision (FP16) or ultra-long context windows will significantly increase VRAM requirements.

README

GLM-4.6

๐Ÿ‘‹ Join our Discord community.
๐Ÿ“– Check out the GLM-4.6 technical blog, technical report(GLM-4.5), and Zhipu AI technical documentation.
๐Ÿ“ Use GLM-4.6 API services on Z.ai API Platform.
๐Ÿ‘‰ One click to GLM-4.6.

Model Introduction

Compared with GLM-4.5, GLM-4.6 brings several key improvements:

  • Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks.
  • Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Codeใ€Clineใ€Roo Code and Kilo Code, including improvements in generating visually polished front-end pages.
  • Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability.
  • More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks.
  • Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios.

We evaluated GLM-4.6 across eight public benchmarks covering agents, reasoning, and coding. Results show clear gains over GLM-4.5, with GLM-4.6 also holding competitive advantages over leading domestic and international models such as DeepSeek-V3.1-Terminus and Claude Sonnet 4.

bench

Inference

Both GLM-4.5 and GLM-4.6 use the same inference method.

you can check our github for more detail.

For general evaluations, we recommend using a sampling temperature of 1.0.

For code-related evaluation tasks (such as LCB), it is further recommended to set:

  • top_p = 0.95
  • top_k = 40

Evaluation

  • For tool-integrated reasoning, please refer to this doc.
  • For search benchmark, we design a specific format for searching toolcall in thinking mode to support search agent, please refer to this. for the detailed template.
ZEN MODE โ€ข README

GLM-4.6

๐Ÿ‘‹ Join our Discord community.
๐Ÿ“– Check out the GLM-4.6 technical blog, technical report(GLM-4.5), and Zhipu AI technical documentation.
๐Ÿ“ Use GLM-4.6 API services on Z.ai API Platform.
๐Ÿ‘‰ One click to GLM-4.6.

Model Introduction

Compared with GLM-4.5, GLM-4.6 brings several key improvements:

  • Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks.
  • Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Codeใ€Clineใ€Roo Code and Kilo Code, including improvements in generating visually polished front-end pages.
  • Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability.
  • More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks.
  • Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios.

We evaluated GLM-4.6 across eight public benchmarks covering agents, reasoning, and coding. Results show clear gains over GLM-4.5, with GLM-4.6 also holding competitive advantages over leading domestic and international models such as DeepSeek-V3.1-Terminus and Claude Sonnet 4.

bench

Inference

Both GLM-4.5 and GLM-4.6 use the same inference method.

you can check our github for more detail.

For general evaluations, we recommend using a sampling temperature of 1.0.

For code-related evaluation tasks (such as LCB), it is further recommended to set:

  • top_p = 0.95
  • top_k = 40

Evaluation

  • For tool-integrated reasoning, please refer to this doc.
  • For search benchmark, we design a specific format for searching toolcall in thinking mode to support search agent, please refer to this. for the detailed template.

๐Ÿ“ Limitations & Considerations

  • โ€ข Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • โ€ข VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • โ€ข FNI scores are relative rankings and may change as new models are added.
  • โš  License Unknown: Verify licensing terms before commercial use.
  • โ€ข Source: Unknown
๐Ÿ“œ

Cite this model

Academic & Research Attribution

BibTeX
@misc{hf_model__zai_org__glm_4.6,
  author = {zai-org},
  title = {undefined Model},
  year = {2026},
  howpublished = {\url{https://huggingface.co/zai-org/glm-4.6}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
zai-org. (2026). undefined [Model]. Free2AITools. https://huggingface.co/zai-org/glm-4.6
๐Ÿ”„ Daily sync (03:00 UTC)

AI Summary: Based on Hugging Face metadata. Not a recommendation.

๐Ÿ“Š FNI Methodology ๐Ÿ“š Knowledge Baseโ„น๏ธ Verify with original source

๐Ÿ›ก๏ธ Model Transparency Report

Verified data manifest for traceability and transparency.

100% Data Disclosure Active

๐Ÿ†” Identity & Source

id
hf-model--zai-org--glm-4.6
author
zai-org
tags
transformerssafetensorsglm4_moetext-generationconversationalenzharxiv:2508.06471license:mitendpoints_compatibleregion:us

โš™๏ธ Technical Specs

architecture
Glm4MoeForCausalLM
params billions
356.79
context length
4,096
vram gb
270.1
vram is estimated
true
vram formula
VRAM โ‰ˆ (params * 0.75) + 2GB (KV) + 0.5GB (OS)

๐Ÿ“Š Engagement & Metrics

likes
1,143
downloads
332,556

Free2AITools Constitutional Data Pipeline: Curated disclosure mode active. (V15.x Standard)