Chinese long context models

Models that keep accuracy at 128K+ tokens — Kimi, GLM-Long, Qwen2.5-1M. Useful for long-document Q&A, whole-repo code review, or multi-hour transcripts where chunking would lose the through-line.

ModelCreatorContextOpen weightTags
Qwen3 72B InstructAlibaba131KYeschat · reasoning · multilingual
DeepSeek V3DeepSeek AI66KYesreasoning · coding · tool_use
Llama 3.3 70B InstructMeta131KYeschat · coding · reasoning

Other capabilities