Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at a fraction of the compute cost. Supports multimodal input including…
AI Models
Gemma 4 31B Instruct is Google DeepMind's 30.7B dense multimodal model supporting text and image input with text output. Features a 256K token context window, configurable thinking/reasoning mode, native function calling, and multilingual support across 140+ languages. Strong on coding,……
Qwen 3.6 Plus builds on a hybrid architecture that combines efficient linear attention with sparse mixture-of-experts routing, enabling strong scalability and high-performance inference. 相比于 3.5 系列, it delivers major gains in agentic coding, front-end development, and overall reasoning,……
GLM-5V-Turbo is Z.ai’s first native multimodal agent foundation model, built for vision-based coding and agent-driven tasks. It natively handles image, 视频, and text inputs, excels at long-horizon planning, complex coding, and task execution, and works seamlessly with agents to complete…
Trinity Large Thinking is a powerful open source reasoning model from the team at Arcee AI. It shows strong performance in PinchBench, agentic workloads, and reasoning tasks. It is free in open claw for the first five days. Launch video:……
格罗克 4.20 Multi-Agent is a variant of xAI’s Grok 4.20 designed for collaborative, agent-based workflows. Multiple agents operate in parallel to conduct deep research, coordinate tool use, and synthesize information across complex tasks. Reasoning effort behavior: - low / medium:……
格罗克 4.20 is xAI's newest flagship model with industry-leading speed and agentic tool calling capabilities. It combines the lowest hallucination rate on the market with strict prompt adherance, delivering consistently precise and truthful responses. Reasoning can be enabled/disabled using the…
Full-length songs are priced at $0.08 per song. Lyria 3 is Google's family of music generation models, available through the Gemini API. With Lyria 3, you can generate high-quality, 48kHz stereo audio from text prompts or from images. These models…
30 second duration clips are priced at $0.04 per clip. Lyria 3 is Google's family of music generation models, available through the Gemini API. With Lyria 3, you can generate high-quality, 48来自文本提示或图像的 kHz 立体声音频...
KAT-Coder-Pro V2 is the latest high-performance model in KwaiKAT’s KAT-Coder series, designed for complex enterprise-grade software engineering and SaaS integration. It builds on the agentic coding strengths of earlier versions, 专注于大规模生产环境, 多系统协调, 和…
Reka Edge is an extremely efficient 7B multimodal vision-language model that accepts image/video+text inputs and generates text outputs. This model is optimized specifically to deliver industry-leading performance in image understanding, video analysis, object detection, and agentic tool-use.
MiMo-V2-Omni is a frontier omni-modal model that natively processes image, 视频, and audio inputs within a unified architecture. It combines strong multimodal perception with agentic capability - visual grounding, multi-step planning, 工具使用, and code execution - making it well-suited…
MiMo-V2-Pro is Xiaomi's flagship foundation model, featuring over 1T total parameters and a 1M context length, deeply optimized for agentic scenarios. It is highly adaptable to general agent frameworks like OpenClaw. 它在……领域跻身全球顶尖行列。
MiniMax-M2.7是专为自主设计的下一代大型语言模型, 现实世界的生产力和持续改进. 旨在积极参与自身的发展, M2.7通过多智能体协作集成先进的智能体能力, 使其能够计划, 执行, 并完善复杂的任务……
GPT-5.4 nano 是 GPT-5.4 系列中最轻量且最具成本效益的变体, 针对速度关键和大批量任务进行了优化. 它支持文本和图像输入,专为低延迟用例(例如分类)而设计, 数据提取, 排行, 和分代理...
GPT-5.4 mini将GPT-5.4的核心功能以更快的速度, 针对高吞吐量工作负载优化的更高效模型. 支持文本和图像输入,跨推理性能强劲, 编码, 和工具的使用, 同时减少大规模的延迟和成本……
小米斯特拉尔 4 是 Mistral Small 系列的下一个主要版本, 将多个旗舰 Mistral 型号的功能统一到一个系统中. 它结合了裁判官的强有力的推理, Pixtral 的多模态理解, 和代理编码能力来自……
GLM-5 Turbo 是 Z.ai 的新模型,专为在 OpenClaw 场景等代理驱动环境中实现快速推理和强大性能而设计. 它针对涉及长执行链的现实世界代理工作流程进行了深度优化, 改进了复杂指令分解, tool…
NVIDIA 神经元 3 Super是120B参数开放式混合动力MoE模型, 仅激活 12B 参数即可在复杂的多代理应用程序中实现最大计算效率和准确性. 基于具有多令牌预测的混合 Mamba-Transformer Mixture-of-Experts 架构构建 (MTP), it delivers over 50%…
NVIDIA 神经元 3 Super是120B参数开放式混合动力MoE模型, 仅激活 12B 参数即可在复杂的多代理应用程序中实现最大计算效率和准确性. 基于具有多令牌预测的混合 Mamba-Transformer Mixture-of-Experts 架构构建 (MTP), it delivers over 50%…







