Posts

#4180 Granite-4.0-H-Small, 32-Billion/9-Billion Activated Parameters; +Discussion Board & Genetic Programming

#4180 Granite-4.0-H-Small, 32-Billion/9-Billion Activated Parameters; +Discussion Board & Genetic Programming See also #4178   Event-Driven AI Agent Discussion Board #4180  Tell me about the capabilities of Granite-4.0-H-Small, 32-Billion/9-Billion activated parameters Released by IBM in October 2025, the Granite-4.0-H-Small is a flagship "workhorse" model designed for enterprise efficiency. It utilizes a novel Hybrid Mamba-Transformer architecture and a Mixture-of-Experts (MoE) strategy to achieve high performance with a significantly reduced memory footprint. Core Architecture & Parameters The model distinguishes itself by decoupling its total capacity from its active compute: Total Parameters: 32 Billion. Activated Parameters: 9 Billion (only these are used per token during inference). Hybrid Design: It combines Mamba-2 (State Space Model) layers with selective Transformer attention layers in a roughly 9:1 ratio. Efficiency: This architecture results ...