32B long-context instruct model with RL alignment, IF, tool use, and enterprise optimization.
5.3K
Granite-4.0-H-Small is a 32B parameter long-context instruct model finetuned from Granite-4.0-H-Small-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets. This model is developed using a diverse set of techniques with a structured chat format, including supervised finetuning, model alignment using reinforcement learning, and model merging. Granite 4.0 instruct models feature improved instruction following (IF) and tool-calling capabilities, making them more effective in enterprise applications.
| Attribute | Details |
|---|---|
| Provider | Granite Team, IBM |
| Architecture | granitehybrid |
| Cutoff date | Not disclosed |
| Languages | English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, Chinese (extensible via finetuning) |
| Tool calling | ✅ |
| Input modalities | Text |
| Output modalities | Text |
| License | Apache 2.0 |
| Model variant | Parameters | Quantization | Context window | VRAM¹ | Size |
|---|---|---|---|---|---|
ai/granite-4.0-h-small:32Bai/granite-4.0-h-small:32B-Q4_K_Mai/granite-4.0-h-small:latest | 32.21 B | MOSTLY_Q4_K_M | 1M tokens | 18.80 GiB | 18.14 GB |
¹: VRAM estimated based on model characteristics.
latest→32B
docker model run ai/granite-4.0-h-small
| Category | Metric | Granite-4.0-h-Small |
|---|---|---|
| General Tasks | ||
| MMLU (5-shot) | 78.44 | |
| MMLU-Pro (5-shot, CoT) | 55.47 | |
| BBH (3-shot, CoT) | 81.62 | |
| AGI EVAL (0-shot, CoT) | 70.63 | |
| GPQA (0-shot, CoT) | 40.63 | |
| Alignment Tasks | ||
| AlpacaEval 2.0 | 42.48 | |
| IFEval (Instruct, Strict) | 89.87 | |
| IFEval (Prompt, Strict) | 85.22 | |
| IFEval (Average) | 87.55 | |
| ArenaHard | 46.48 | |
| Math Tasks | ||
| GSM8K (8-shot) | 87.27 | |
| GSM8K Symbolic (8-shot) | 87.38 | |
| Minerva Math (0-shot, CoT) | 74.00 | |
| DeepMind Math (0-shot, CoT) | 59.33 | |
| Code Tasks | ||
| HumanEval (pass@1) | 88.00 | |
| HumanEval+ (pass@1) | 83.00 | |
| MBPP (pass@1) | 84.00 | |
| MBPP+ (pass@1) | 71.00 | |
| CRUXEval-O (pass@1) | 50.25 | |
| BigCodeBench (pass@1) | 46.23 | |
| Tool Calling Tasks | ||
| BFCL v3 | 64.69 | |
| Multilingual Tasks | ||
| MULTIPLE (pass@1) | 57.37 | |
| MMMLU (5-shot) | 69.69 | |
| INCLUDE (5-shot) | 63.97 | |
| MGSM (8-shot) | 38.72 | |
| Safety | ||
| SALAD-Bench | 97.30 | |
| AttaQ | 86.64 |
Content type
Model
Digest
sha256:7a1eb06e9…
Size
18.1 GB
Last updated
4 months ago
docker model pull ai/granite-4.0-h-small:32BPulls:
71
Last week