| Model | Min Training Examples | Typical Use |
|---|---|---|
| o4-mini | 10 | Reasoning-focused customization (New in 2026) |
| GPT-4o / GPT-5.4 | 10 | High-quality custom behavior |
| GPT-4o mini | 10 | Cost-effective custom models |
As of April 2026, Foundry supports Global Training for models like o4-mini. This allows you to launch fine-tuning jobs across 13+ Azure regions, offering lower per-token training rates compared to standard regional training.
For reasoning models (o-series), Foundry provides Reinforcement Fine-Tuning (RFT). Unlike Supervised Fine-Tuning (which teaches formatting or style), RFT aligns model behavior with complex business logic by explicitly rewarding accurate reasoning paths.
{"messages": [
{"role": "system", "content": "You are a legal contract analyzer."},
{"role": "user", "content": "Analyze this NDA clause: ..."},
{"role": "assistant", "content": "Risk Level: Medium. Key concerns: ..."}
]}