Bookkeeping Service Providers

  • Accounting
  • Bookkeeping
  • US Taxation
  • Financial Planning
  • Accounting Software
  • Small Business Finance
You are here: Home / CLOUD / Announcing new fine-tuning models and techniques in Azure AI Foundry

Announcing new fine-tuning models and techniques in Azure AI Foundry

May 12, 2025 by cbn Leave a Comment

Today, we’re excited to announce three major enhancements to model fine-tuning in Azure AI Foundry—Reinforcement Fine-Tuning (RFT) with o4-mini (coming soon), Supervised Fine-Tuning (SFT) for the GPT-4.1-nano and Llama 4 Scout model (available now). These updates reflect our continued commitment to empowering organizations with tools to build highly customized, domain-adapted AI systems for real-world impact. 

With these new models, we’re unblocking two major avenues of LLM customization: GPT-4.1-nano is a powerful small model, ideal for distillation, while o4-mini is the first reasoning model you can fine-tune, and Llama 4 Scout is a best-in-class open source model. 

Design, customize, and manage AI apps with Azure today

Reinforcement Fine-Tuning with o4-mini 

Reinforcement Fine-Tuning introduces a new level of control for aligning model behavior with complex business logic. By rewarding accurate reasoning and penalizing undesirable outputs, RFT improves model decision-making in dynamic or high-stakes environments.

Coming soon for the o4-mini model, RFT unlocks new possibilities for use cases requiring adaptive reasoning, contextual awareness, and domain-specific logic—all while maintaining fast inference performance.

Real world impact: DraftWise 

DraftWise, a legal tech startup, used reinforcement fine-tuning (RFT) in Azure AI Foundry Models to enhance the performance of reasoning models tailored for contract generation and review. Faced with the challenge of delivering highly contextual, legally sound suggestions to lawyers, DraftWise fine-tuned Azure OpenAI models using proprietary legal data to improve response accuracy and adapt to nuanced user prompts. This led to a 30% improvement in search result quality, enabling lawyers to draft contracts faster and focus on high-value advisory work. 

Reinforcement fine-tuning on reasoning models is a potential game changer for us. It’s helping our models understand the nuance of legal language and respond more intelligently to complex drafting instructions, which promises to make our product significantly more useful to lawyers in real time.

—James Ding, founder and CEO of DraftWise.

When should you use Reinforcement Fine-Tuning?

Reinforcement Fine-Tuning is best suited for use cases where adaptability, iterative learning, and domain-specific behavior are essential. You should consider RFT if your scenario involves: 

  1. Custom Rule Implementation: RFT thrives in environments where decision logic is highly specific to your organization and cannot be easily captured through static prompts or traditional training data. It enables models to learn flexible, evolving rules that reflect real-world complexity. 
  1. Domain-Specific Operational Standards: Ideal for scenarios where internal procedures diverge from industry norms—and where success depends on adhering to those bespoke standards. RFT can effectively encode procedural variations, such as extended timelines or modified compliance thresholds, into the model’s behavior. 
  1. High Decision-Making Complexity: RFT excels in domains with layered logic and variable-rich decision trees. When outcomes depend on navigating numerous subcases or dynamically weighing multiple inputs, RFT helps models generalize across complexity and deliver more consistent, accurate decisions. 

Example: Wealth advisory at Contoso Wellness 

To showcase the potential of RFT, consider Contoso Wellness, a fictitious wealth advisory firm. Using RFT, the o4-mini model learned to adapt to unique business rules, such as identifying optimal client interactions based on nuanced patterns like the ratio of a client’s net worth to available funds. This enabled Contoso to streamline their onboarding processes and make more informed decisions faster.

Supervised Fine-Tuning now available for GPT-4.1-nano 

We’re also bringing Supervised Fine-Tuning (SFT) to the GPT-4.1-nano model—a small but powerful foundation model optimized for high-throughput, cost-sensitive workloads. With SFT, you can instill your model with company-specific tone, terminology, workflows, and structured outputs—all tailored to your domain. This model will be available for fine-tuning in the coming days. 

Why Fine-tune GPT-4.1-nano? 

  • Precision at Scale: Tailor the model’s responses while maintaining speed and efficiency. 
  • Enterprise-Grade Output: Ensure alignment with business processes and tone-of-voice. 
  • Lightweight and Deployable: Perfect for scenarios where latency and cost matter—such as customer service bots, on-device processing, or high-volume document parsing. 

Compared to larger models, 4.1-nano delivers faster inference and lower compute costs, making it well suited for large-scale workloads like: 

  • Customer support automation, where models must handle thousands of tickets per hour with consistent tone and accuracy. 
  • Internal knowledge assistants that follow company style and protocol in summarizing documentation or responding to FAQs. 

As a small, fast, but highly capable model, GPT-4.1-nano makes a great candidate for distillation as well. You can use models like GPT-4.1 or o4 to generate training data—or capture production traffic with stored completions—and teach 4.1-nano to be just as smart!

Fine-tune gpt-4.1-nano demo in Azure AI Foundry.

Llama 4 Fine-Tuning now available 

We’re also excited to announce support for fine-tuning Meta’s Llama 4 Scout—a cutting edge,17 billion active parameter model which offers an industry leading context window of 10M tokens while fitting on a single H100 GPU for inferencing. It’s a best-in-class model, and more powerful than all previous generation llama models. 

Llama 4 fine-tuning is available in our managed compute offering, allowing you to fine-tune and inference using your own GPU quota. Available in both Azure AI Foundry and as Azure Machine Learning components, you have access to additional hyperparameters for deeper customization compared to our serverless experience.

Get started with Azure AI Foundry today

Azure AI Foundry is your foundation for enterprise-grade AI tuning. These fine-tuning enhancements unlock new frontiers in model customization, helping you build intelligent systems that think and respond in ways that reflect your business DNA.

  • Use Reinforcement Fine-tuning with o4-mini to build reasoning engines that learn from experience and evolve over time. Coming soon in Azure AI Foundry, with regional availability for East US2 and Sweden Central. 
  • Use Supervised Fine-Tuning with 4.1-nano to scale reliable, cost-efficient, and highly customized model behaviors across your organization. Available now in Azure AI Foundry in North Central US and Sweden Central. 
  • Try Llama 4 scout fine tuning to customize a best-in-class open source model. Available now in Azure AI Foundry model catalog and Azure Machine Learning. 

With Azure AI Foundry, fine-tuning isn’t just about accuracy—it’s about trust, efficiency, and adaptability at every layer of your stack. 

Explore further: 

  • Get started with Azure AI Foundry.
  • Documentation on fine-tuning in Azure AI Foundry.

We’re just getting started. Stay tuned for more model support, advanced tuning techniques, and tools to help you build AI that’s smarter, safer, and uniquely yours. 

A colorful background with a white line

Azure AI services

Build cutting-edge, market-ready AI applications with out-of-the-box and customizable APIs and models

Get started with Azure >

The post Announcing new fine-tuning models and techniques in Azure AI Foundry appeared first on Microsoft Azure Blog.

Share on FacebookShare on TwitterShare on Google+Share on LinkedinShare on Pinterest

Filed Under: CLOUD

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Archives

  • May 2025
  • April 2025
  • March 2025
  • February 2025
  • January 2025
  • December 2024
  • November 2024
  • October 2024
  • July 2024
  • June 2024
  • May 2024
  • April 2024
  • March 2024
  • February 2024
  • January 2024
  • December 2023
  • October 2023
  • September 2023
  • August 2023
  • July 2023
  • June 2023
  • May 2023
  • April 2023
  • March 2023
  • February 2023
  • January 2023
  • December 2022
  • November 2022
  • October 2022
  • September 2022
  • August 2022
  • July 2022
  • June 2022
  • May 2022
  • April 2022
  • March 2022
  • February 2022
  • January 2022
  • December 2021
  • November 2021
  • October 2021
  • September 2021
  • August 2021
  • May 2021
  • April 2021
  • September 2020
  • August 2020
  • July 2020
  • June 2020
  • May 2020
  • April 2020
  • March 2020
  • February 2020
  • January 2020
  • December 2019
  • November 2019
  • October 2019
  • September 2019
  • August 2019
  • July 2019
  • June 2019
  • May 2019
  • April 2019
  • March 2019
  • February 2019
  • January 2019
  • December 2018
  • November 2018
  • October 2018
  • September 2018
  • August 2018
  • July 2018
  • June 2018
  • May 2018
  • April 2018
  • March 2018
  • February 2018
  • January 2018
  • December 2017
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • July 2017
  • May 2017
  • April 2017
  • March 2017
  • February 2017
  • January 2017
  • March 2016

Recent Posts

  • Agentic DevOps: Evolving software development with GitHub Copilot and Microsoft Azure
  • Powering the next AI frontier with Microsoft Fabric and the Azure data portfolio 
  • Azure AI Foundry: Your AI App and agent factory
  • Announcing new fine-tuning models and techniques in Azure AI Foundry
  • Key network security takeaways from RSAC 2025

Recent Comments

    Categories

    • Accounting
    • Accounting Software
    • BlockChain
    • Bookkeeping
    • CLOUD
    • Data Center
    • Financial Planning
    • IOT
    • Machine Learning & AI
    • SECURITY
    • Uncategorized
    • US Taxation

    Categories

    • Accounting (145)
    • Accounting Software (27)
    • BlockChain (18)
    • Bookkeeping (205)
    • CLOUD (1,310)
    • Data Center (214)
    • Financial Planning (345)
    • IOT (260)
    • Machine Learning & AI (41)
    • SECURITY (611)
    • Uncategorized (1,284)
    • US Taxation (17)

    Subscribe Our Newsletter

     Subscribing I accept the privacy rules of this site

    Copyright © 2025 · News Pro Theme on Genesis Framework · WordPress · Log in