Introducing Phi-3: Redefining what's possible with SLMs | Microsoft Azure Blog (2024)

We are excited to introduce Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks. This release expands the selection of high-quality models for customers, offering more practical choices as they compose and build generative AI applications.

Starting today, Phi-3-mini, a 3.8B language model is available on Microsoft Azure AI Studio, Hugging Face, and Ollama.

  • Phi-3-mini is available in two context-length variants—4K and 128K tokens. It is the first model in its class to support a context window of up to 128K tokens, with little impact on quality.
  • It is instruction-tuned, meaning that it’s trained to follow different types of instructions reflecting how people normally communicate. This ensures the model is ready to use out-of-the-box.
  • It is available on Azure AI to take advantage of the deploy-eval-finetune toolchain, and is available on Ollama for developers to run locally on their laptops.
  • It has been optimized for ONNX Runtime with support for Windows DirectML along with cross-platform support across graphics processing unit (GPU), CPU, and even mobile hardware.
  • It is also available as an NVIDIA NIM microservice with a standard API interface that can be deployed anywhere. And has been optimized for NVIDIA GPUs.

In the coming weeks, additional models will be added to Phi-3 family to offer customers even more flexibility across the quality-cost curve. Phi-3-small (7B) and Phi-3-medium (14B) will be available in the Azure AI model catalog and other model gardens shortly.  

Microsoft continues to offer the best models across the quality-cost curve and today’s Phi-3 release expands the selection of models with state-of-the-art small models.

Groundbreaking performance at a small size

Phi-3 models significantly outperform language models of the same and larger sizes on key benchmarks (see benchmark numbers below, higher is better). Phi-3-mini does better than models twice its size, and Phi-3-small and Phi-3-medium outperform much larger models, including GPT-3.5T.

All reported numbers are produced with the same pipeline to ensure that the numbers are comparable. As a result, these numbers may differ from other published numbers due to slight differences in the evaluation methodology. More details on benchmarks are provided in our technical paper.

Note: Phi-3 models do not perform as well on factual knowledge benchmarks (such as TriviaQA) as the smaller model size results in less capacity to retain facts.

Introducing Phi-3: Redefining what's possible with SLMs | Microsoft Azure Blog (2)

Safety-first model design

Responsible ai principles

Learn about our approach

Phi-3 models were developed in accordance with the Microsoft Responsible AI Standard, which is a company-wide set of requirements based on the following six principles: accountability, transparency, fairness, reliability and safety, privacy and security, and inclusiveness. Phi-3 models underwent rigorous safety measurement and evaluation, red-teaming, sensitive use review, and adherence to security guidanceto help ensure that these models are responsibly developed, tested, and deployed in alignment with Microsoft’s standards and best practices.

Building on our prior work with Phi models (“Textbooks Are All You Need”), Phi-3 models are also trained using high-quality data. They were further improved with extensive safety post-training, including reinforcement learning from human feedback (RLHF), automated testing and evaluations across dozens of harm categories, and manual red-teaming. Our approach to safety training and evaluations are detailed in our technical paper, and we outline recommended uses and limitations in the model cards. See the model card collection.

Unlocking new capabilities

Microsoft’s experience shipping copilots and enabling customers to transform their businesses with generative AI using Azure AI has highlighted the growing need for different-size models across the quality-cost curve for different tasks. Small language models, like Phi-3, are especially great for:

  • Resource constrained environments including on-device and offline inference scenarios.
  • Latency bound scenarios where fast response times are critical.
  • Cost constrained use cases, particularly those with simpler tasks.

For more on small language models, see our Microsoft Source Blog.

Thanks to their smaller size, Phi-3 models can be used in compute-limited inference environments. Phi-3-mini, in particular, can be used on-device, especially when further optimized with ONNX Runtime for cross-platform availability. The smaller size of Phi-3 models also makes fine-tuning or customization easier and more affordable. In addition, their lower computational needs make them a lower cost option with much better latency. The longer context window enables taking in and reasoning over large text content—documents, web pages, code, and more. Phi-3-mini demonstrates strong reasoning and logic capabilities, making it a good candidate for analytical tasks.

Customers are already building solutions with Phi-3. One example where Phi-3 is already demonstrating value is in agriculture, where internet might not be readily accessible. Powerful small models like Phi-3 along with Microsoft copilot templates are available to farmers at the point of need and provide the additional benefit of running at reduced cost, making AI technologies even more accessible.

ITC, a leading business conglomerate based in India, is leveraging Phi-3 as part of their continued collaboration with Microsoft on the copilot for Krishi Mitra, a farmer-facing app that reaches over a million farmers.

Our goal with the Krishi Mitra copilot is to improve efficiency while maintaining the accuracy of a large language model. We are excited to partner with Microsoft on using fine-tuned versions of Phi-3 to meet both our goals—efficiency and accuracy!

Saif Naik, Head of Technology, ITCMAARS

Originating in Microsoft Research, Phi models have been broadly used, with Phi-2 downloaded over 2 million times. The Phi series of models have achieved remarkable performance with strategic data curation and innovative scaling. Starting with Phi-1, a model used for Python coding, to Phi-1.5, enhancing reasoning and understanding, and then to Phi-2, a 2.7 billion-parameter model outperforming those up to 25 times its size in language comprehension.1 Each iteration has leveraged high-quality training data and knowledge transfer techniques to challenge conventional scaling laws.

Get started today

To experience Phi-3 for yourself, start with playing with the model on Azure AI Playground. You can also find the model on the Hugging Chat playground. Start building with and customizing Phi-3 for your scenarios using the Azure AI Studio. Join us to learn more about Phi-3 during a special live stream of the AI Show. 

1 Microsoft Research Blog, Phi-2: The surprising power of small language models, December 12, 2023.

Introducing Phi-3: Redefining what's possible with SLMs | Microsoft Azure Blog (2024)

FAQs

What is Microsoft Phi 3? ›

Introducing Phi-3: Redefining What's Possible with SLMs

Microsoft announced the introduction of Phi-3, a revolutionary new architecture that leverages the power of SLMS. Phi-3 redefines what's possible with language models, enabling faster and more accurate text processing.

What can I do with Azure? ›

Use cases
  1. Application development. Build, manage, and continuously deliver cloud apps—with any platform or language.
  2. AI. Analyze images, comprehend speech, and make predictions using data.
  3. Cloud migration and modernization. ...
  4. Data and analytics. ...
  5. Hybrid cloud and infrastructure. ...
  6. Internet of Things. ...
  7. Security and governance.

How many Azure services are there? ›

The Azure cloud platform is more than 200 products and cloud services designed to help you bring new solutions to life—to solve today's challenges and create the future.

Is Microsoft Azure free? ›

The Azure free account includes certain types of specific services—and certain amounts of those services—for free. To enable your production scenarios, you may need to use resources beyond the free amounts. If you choose to move to pay as you go, you'll be billed for those additional resources at pay-as-you-go rates.

What is E3 in Azure? ›

Endostar E3 Azure is an innovative system created with Azure HT Technology which is our specially designed heat-treatment manufacturing process. The goal of Azure HT Technology was to create files that are extremely flexible and resistant to breaking even in the most complicated clinical cases.

What does PHI stand for computers? ›

The definition of PHI is protected health information. It is a subset of PII that is protected by the HIPAA Privacy Act of 1996.

What is the main purpose of Azure? ›

The Azure cloud platform is more than 200 products and cloud services designed to help you bring new solutions to life—to solve today's challenges and create the future. Build, run, and manage applications across multiple clouds, on-premises, and at the edge, with the tools and frameworks of your choice.

What is the best thing about Azure? ›

Azure is the only consistent hybrid cloud, delivers unparalleled developer productivity, provides comprehensive, multilayered security, including the largest compliance coverage of any cloud provider, and you'll pay less for Azure as AWS is five times more expensive than Azure for Windows Server and SQL Server.

What is the benefit of using Azure? ›

One of the biggest benefits of Azure is its cost savings. Migrating to Microsoft Azure can lead to significant cost savings for businesses. With Azure, there are no upfront costs for hardware or infrastructure, and organizations only pay for the services they use.

What is the most used service on Azure? ›

What Are the Top 10 Most Used Microsoft Azure Services?
  • Azure DevOps (Development Operations)
  • Azure Virtual Machines.
  • Azure Blob Storage.
  • Azure AD (Active Directory)
  • Azure Cosmos DB (Database)
  • Logic Apps.
  • Azure Data Factory.
  • Azure CDN (Content Delivery Network)
Apr 6, 2023

What is the new name for Azure? ›

On July 11, 2023, Microsoft announced the renaming of Azure AD to Microsoft Entra ID to improve consistency with other Microsoft cloud products. The name change took place on July 15, 2023.

What is Azure in simple terms? ›

Azure is a huge collection of servers and networking hardware, which runs a complex set of distributed applications. These applications orchestrate the configuration and operation of virtualized hardware and software on those servers. The orchestration of these servers is what makes Azure so powerful.

Is it hard to learn Azure? ›

If you are already AWS certified, you may be wondering if Azure is easy to learn? The short answer is yes, Azure is easy to learn if you already know AWS for many reasons: Basic architecture concepts are the same across all the cloud vendors. Cloud native concepts (IaaS, PaaS, SaaS) are the same across all cloud ...

Why Azure is so expensive? ›

Scale and performance needs: Higher-tier services for better performance and increased scalability come with higher costs. For businesses requiring high-performance computing, large-scale storage, or extensive data processing capabilities, Azure can be costly.

Can I use Azure for free forever? ›

We'll never charge you unless you decide to move to pay-as-you-go pricing. If you move to pay as you go, you'll only be charged for services that you use above the free monthly amounts. You can check your usage of free services in the Azure portal.

What is the Microsoft ChatGPT called? ›

Copilot works just like ChatGPT. You can ask it anything, and using an LLM (large language model), it will understand and respond in natural human language. But Microsoft has also been slowly rolling out Copilot features, and most of them are available to use now.

What's included in MS E3? ›

The E3 suite encompasses Office 365 ProPlus Word, Excel, PowerPoint, OneNote, Publisher, co-authoring, Outlook Exchange Email, Skype for Business, Instant Messaging, voicemail, video dial-in conferencing, SharePoint collaboration and Teams, OneDrive, Yammer, enterprise social networking, Encryption, Data Loss ...

What is Microsoft's version of ChatGPT? ›

ChatGPT and Microsoft Copilot are both artificial intelligence (AI) technologies that were developed with the intent of helping you accomplish tasks and activities faster and more efficiently. While they may seem similar, there are significant differences between the two.

What is the difference between Microsoft A1 and E3? ›

Microsoft 365 E3 includes features like Team & Productivity, Identity & Access Management, Compliance and Security. Office 365 A1 for Faculty is known for functionalities like Outlook, Word, Excel, PowerPoint, Collaboration Tools, Messaging & Video Conferencing and Cloud-Based.

Top Articles
Latest Posts
Article information

Author: Dan Stracke

Last Updated:

Views: 5934

Rating: 4.2 / 5 (43 voted)

Reviews: 90% of readers found this page helpful

Author information

Name: Dan Stracke

Birthday: 1992-08-25

Address: 2253 Brown Springs, East Alla, OH 38634-0309

Phone: +398735162064

Job: Investor Government Associate

Hobby: Shopping, LARPing, Scrapbooking, Surfing, Slacklining, Dance, Glassblowing

Introduction: My name is Dan Stracke, I am a homely, gleaming, glamorous, inquisitive, homely, gorgeous, light person who loves writing and wants to share my knowledge and understanding with you.