LATEST NEWS :
FREE Orientation @ Kapoorthala Branch on 30th April 2024 , FREE Workshop @ Kanpur Branch on 29th April , New Batch of Modern History W.e.f. 01.05.2024 , Indian Economy @ Kanpur Branch w.e.f. 25.04.2024. Interested Candidates may join these workshops and batches .
Print Friendly and PDF

Phi-3-mini

27.04.2024

 

Phi-3-mini

 

For Prelims: About Phi-3-mini, How is Phi-3-mini different from Large Language Models (LLMs)?

 

Why in the news?                  

           Recently, Microsoft unveiled the latest version of its ‘lightweight’ AI model, the Phi-3-Mini.

 

About Phi-3-mini:

  • It is believed to be the first among the three small models that Microsoft is planning to release.
  • It has reportedly outperformed models of the same size and the next size up across a variety of benchmarks, in areas like language, reasoning, coding, and maths.
  • It is the first model in its class to support a context window of up to 128K tokens, with little impact on quality.
  • The amount of conversation that an AI can read and write at any given time is called the context window, and is measured in tokens.
  • It is a 3.8B language model and is available on AI development platforms such as Microsoft Azure AI Studio, HuggingFace, and Ollama.
  • Phi-3-mini is available in two variants, one with 4K context length, and another with 128K tokens.

 

How is Phi-3-mini different from Large Language Models (LLMs)?

  • Phi-3-mini is a Small Language Model (SLM).
  • SLMs are more streamlined versions of large language models. When compared to LLMs, smaller AI models are also cost-effective to develop and operate, and they perform better on smaller devices like laptops and smartphones.
  • SLMs are great for resource-constrained environments including on-device and offline inference scenarios and such models are good for scenarios where fast response times are critical, say for chatbots or virtual assistants.
  • SLMs can be customized for specific tasks and achieve accuracy and efficiency in doing them. Most SLMs undergo targeted training, demanding considerably less computing power and energy compared to LLMs.
  • SLMs also differ when it comes to inference speed and latency. Their compact size allows for quicker processing. Their cost makes them appealing to smaller organizations and research groups.

                                                                      Source: Indian Express