Microsoft unveiled Phi-3-mini, a compact AI model, that runs on your phone. Designed for simplicity, it excels in language, coding, and math tasks, and shows impressive benchmark results to highly rivals models like Mixtral 8x7B and GPT-3.5.
SpreadsheetLLM: Microsoft’s Latest AI Innovation for Smart Data Management
Microsoft has launched phi-3-mini, a 3.8 billion parameter language model trained on 3.3 trillion tokens, whose overall performance, as measured by both academic benchmarks and internal testing, rivals that of models such as Mixtral 8x7B and GPT-3.5.
According to the technical report released by Microsoft, the Phi-3 model, despite being small enough to be deployed on a phone,. The model is also further aligned for robustness, safety, and chat format.
Microsoft has also introduced Phi-3-Small and Phi-3-Medium models, both significantly more capable than Phi-3-Mini. Phi-3-Small, with 7 billion parameters, utilizes the tiktoken tokenizer for improved multilingual tokenization. It boasts a vocabulary size of 100,352 and a default context length of 8K.
Also Read: Microsoft AI Video App Makes Mona Lisa Painting Sing; Video Gets Over 7 Million Views
Also Read: Microsoft’s $2.9B Boost for Japan’s AI and Cloud Capabilities
Phi-3-mini was developed following Microsoft’s responsible AI principles. The overall approach consisted of safety alignment in post-training, red-teaming, automated testing, and evaluations across dozens of RAI harm categories. Helpfulness and harmlessness preference datasets, along with modifications and multiple in-house generated datasets, were leveraged to address the RAI harm categories in safety post-training.
An independent red team at Microsoft iteratively examined phi-3-mini further to identify areas of improvement during the post-training process. Based on their feedback, Microsoft curated additional datasets tailored to address their insights, thereby refining the post-training dataset. This process resulted in a significant decrease in harmful response rates
Also Read: Mustafa Suleyman announces new Microsoft AI hub in London
Phi-3 mini demonstrates a similar level of language understanding and reasoning ability as much larger models, but its size fundamentally limits it for certain tasks. For example, it cannot store extensive “factual knowledge,” resulting in lower performance on tasks such as TriviaQA.
Microsoft believes such weaknesses can be addressed by augmenting the model with a search engine. Additionally, the model’s language capabilities are mostly restricted to English, highlighting the need to explore multilingual capabilities for Small Language Models.
As per the reports, Microsoft said, Phi-3-mini will be available immediately on Microsoft cloud service platform Azure’s AI model catalog, machine learning model platform Hugging Face, and Ollama, a framework for running models on a local machine.
This post was last modified on April 23, 2024 5:46 am
Perplexity AI Voice Assistant is a smart tool for Android devices that lets users perform…
Meta AI is a personal voice assistant app powered by Llama 4. It offers smart,…
On April 23, 2025, current President Donald J. Trump signed an executive order to advance…
Google is launching The Android Show: I/O Edition, featuring Android ecosystem president Sameer Samat, to…
The top 11 generative AI companies in the world are listed below. These companies have…
Google has integrated Veo 2 video generation into the Gemini app for Advanced subscribers, enabling…