Microsoft unveiled Phi-3-mini, a compact AI model, that runs on your phone. Designed for simplicity, it excels in language, coding, and math tasks, and shows impressive benchmark results to highly rivals models like Mixtral 8x7B and GPT-3.5.
SpreadsheetLLM: Microsoft’s Latest AI Innovation for Smart Data Management
Microsoft has launched phi-3-mini, a 3.8 billion parameter language model trained on 3.3 trillion tokens, whose overall performance, as measured by both academic benchmarks and internal testing, rivals that of models such as Mixtral 8x7B and GPT-3.5.
According to the technical report released by Microsoft, the Phi-3 model, despite being small enough to be deployed on a phone,. The model is also further aligned for robustness, safety, and chat format.
Microsoft has also introduced Phi-3-Small and Phi-3-Medium models, both significantly more capable than Phi-3-Mini. Phi-3-Small, with 7 billion parameters, utilizes the tiktoken tokenizer for improved multilingual tokenization. It boasts a vocabulary size of 100,352 and a default context length of 8K.
Also Read: Microsoft AI Video App Makes Mona Lisa Painting Sing; Video Gets Over 7 Million Views
Also Read: Microsoft’s $2.9B Boost for Japan’s AI and Cloud Capabilities
Phi-3-mini was developed following Microsoft’s responsible AI principles. The overall approach consisted of safety alignment in post-training, red-teaming, automated testing, and evaluations across dozens of RAI harm categories. Helpfulness and harmlessness preference datasets, along with modifications and multiple in-house generated datasets, were leveraged to address the RAI harm categories in safety post-training.
An independent red team at Microsoft iteratively examined phi-3-mini further to identify areas of improvement during the post-training process. Based on their feedback, Microsoft curated additional datasets tailored to address their insights, thereby refining the post-training dataset. This process resulted in a significant decrease in harmful response rates
Also Read: Mustafa Suleyman announces new Microsoft AI hub in London
Phi-3 mini demonstrates a similar level of language understanding and reasoning ability as much larger models, but its size fundamentally limits it for certain tasks. For example, it cannot store extensive “factual knowledge,” resulting in lower performance on tasks such as TriviaQA.
Microsoft believes such weaknesses can be addressed by augmenting the model with a search engine. Additionally, the model’s language capabilities are mostly restricted to English, highlighting the need to explore multilingual capabilities for Small Language Models.
As per the reports, Microsoft said, Phi-3-mini will be available immediately on Microsoft cloud service platform Azure’s AI model catalog, machine learning model platform Hugging Face, and Ollama, a framework for running models on a local machine.
This post was last modified on April 23, 2024 5:46 am
Rish Gupta is an Indian entrepreneur who serves as the chief executive officer (CEO) of…
Are you looking to advance your engineering career in the field of robotics? Check out…
Artificial intelligence is a topic that has recently made internet users all over the world…
Boost your learning journey with the power of AI communities. The article below highlights the…
Demystify the world of Artificial Intelligence with our comprehensive AI Glossary and Terminologies Cheat Sheet.…
Scott Wu is the co-founder and Chief Executive Officer of Cognition Labs, an artificial intelligence…