Microsoft Releases Phi-2, Outperforms Gemini Nano, Mistral 7B, and Llama 2 Models

Share via:

Microsoft has released its Small Language Model (SML) Phi-2, a 2.7 billion-parameter language model showcasing exceptional reasoning and language understanding abilities.

Phi-2, a Transformer-based model with a next-word prediction objective, underwent training on 1.4T tokens from a mix of Synthetic and Web datasets for NLP and coding. The training process, conducted over 14 days using 96 A100 GPUs, resulted in Phi-2, a base model without alignment through reinforcement learning from human feedback (RLHF) or fine-tuning instructions.

Despite its modest 2.7 billion parameters, Phi-2 outperforms Mistral and Llama-2 models, both at 7B and 13B parameters, across various aggregated benchmarks. Particularly noteworthy is its superior performance compared to the significantly larger 70B-parameter Llama-2 model in multi-step reasoning tasks, such as coding and math.

Furthermore, Phi-2 matches or outperforms the recently-announced Google Gemini Nano 2, despite being smaller in size.

Microsoft couldn’t help but make a subtle reference to Google’s staged demo video for Gemini, which received significant criticism. In the video, Google showcased its upcoming AI model, Gemini Ultra, solving complex physics problems and rectifying students’ errors. 

Interestingly, Microsoft highlighted that despite Phi-2 likely being a fraction of the size of Gemini Ultra, it demonstrated the ability to provide accurate answers and correct students using similar prompts.

The post Microsoft Releases Phi-2, Outperforms Gemini Nano, Mistral 7B, and Llama 2 Models appeared first on Analytics India Magazine.

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Popular

More Like this

Microsoft Releases Phi-2, Outperforms Gemini Nano, Mistral 7B, and Llama 2 Models

Microsoft has released its Small Language Model (SML) Phi-2, a 2.7 billion-parameter language model showcasing exceptional reasoning and language understanding abilities.

Phi-2, a Transformer-based model with a next-word prediction objective, underwent training on 1.4T tokens from a mix of Synthetic and Web datasets for NLP and coding. The training process, conducted over 14 days using 96 A100 GPUs, resulted in Phi-2, a base model without alignment through reinforcement learning from human feedback (RLHF) or fine-tuning instructions.

Despite its modest 2.7 billion parameters, Phi-2 outperforms Mistral and Llama-2 models, both at 7B and 13B parameters, across various aggregated benchmarks. Particularly noteworthy is its superior performance compared to the significantly larger 70B-parameter Llama-2 model in multi-step reasoning tasks, such as coding and math.

Furthermore, Phi-2 matches or outperforms the recently-announced Google Gemini Nano 2, despite being smaller in size.

Microsoft couldn’t help but make a subtle reference to Google’s staged demo video for Gemini, which received significant criticism. In the video, Google showcased its upcoming AI model, Gemini Ultra, solving complex physics problems and rectifying students’ errors. 

Interestingly, Microsoft highlighted that despite Phi-2 likely being a fraction of the size of Gemini Ultra, it demonstrated the ability to provide accurate answers and correct students using similar prompts.

The post Microsoft Releases Phi-2, Outperforms Gemini Nano, Mistral 7B, and Llama 2 Models appeared first on Analytics India Magazine.

Disclaimer

We strive to uphold the highest ethical standards in all of our reporting and coverage. We StartupNews.fyi want to be transparent with our readers about any potential conflicts of interest that may arise in our work. It’s possible that some of the investors we feature may have connections to other businesses, including competitors or companies we write about. However, we want to assure our readers that this will not have any impact on the integrity or impartiality of our reporting. We are committed to delivering accurate, unbiased news and information to our audience, and we will continue to uphold our ethics and principles in all of our work. Thank you for your trust and support.

Website Upgradation is going on for any glitch kindly connect at office@startupnews.fyi

More like this

Swift, UBS, Chainlink pilot tokenized fund settlement

Swift, UBS and Chainlink aim to modernize traditional...

PayPay expands digital wage payment system in Japan

PayPay's digital wage payment service was initially exclusive...

Apple @ Work Podcast: Password security 101

Apple @ Work is exclusively brought to you...

Popular

Upcoming Events

Startup Information that matters. Get in your inbox Daily!