Microsoft just released Phi-3.5-MoE, an advanced AI model that rivals the reasoning capabilities of much larger models while Microsoft just released it with efficient architecture.
The details:
- Phi-3.5-MoE uses a new mixture-of-experts (MoE) approach, which selectively activates only the most relevant parts of the model for each task to save compute power.
- The new model excels at understanding and following complex instructions and can handle up to ~125,000 words in a single prompt.
- In head-to-head benchmarks, Phi-3.5-MoE outperformed popular models like Meta’s Llama 3 8B and Google’s Gemma 2 9B, but fell short against OpenAI’s GPT-4o mini.
- Microsoft made the model available under an open-source MIT license on Hugging Face.
Why it matters: While the mainstream media focuses on the most advanced large language model, there’s also another race amongst tech giants for the smartest, fastest, and smallest AI. Breakthroughs like Phi-3.5-MoE are paving the way for advanced AI models to run directly and privately on our mobile devices.Wanna know how to create an advanced AI modal using Phi-3.5-MoE?Register today