Language models could soon be running on our smartphones.
The race for the best small language models (SMLs) is on! With more and more smartphone makers looking to run AI models right on your device, tech giants are battling it out to release the most powerful SMLs to the market. Today, Microsoft just made a big move with the release of Phi-3, a new family of small but seriously impressive language models.
Isn’t it interesting to use an AI chatbot like ChatGPT on our smartphones without the need to be connected to the internet?
The Phi-3 family is a set of AI models that are designed to be the most capable and cost-effective SLMs available, outperforming models of similar and larger sizes across various benchmarks in language, reasoning, coding, and math.
The Phi-3 family offers a range of models along the quality-cost curve, providing customers with practical choices when building generative AI applications.
This flexibility is crucial for developers and companies looking to integrate AI capabilities into their products without breaking the bank.
These models achieve groundbreaking performance on key benchmarks despite their relatively small size. However, the smaller model size limits their performance on factual knowledge benchmarks like TriviaQA.
The image below compares the quality (as measured by performance on the Massive Multitask Language Understanding benchmark) vs. size (in billions of active parameters) for various SLMs.
The key takeaways are:
Take a look at this example of Phi-3 mini running on a mobile phone.
That’s pretty impressive, isn’t it? The fact that we can now run AI models like this directly on our smartphones is a game-changer.
If you want to learn more about Phi-3, check out the whitepaper here.
Phi-3 mini is currently available on the following platforms:
On Azure AI Studio, the Phi-3-mini 4k and 128k instruct are already available.
Here’s what the dashboard looks like:
The model in HuggingChat seems to be even more capable and also has the ability to search the internet, which is a fantastic feature.
It’s pretty wild to think that we’ll soon be carrying around smartphones with built-in AI models that can rival the likes of ChatGPT. I’m incredibly excited to see how this technology evolves and what new apps and experiences it enables.
While Phi-3 and similar small language models are still limited compared to their much larger counterparts like GPT-4 in terms of knowledge, reasoning, and generation capabilities, they represent a significant step forward in on-device AI. There will likely continue to be a gap in capabilities between device-based models and cloud-based models for the foreseeable future, but that gap is closing faster than many of us expected.
But overall, I’m optimistic — I think on-device AI has the potential to make our gadgets feel a whole lot smarter. So, what do you think? Are you ready to welcome Phi-3 and other small language models into your pocket?
Software engineer, writer, solopreneur