Small Language Models (SLM) are trained on focused datasets, making them very efficient at tasks like analyzing customer feedback, generating product descriptions, or handling specialized industry ...
In a new case study, Hugging Face researchers have demonstrated how small language models (SLMs) can be configured to outperform much larger models. Their findings show that a Llama 3 model with 3B ...
Sometimes the best way to solve a complex problem is to take a page from a children’s book. That’s the lesson Microsoft researchers learned by figuring out how to pack more punch into a much smaller ...
AI thrives on data but feeding it the right data is harder than it seems. As enterprises scale their AI initiatives, they face the challenge of managing diverse data pipelines, ensuring proximity to ...
As tech companies race to deliver on-device AI, we are seeing a growing body of research and techniques for creating small language models (SLMs) that can run on resource-constrained devices. The ...
The original version of this story appeared in Quanta Magazine. Large language models work well because they’re so large. The latest models from OpenAI, Meta, and DeepSeek use hundreds of billions of ...
Microsoft announced a new version of its small language model, Phi-3, which can look at images and tell you what’s in them. Phi-3-vision is a multimodal model — aka it can read both text and images — ...
Microsoft just released its latest small language model that can operate directly on the user’s computer. If you haven’t followed the AI industry closely, you might be asking: what exactly is a small ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results