close
close

Meta AI Develops Compact Language Model for Mobile Devices


We want to hear from you! Take our short AI survey and share your thoughts on the current state of AI, how you’re implementing it, and what you expect in the future. Learn more


Meta AI researchers have unveiled MobileLLM, a new approach to building efficient language models designed for smartphones and other resource-constrained devices. Published June 27, 2024, the work challenges assumptions about the necessary size of effective AI models.

The research team, consisting of members from Meta Reality Labs, PyTorch, and Meta AI Research (FAIR), focused on optimizing models with fewer than 1 billion parameters. That’s a fraction of the size of models like GPT-4, which are estimated to have more than a trillion parameters.

Yann LeCun, Chief AI Scientist at Meta, highlighted key aspects of X (formerly known as Twitter) research:


Countdown to VB Transform 2024

Join enterprise leaders in San Francisco July 9-11 for our flagship AI event. Connect with peers, explore the opportunities and challenges of Generative AI, and learn how to integrate AI applications into your industry. Register now


Key innovations in MobileLLM include:

  1. Prioritizing model depth over width
  2. Implementing shared embedding and attention for group queries
  3. Using the innovative technique of instantaneous block weight sharing

These design choices allowed MobileLLM to outperform previous models of similar size by 2.7% to 4.3% on typical test tasks. While these single-digit improvements may seem small, they represent significant progress in the highly competitive field of language model development.

Interestingly, the 350 million parameter version of MobileLLM showed comparable accuracy to the much larger 7 billion LLaMA-2 model for some API call tasks. This suggests that for some specific applications, more compact models can offer similar functionality while using significantly fewer computational resources.

Source: “MobileLLM: Optimizing Sub-Billion Parameter Language Models for On-Device Use Cases,” Zechun Liu et al., Meta

The development of MobileLLM is in line with a growing interest in more efficient AI models. As progress in very large language models shows signs of slowing, researchers are increasingly exploring the potential of more compact, specialized designs. The focus on performance and on-device implementation puts MobileLLM in a similar category to what some researchers call Small Language Models (SLMs), despite the “LLM” in the name.

Although MobileLLM is not yet available for public use, Meta has made the pre-training code available, allowing other researchers to build on their work. As the technology matures, it could enable more advanced AI capabilities on personal devices, although the timeline and exact capabilities remain uncertain.

The development of MobileLLM is an important step in making advanced AI more accessible and sustainable. It challenges the notion that effective language models need to be massive, potentially opening up new possibilities for AI applications on personal devices.