Meta makes its MobileLLM open for researchers


Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Meta AI has announced the open-source release of MobileLLM, a set of language models optimized for mobile devices, with model checkpoints and code now accessible on Hugging Face. However, it is presently only available under a Creative Commons 4.0 non-commercial license, meaning enterprises can’t use it on commercial products.

Originally described in a research paper published in July 2024 and covered by VentureBeat, MobileLLM is now fully available with open weights, marking a significant milestone for efficient, on-device AI.

The release of these open weights makes MobileLLM a more direct, if roundabout, competitor to Apple Intelligence, Apple’s on-device/private cloud hybrid AI solution made up of multiple models, shipping out to users of its iOS 18 operating system in the U.S. and outside the EU this week. However, being restricted to research use and requiring downloading and installation from Hugging Face, it’s likely to remain limited to a computer science and academic audience for now.

More efficiency for mobile devices

MobileLLM aims to tackle the challenges of deploying AI models on smartphones and other resource-constrained devices.

With parameter counts ranging from 125 million to 1 billion, these models are designed to operate within the limited memory and energy capacities typical of mobile hardware.

By emphasizing architecture over sheer size, Meta’s research suggests that well-designed compact models can deliver robust AI performance directly on devices.

Resolving scaling issues

The design philosophy behind MobileLLM deviates from traditional AI scaling laws that emphasize width and large parameter counts.

Meta AI’s research instead focuses on deep, thin architectures to maximize performance, improving how abstract concepts are captured by the model.

Yann LeCun, Meta’s Chief AI Scientist, highlighted the importance of these depth-focused strategies in enabling advanced AI on everyday hardware.

MobileLLM incorporates several innovations aimed at making smaller models more effective:

Depth Over Width: The models employ deep architectures, shown to outperform wider but shallower ones in small-scale scenarios.

Embedding Sharing Techniques: These maximize weight efficiency, crucial for maintaining compact model architecture.

Grouped Query Attention: Inspired by work from Ainslie et al. (2023), this method optimizes attention mechanisms.

Immediate Block-wise Weight Sharing: A novel strategy to reduce latency by minimizing memory movement, helping keep execution efficient on mobile devices.

Performance Metrics and Comparisons

Despite their compact size, MobileLLM models excel on benchmark tasks. The 125 million and 350 million parameter versions show 2.7% and 4.3% accuracy improvements over previous state-of-the-art (SOTA) models in zero-shot tasks.

Remarkably, the 350M version even matches the API calling performance of the much larger Meta Llama-2 7B model.

These gains demonstrate that well-architected smaller models can handle complex tasks effectively.

Designed for smartphones and the edge

MobileLLM’s release aligns with Meta AI’s broader efforts to democratize access to advanced AI technology.

With the increasing demand for on-device AI due to cloud costs and privacy concerns, models like MobileLLM are set to play a pivotal role.

The models are optimized for devices with memory constraints of 6-12 GB, making them practical for integration into popular smartphones like the iPhone and Google Pixel.

Open but non-commercial

Meta AI’s decision to open-source MobileLLM reflects the company’s stated commitment to collaboration and transparency. Unfortunately, the licensing terms prohibit commercial usage for now, so only researchers can benefit.

By sharing both the model weights and pre-training code, they invite the research community to build on and refine their work.

This could accelerate innovation in the field of small language models (SLMs), making high-quality AI accessible without reliance on extensive cloud infrastructure.

Developers and researchers interested in testing MobileLLM can now access the models on Hugging Face, fully integrated with the Transformers library. As these compact models evolve, they promise to redefine how advanced AI operates on everyday devices.



Source link

Share

Latest Updates

Frequently Asked Questions

Related Articles

Get this 15-inch HP Ryzen laptop with 16GB of RAM for nearly half off

Eight hundred bucks for a laptop with an older processor isn’t a great...

Zuckerberg Firing Hundreds of AI Developers After Hiring Spree

Mark Zuckerberg’s Meta is once again shaking up its artificial intelligence unit: as...

AI Models Get Brain Rot, Too

AI models may be a bit like humans, after all.A new study from...

Google claims first ‘verifiable’ quantum advantage for Willow chip

Google has claimed that its quantum processor Willow has achieved the first “verifiable”...
custom cakes home inspections business brokerage life counseling rehab center residences chiropractic clinic surf school merchant advisors poker room med spa facility services creative academy tea shop life coach restaurant life insurance fitness program electrician NDIS provider medical academy sabung ayam online judi bola judi bola judi bola judi bola Slot Mahjong slot mahjong Slot Mahjong judi bola sabung ayam online mahjong ways mahjong ways mahjong ways judi bola SV388 SABUNG AYAM ONLINE GA28 judi bola online sabung ayam online live casino online live casino online SV388 SV388 SV388 SV388 SV388 Mix parlay sabung ayam online SV388 SBOBET88 judi bola judi bola judi bola Reset Pola Blackjack Jadi Kasus Study Mahjong Ways Mahjong Ways Mahjong Ways Mahjong Ways sabung ayam online sabung ayam online judi bola sabung ayam online judi bola Judi Bola Sabung Ayam Online Live Casino Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online Sabung Ayam Online sabung ayam online judi bola mahjong ways sabung ayam online judi bola mahjong ways mahjong ways sabung ayam online sv388 Sv388 judi bola judi bola judi bola JUARA303 Mahjong ways Judi Bola Judi Bola Sabung Ayam Online Live casino mahjong ways 2 sabung ayam online sabung ayam online mahjong ways mahjong ways mahjong ways SV388 SBOBET88 judi bola judi bola judi bola judi bola judi bola https://himakom.fisip.ulm.ac.id/ SABUNG AYAM ONLINE MIX PARLAY SLOT GACOR JUDI BOLA SV388 LIVE CASINO LIVE CASINO ONLINE Judi Bola Online SABUNG AYAM ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE JUDI BOLA ONLINE LIVE CASINO ONLINE LIVE CASINO ONLINE sabung ayam online Portal SV388 SBOBET88 SABUNG AYAM ONLINE JUDI BOLA ONLINE CASINO ONLINE MAHJONG WAYS 2 sabung ayam online judi bola SABUNG AYAM ONLINE JUDI BOLA ONLINE Sabung Ayam Online JUDI BOLA Sabung Ayam Online JUDI BOLA SV388, WS168 & GA28 SBOBET88 SV388, WS168 & GA28 SBOBET88 SBOBET88 CASINO ONLINE SLOT GACOR Sabung Ayam Online judi bola