Share this article
Latest news
With KB5043178 to Release Preview Channel, Microsoft advises Windows 11 users to plug in when the battery is low
Copilot in Outlook will generate personalized themes for you to customize the app
Microsoft will raise the price of its 365 Suite to include AI capabilities
Death Stranding Director’s Cut is now Xbox X|S at a huge discount
Outlook will let users create custom account icons so they can tell their accounts apart easier
The MediaPipe LLM Inference API lets you run LLMs on Android and iOS
The new API allows you to use Gemma, Falcon, Phi 2, and Stable LM
2 min. read
Published onMarch 12, 2024
published onMarch 12, 2024
Share this article
Read our disclosure page to find out how can you help Windows Report sustain the editorial teamRead more
Google’s Experimental MediaPipe LLM Inference API allows you to bringlarge language modelsto your Android and iOS devices. Furthermore, the experimental API can also run LLMs on web platforms. In addition, the API provides initial support for Gemma, Falcon, Phi 2, and Stable LM.
However, it is still under active development. On top of that, on Android devices, production applications with LLMs can useAndroid AICoreto accessGemini APIorGemini Nano.
How do I run LLMs on Android/iOS?
To run LLMs on Android and iOS, you can use the MediaPipe LLM Inference API. However, there are other third-party applications, such asMLC LLM. Yet, the Android AICore facilitates the use of Gemini-powered solutions. Moreover, the AICore can use hardware-specific neural accelerators. You can also use a sample of theMediaPipe LLM Inference. In addition, if you have access to Partner Dash, you could try theWeb Demo.
The MediaPipe LLM Inference API allows large language models to runentirely on-device across platforms. It is also easy to use in just a few steps, so you can use LLMs on devices with slightly lower specs. However, you shouldn’t expect them to work at maximum capacity unless your device is high-end. Hopefully, soon, they will add better optimizations and allow lower-spec devices,like phones, to run the LLMs smoother through the API.
The MediaPipe LLM Inference API allows you to run large language models on various devices. In addition, it is easy to use, and someRedditorsconsider it a great opportunity. Furthermore, the API is experimental and will receive more updates and features in the future. However, for now, you can use it for text-to-text generation. Also, it allows you to pick from multiple models to meet your specific needs.
By the way, if you encounter any compatibility issues, check out theLLM Conversion guide.
What do you think? Are you going to use the MediaPipe LLM Inference API? Let us know in the comments.
More about the topics:Google,iOS
Sebastian Filipoiu
Sebastian is a content writer with a desire to learn everything new about AI and gaming. So, he spends his time writing prompts on various LLMs to understand them better. Additionally, Sebastian has experience fixing performance-related problems in video games and knows his way around Windows. Also, he is interested in anything related to quantum technology and becomes a research freak when he wants to learn more.
User forum
0 messages
Sort by:LatestOldestMost Votes
Comment*
Name*
Email*
Commenting as.Not you?
Save information for future comments
Comment
Δ
Sebastian Filipoiu
Sebastian is a content writer with a desire to learn everything new about AI and gaming.