Share this article

Latest news

With KB5043178 to Release Preview Channel, Microsoft advises Windows 11 users to plug in when the battery is low

Copilot in Outlook will generate personalized themes for you to customize the app

Microsoft will raise the price of its 365 Suite to include AI capabilities

Death Stranding Director’s Cut is now Xbox X|S at a huge discount

Outlook will let users create custom account icons so they can tell their accounts apart easier

The MediaPipe LLM Inference API lets you run LLMs on Android and iOS

The new API allows you to use Gemma, Falcon, Phi 2, and Stable LM

2 min. read

Published onMarch 12, 2024

published onMarch 12, 2024

Share this article

Read our disclosure page to find out how can you help Windows Report sustain the editorial teamRead more

Google’s Experimental MediaPipe LLM Inference API allows you to bringlarge language modelsto your Android and iOS devices. Furthermore, the experimental API can also run LLMs on web platforms. In addition, the API provides initial support for Gemma, Falcon, Phi 2, and Stable LM.

However, it is still under active development. On top of that, on Android devices, production applications with LLMs can useAndroid AICoreto accessGemini APIorGemini Nano.

How do I run LLMs on Android/iOS?

How do I run LLMs on Android/iOS?

To run LLMs on Android and iOS, you can use the MediaPipe LLM Inference API. However, there are other third-party applications, such asMLC LLM. Yet, the Android AICore facilitates the use of Gemini-powered solutions. Moreover, the AICore can use hardware-specific neural accelerators. You can also use a sample of theMediaPipe LLM Inference. In addition, if you have access to Partner Dash, you could try theWeb Demo.

The MediaPipe LLM Inference API allows large language models to runentirely on-device across platforms. It is also easy to use in just a few steps, so you can use LLMs on devices with slightly lower specs. However, you shouldn’t expect them to work at maximum capacity unless your device is high-end. Hopefully, soon, they will add better optimizations and allow lower-spec devices,like phones, to run the LLMs smoother through the API.

The MediaPipe LLM Inference API allows you to run large language models on various devices. In addition, it is easy to use, and someRedditorsconsider it a great opportunity. Furthermore, the API is experimental and will receive more updates and features in the future. However, for now, you can use it for text-to-text generation. Also, it allows you to pick from multiple models to meet your specific needs.

By the way, if you encounter any compatibility issues, check out theLLM Conversion guide.

What do you think? Are you going to use the MediaPipe LLM Inference API? Let us know in the comments.

More about the topics:Google,iOS

Sebastian Filipoiu

Sebastian is a content writer with a desire to learn everything new about AI and gaming. So, he spends his time writing prompts on various LLMs to understand them better. Additionally, Sebastian has experience fixing performance-related problems in video games and knows his way around Windows. Also, he is interested in anything related to quantum technology and becomes a research freak when he wants to learn more.

User forum

0 messages

Sort by:LatestOldestMost Votes

Comment*

Name*

Email*

Commenting as.Not you?

Save information for future comments

Comment

Δ

Sebastian Filipoiu

Sebastian is a content writer with a desire to learn everything new about AI and gaming.