Google Gemma (@googlegemma) on X
gemini gemma google
| Source: Mastodon | Original article
Google’s AI team has posted a short video on X showing how to run the latest Gemma 4 model directly on an iPhone, completely offline. The demonstration highlights that the model can handle long‑context prompts without touching the cloud, eliminating data‑transfer fees, API costs and any recurring subscription. The clip, shared from the @googlegemma account, walks viewers through the installation steps and showcases a real‑time chat session that runs entirely on the device’s processor.
The move matters because it pushes the frontier of edge AI from laptops and servers to handheld consumer hardware. By leveraging the same research that underpins Google’s Gemini series, Gemma 4 offers a lightweight yet capable large‑language model that can be embedded in apps without exposing user data to external servers. For Nordic users, where privacy regulations are strict and mobile connectivity can be spotty in remote areas, an offline LLM opens new possibilities for secure personal assistants, on‑device translation and localized content generation. It also signals Google’s intent to compete with Apple’s own on‑device language models and with Meta’s open‑source initiatives, potentially reshaping the economics of AI‑powered mobile services.
As we reported on 16 April, the Gemma family already proved its efficiency on CPUs, with Gemma2B out‑performing GPT‑3.5 Turbo in benchmark tests. The iPhone rollout suggests Google is now translating that efficiency into a consumer‑ready form factor. The next steps to watch include performance benchmarks on Apple’s M‑series chips, the release of developer toolkits for iOS integration, and whether Google will extend offline support to other platforms such as Android tablets or wearables. Industry observers will also be keen to see how the model’s accuracy and safety controls hold up when stripped of cloud‑based moderation layers.
Sources
Back to AIPULSEN