Google AI Edge Gallery logo

Google AI Edge Gallery

Bring on-device function calling to iPhone

2026-02-28

Product Introduction

  1. Definition: The Google AI Edge Gallery is an experimental iOS utility application (Category: Utilities) developed by Google. It functions as an on-device platform for running various open-source Generative AI models locally on iPhones (iOS 17.0+), iPads, Macs (macOS 14.0+, Apple Silicon), and Apple Vision Pro (visionOS 1.0+), without requiring an internet connection after initial model download.
  2. Core Value Proposition: It exists to demonstrate and provide user access to cutting-edge on-device AI capabilities, specifically enabling fully offline generative AI tasks and real-time function calling (like Mobile Actions) directly on Apple hardware. Its primary goal is to showcase the potential of local AI processing for privacy, speed, and accessibility.

Main Features

  1. Mobile Actions: This core feature leverages the compact 270M parameter FunctionGemma model to interpret natural language voice commands and execute corresponding device actions entirely offline. It translates phrases like "Create a calendar event for lunch tomorrow" or "Turn on the flashlight" into actual system functions (e.g., Calendar API access, Torch control) using on-device function calling, a first for iPhones. No data leaves the device.
  2. Model Selection & Switching: Users can download and seamlessly switch between different open-source generative AI models sourced primarily from Hugging Face. This allows direct comparison of model performance (speed, output quality) for specific tasks like text generation or image understanding, all running locally on the device's hardware (CPU/Neural Engine).
  3. Multi-Modal Input Processing:
    • Ask Image: Upload or capture images locally. Use offline image understanding models to ask questions about the content (e.g., "What's in this picture?", "Describe this diagram"). The app provides descriptions, solves problems, or identifies objects without cloud processing.
    • Audio Scribe: Upload or record audio clips. Utilize on-device speech-to-text (STT) models for offline transcription into text or translation into other languages, processing audio data locally for enhanced privacy.
  4. AI Task Modules:
    • Prompt Lab: Execute single-turn LLM prompts locally for tasks like text summarization, rewriting, code generation, or creative writing using the selected offline generative model.
    • AI Chat: Engage in multi-turn conversations with the chosen local AI model, demonstrating contextual understanding and response generation entirely on-device.
    • Tiny Garden: An experimental, fully offline mini-game utilizing natural language commands (e.g., "Plant a sunflower", "Water the roses") to interact with the game world, showcasing playful local AI interaction.
  5. Performance Insights: Provides real-time on-device AI benchmarks during model operation, displaying key metrics like Time To First Token (TTFT), token decode speed, and overall task latency, offering transparency into the local processing efficiency.

Problems Solved

  1. Pain Point: Addresses the privacy concerns and latency issues inherent in cloud-based AI services by processing all user data (voice, images, audio, text) exclusively on-device, ensuring sensitive information never leaves the user's iPhone, iPad, or Mac.
  2. Pain Point: Enables offline AI functionality where internet connectivity is unavailable, unreliable, or undesirable (e.g., travel, remote areas, privacy-sensitive environments), allowing access to advanced AI features like image analysis, translation, or device control without a network.
  3. Target Audience:
    • Privacy-Conscious Users: Individuals wary of sending personal data (photos, voice recordings, messages) to cloud servers.
    • Tech Enthusiasts & Developers: Users interested in experimenting with cutting-edge on-device generative AI models and comparing their local performance.
    • Professionals Needing Offline Tools: Travelers, field workers, or anyone requiring reliable AI tools (translation, note transcription, image analysis) without internet access.
    • iOS Power Users: Individuals seeking to leverage the latest local AI capabilities like offline voice commands for device control (Mobile Actions) on their Apple devices.
  4. Use Cases:
    • Creating calendar events or setting reminders via voice command while offline (Mobile Actions).
    • Identifying objects or understanding diagrams in photos without uploading them (Ask Image).
    • Transcribing meeting notes or translating a foreign language conversation recorded locally (Audio Scribe).
    • Experimenting with different open-source LLMs for text generation or coding help entirely offline.
    • Quickly summarizing long articles or rewriting text snippets without an internet connection (Prompt Lab).

Unique Advantages

  1. Differentiation: Unlike most AI apps (including many from Google itself like Gemini) that rely on cloud servers, Google AI Edge Gallery operates fully offline after setup. It uniquely offers real on-device function calling (Mobile Actions) on iOS, a capability typically restricted to cloud APIs or platform-level integrations. It also differs by allowing direct user comparison of multiple local AI models within one app.
  2. Key Innovation: The core innovation is the deployment and efficient execution of relatively capable generative AI models (like the 270M FunctionGemma for Mobile Actions and other open-source models) directly on mobile and edge devices (iPhone, iPad, Mac, Vision Pro) using local hardware acceleration (Apple Neural Engine). This showcases significant advancements in model compression and on-device inference optimization to enable complex tasks like multi-modal understanding and function calling without internet connectivity.

Frequently Asked Questions (FAQ)

  1. Does Google AI Edge Gallery work without internet? Yes, Google AI Edge Gallery is designed for fully offline AI processing. Once the desired AI models (e.g., FunctionGemma for Mobile Actions, or models for Ask Image/Audio Scribe) are downloaded within the app, all features including voice commands, image analysis, audio transcription, and chatting work entirely on-device without needing an internet connection.
  2. What iPhone models support Google AI Edge Gallery? Google AI Edge Gallery requires iOS 17.0 or later. It leverages the device's processor (CPU and Apple Neural Engine) for on-device AI acceleration. Performance may vary based on the specific iPhone model's hardware capabilities, but it is compatible with all devices running iOS 17+.
  3. Is my data private with Google AI Edge Gallery? Yes, Google AI Edge Gallery prioritizes on-device privacy. All data processing (voice commands for Mobile Actions, uploaded images for Ask Image, audio clips for Audio Scribe, chat conversations) occurs locally on your iPhone, iPad, or Mac. No audio, images, or personal command data is sent to Google servers for processing core app functions, as confirmed by its offline operation and privacy labels indicating minimal data linkage (primarily diagnostics/analytics).
  4. Can I use different AI models in the app? Yes, a key feature of Google AI Edge Gallery is the ability to download and switch between different open-source generative AI models (primarily from Hugging Face). This allows users to experiment with and compare the performance and output of various models for tasks like text generation or image understanding, all running locally.
  5. What can I do with the Mobile Actions feature? The Mobile Actions feature uses the on-device FunctionGemma model to turn natural voice commands into actual phone actions offline. Supported actions typically include creating calendar events, setting reminders, opening maps for navigation, toggling the flashlight on/off, and potentially other device controls, all processed locally without cloud dependency.

Subscribe to Our Newsletter

Get weekly curated tool recommendations and stay updated with the latest product news