Cai logo

Cai

Press ⌥C on anything tor to run smart actions

2026-04-22

Product Introduction

  1. Definition: Cai is a local AI-native action layer and productivity utility specifically designed for macOS. It functions as an extensible middleware that sits between the operating system's UI and various execution engines—including Large Language Models (LLMs), shell scripts, and third-party APIs. Built with a focus on privacy and efficiency, it enables users to execute complex workflows directly on selected text or images using a system-wide hotkey.

  2. Core Value Proposition: Cai exists to eliminate "context switching" and "manual copy-paste loops" by providing a frictionless interface for local intelligence. Its primary value lies in its ability to bring "bring-your-own-brain" AI capabilities directly into any macOS application (Terminal, Slack, VS Code, Browser) without requiring a cloud connection, a subscription, or account registration. By leveraging the Apple Silicon MLX framework, it offers high-performance local inference that respects user privacy and data sovereignty.

Main Features

  1. Built-in Local Inference with Ministral 3B: Cai ships with the Ministral 3B model, pre-configured via the MLX framework for optimized performance on Apple Silicon (M1/M2/M3/M4 chips). This allows for zero-setup, offline AI processing, including text summarization, grammar correction, and code explanation, without any external API dependencies.

  2. Multi-Engine Model Support: Beyond its built-in capabilities, Cai acts as a gateway for diverse AI backends. It supports any HuggingFace MLX model, local servers via Ollama and LM Studio, and native Apple Intelligence (on supported macOS versions). For users requiring massive scale, it provides optional integration with cloud providers like OpenRouter and OpenAI via API keys.

  3. Custom Action & Shell Scripting Engine: Users can define custom actions using shell scripts (e.g., Bash, Zsh) or URL templates. By using the {{selected text}} placeholder, Cai can pipe highlighted content directly into terminal commands (like lsof or gh pr list) or custom webhooks. This transforms the tool from a simple AI wrapper into a powerful automation layer for developers and power users.

  4. On-Device OCR & Image Processing: Cai includes an integrated Optical Character Recognition (OCR) engine. Users can take a screenshot or select an image containing text (such as a terminal error or a non-selectable PDF), and Cai will extract the text to immediately run actions like "Fix Bug" or "Translate" on the captured content.

  5. Deep Integration with GitHub & Linear: Through native connectors, Cai allows users to convert any selected text or error message into a structured GitHub Issue or Linear Ticket. This integration facilitates a "one-click" project management workflow, populating titles and descriptions using AI based on the highlighted context.

Problems Solved

  1. Pain Point: Friction in AI Workflows: Traditional AI assistants require users to copy text, switch to a browser or app, paste the text, and wait for a response. Cai solves this by operating inline; the ⌥C shortcut triggers actions directly over the selection, maintaining the user's "flow state."

  2. Pain Point: Data Privacy & Security: Many corporate environments prohibit the use of cloud-based AI due to telemetry and data retention concerns. Cai solves this by being 100% local by default, featuring an MIT license, and requiring no account or cloud-based telemetry.

  3. Target Audience:

    • Software Engineers: For debugging terminal output, generating git commands, and managing GitHub issues.
    • Technical Project Managers: For quickly triaging messages from Slack/Teams into Linear tickets.
    • Content Creators & Researchers: For summarizing long-form content, translating strings, and extracting text from images.
    • Privacy Enthusiasts: Users who want the power of LLMs without the privacy trade-offs of cloud-first platforms.
  4. Use Cases:

    • Selecting a stack trace in a terminal and running a "Fix Code" action.
    • Highlighting an email in a foreign language and using a "Translate" custom prompt.
    • Extracting text from a design mockup (OCR) and sending it to a Slack channel.
    • Running a custom shell script to kill a process based on a selected port number.

Unique Advantages

  1. Differentiation: Unlike standard clipboard managers (e.g., Maccy) which focus on storage, Cai focuses on execution. Unlike AI chat interfaces (e.g., ChatGPT Desktop), Cai is a "headless" action layer that works within the context of other apps. Compared to Raycast, Cai offers a more specialized, open-source focus on LLM integration and local MLX model support without the overhead of a proprietary ecosystem.

  2. Key Innovation: The application’s core innovation is the marriage of the Model Context Protocol (MCP) principles with macOS Accessibility permissions. By simulating system-level commands, it can intercept and act upon UI elements in any application, effectively turning the entire operating system into an interactive, AI-augmented canvas.

Frequently Asked Questions (FAQ)

  1. Is Cai really free and open source? Yes, Cai is released under the MIT License. There are no subscriptions, no trials, and no hidden fees. All source code is accessible for audit and community contribution. While the app is free, using optional cloud providers like OpenRouter would require your own API key and associated costs.

  2. Does Cai work entirely offline? Yes. If you use the built-in Ministral 3B model, a local Ollama/LM Studio instance, or Apple Intelligence, all data processing stays on your machine. No internet connection is required for these local workflows, making it ideal for secure or remote environments.

  3. How does Cai differ from Raycast or Maccy? While Maccy is a dedicated clipboard manager for history retrieval, Cai is an "action layer" that processes your current selection. Compared to Raycast, Cai is more focused on the "Bring Your Own Model" (BYOM) philosophy and provides deeper, scriptable integration for LLMs specifically, rather than being a general-purpose launcher.

  4. Which Mac models are compatible with Cai? Cai is optimized for macOS. While it runs on Intel-based Macs using cloud APIs or local servers, the high-performance local AI features (MLX) are specifically designed for Apple Silicon (M1, M2, M3, M4) to utilize the Neural Engine and Unified Memory for maximum speed.

Subscribe to Our Newsletter

Get weekly curated tool recommendations and stay updated with the latest product news