AskCodi logo

AskCodi

Custom LLMs, without training. Use via openai compatible api

2025-11-26

Product Introduction

  1. AskCodi is an OpenAI-compatible LLM orchestration platform that enables developers to build and deploy custom "virtual models" combining prompts, reasoning workflows, review processes, and security guardrails. It acts as a unified interface for multiple AI providers (OpenAI, Anthropic, Google, and open-source models) while allowing teams to define reusable coding models with embedded organizational standards.
  2. The core value lies in streamlining AI-driven development workflows by abstracting complex LLM management, enabling consistent code generation, and enforcing quality controls across all AI interactions. It reduces repetitive prompt engineering while providing enterprise-grade security and cost transparency.

Main Features

  1. The platform offers a unified OpenAI-compatible API endpoint that works with GPT-4, Claude 3, Gemini, and open-source models like Llama and Mistral, eliminating the need to manage multiple SDKs or authentication systems. Users simply modify their existing OpenAI API base URL to https://api.askcodi.com while retaining full compatibility with tools like VS Code, JetBrains IDEs, and CLI workflows.
  2. Custom coding models can be created with stacked prompts, multi-step reasoning modes, and automatic review passes that check outputs for bugs or security issues before delivery. Models support PII masking, custom blocking rules, and organizational policies that apply automatically to all requests through predefined model names like secure-fix-bot or typescript-reviewer.
  3. Built-in cost controls provide pass-through pricing with zero markup, allowing teams to mix expensive frontier models with cost-efficient SLMs (Small Language Models) while tracking token usage across all providers. The platform offers granular analytics showing token consumption per model/provider and automatic throttling when approaching usage limits.

Problems Solved

  1. It eliminates the complexity of managing multiple LLM APIs, authentication methods, and rate limits by providing a single standardized interface compatible with existing OpenAI-based tools. Developers no longer need to rewrite code when switching between AI providers or experimenting with model combinations.
  2. The product specifically targets engineering teams and organizations requiring standardized AI coding practices across IDEs, CLIs, and internal tools. It serves both individual developers seeking consistent code quality and enterprises needing centralized control over AI-generated outputs.
  3. Typical use cases include generating project-specific code with baked-in style guidelines, automatically reviewing AI outputs for security vulnerabilities, and enforcing PII redaction across all AI interactions in regulated industries. Teams can deploy models that combine GPT-4 for complex reasoning with CodeLlama for syntax validation in a single API call.

Unique Advantages

  1. Unlike other AI orchestration tools, AskCodi enables true model portability by maintaining full OpenAI API compatibility while adding layers of customization, allowing immediate integration with unmodified developer tools like Continue.dev and Cursor IDE. This eliminates vendor lock-in while adding advanced functionality.
  2. The platform uniquely combines automatic review modes that execute post-generation quality checks and reasoning modes that force multi-step problem-solving, even on base models not natively designed for chain-of-thought workflows. These features operate through simple model name parameters rather than complex API modifications.
  3. Competitive differentiation comes from transparent cost structure (direct pass-through of provider pricing) and the ability to achieve frontier-model quality at reduced costs by orchestrating SLMs. For example, teams can route simple tasks to Mistral-7B via local inference while reserving Claude Opus for critical reasoning, all through the same API endpoint.

Frequently Asked Questions (FAQ)

  1. What counts towards my token usage? All input and output tokens from API calls are counted, including prompts, responses, and system messages. Token consumption is tracked per provider/model combination, with detailed analytics available in real-time through the AskCodi dashboard. Usage against monthly limits is shown with alerts at 75%, 90%, and 100% thresholds.
  2. How are custom coding models billed? Custom models incur no additional fees beyond the underlying LLM provider costs. If a model combines GPT-4 for initial generation and Claude 3 for review, users pay only for tokens consumed by both models during execution. There are no charges for creating, storing, or managing model configurations.
  3. Which AI models and providers are supported? The platform currently supports OpenAI (GPT-4, GPT-4o), Anthropic (Claude 3.5 Sonnet, Opus, Haiku), Google (Gemini Pro, Flash), and open-source models like Llama 3, Mistral, and CodeLlama through partnerships with Replicate and Fireworks.ai. New models are added biweekly based on developer demand and provider releases.

Subscribe to Our Newsletter

Get weekly curated tool recommendations and stay updated with the latest product news