One API, every major model

One API endpoint for every major AI model.

MoleAPI is an AI model gateway. Access GPT, Claude, Gemini, and more through a single OpenAI-compatible endpoint. Manage all your API keys and usage in one place.

Why MoleAPI

More than a proxy — a simpler way to manage AI

One endpoint for multiple model providers, one dashboard for all your keys and usage, switch models anytime without changing code.

One endpoint, many models

Access GPT, Claude, Gemini, and more through a single API endpoint. Switch models by changing one parameter — no code rewrite needed.

Works with your existing tools

Compatible with OpenAI SDK, Cursor, Claude Code, and other tools you already use. Just swap the base URL.

Keys and usage in one place

Create and manage API keys, set spending limits, and see usage breakdowns by model — all from one dashboard.

Product Overview

Models, integrations, and solutions — three ways to explore MoleAPI

See which models we support, what tools you can connect, and which use cases fit.

Supported Models

All major models, one gateway

Access GPT-4o, Claude 3.5, Gemini Pro, and more. Covers chat, reasoning, coding, and embeddings. New models available as soon as they launch.

Chat & reasoning models

GPT-4o, Claude 3.5 Sonnet, Gemini Pro, and more — for conversations, analysis, and complex reasoning tasks.

Code & embeddings

Specialized models for code generation, text embeddings, and RAG pipelines. Same API, different model name.

New models on day one

When providers launch new models, we add them quickly. No need to wait for SDK updates or change your integration.

View all models
Integrations

Works with the tools you already use

Compatible with OpenAI SDK, Cursor, Claude Code, and more. Just change the API base URL — no SDK swap needed.

View integrations
Use Cases

From solo developers to teams

Whether you want to try different models affordably or need centralized AI management for your team, MoleAPI has you covered.

View solutions

Guides

Step-by-step integration and migration guides

Set up from scratch, migrate from OpenAI, configure Cursor or Claude Code — we have tutorials for all of it.

FAQ

Common questions

Answers about pricing, compatibility, and how things work — the questions new users ask most.

How is pricing calculated?

You pay per token, same as using model providers directly. Check the console for current pricing per model. New accounts get free credits to try things out.

Is MoleAPI slower than calling providers directly?

MoleAPI adds minimal latency (typically under 50ms). Responses are streamed directly from the provider, so you get the same speed as a direct connection.

Can I use my existing OpenAI code?

Yes. Just change the base URL and API key — your existing code, SDK, and tools work as-is. No other changes needed.

What happens if a model provider goes down?

You can switch to another model by changing one parameter. Since all models use the same API format, your application logic stays the same.