API Documentation

Complete reference for the OMA-AI API. OpenAI-compatible, privacy-first, powered by Venice.

Overview

OMA-AI provides access to 38 AI models through a simple REST API. The API is OpenAI-compatible, making it easy to switch from other providers.

Base URL

https://oma-ai.com/api

Key Features

Authentication

All API requests require an API key in the Authorization header:

Authorization: Bearer oma_your_api_key

Get your API key from the dashboard.

Quick Start

# Simple chat completion
curl https://oma-ai.com/api/llm \
  -H "Authorization: Bearer oma_your_key" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "deepseek-v3.2",
    "prompt": "Hello, world!"
  }'

Chat Completions

POST /api/llm

Create a chat completion with any of our 38 models.

Request Body

Parameter Type Default Description
prompt string required The user's prompt/message
model string "deepseek-v3.2" Model ID or alias
max_tokens integer 1000 Maximum tokens in response
temperature float 0.7 Sampling temperature (0-2)
web_search boolean false Enable web search with citations
uncensored boolean false Disable content filters

Model Aliases

Use these shortcuts instead of full model IDs:

Alias Model Use Case
default DeepSeek V3.2 Best value
cheap GLM 4.7 Flash Fast, cheap
code Qwen 3 Coder Turbo Code generation
premium Kimi K2.5 Highest quality
uncensored Venice Uncensored No filters

Response

{
  "success": true,
  "response": "The AI response text...",
  "model": "deepseek-v3.2",
  "model_id": "deepseek-v3.2",
  "privacy": "private",
  "tier": "standard",
  "usage": {
    "prompt_tokens": 12,
    "completion_tokens": 156,
    "total_tokens": 168
  },
  "cost": {
    "input_usd": "0.000005",
    "output_usd": "0.000156",
    "total_usd": "0.000161"
  },
  "timestamp": 1709123456789
}

List Models

GET /api/llm

Get a list of all available models with pricing and capabilities.

Response

{
  "success": true,
  "total": 38,
  "models": [
    {
      "id": "deepseek-v3.2",
      "model_id": "deepseek-v3.2",
      "privacy": "private",
      "tier": "standard",
      "pricing": {
        "input": 0.40,
        "output": 1.00,
        "unit": "per 1M tokens"
      },
      "context": 160000,
      "best_for": "Best value, excellent quality"
    }
  ],
  "aliases": { ... },
  "default_model": "deepseek-v3.2"
}

Enable real-time web search with citations on any model:

curl https://oma-ai.com/api/llm \
  -H "Authorization: Bearer oma_your_key" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "deepseek-v3.2",
    "prompt": "What are the latest developments in AI?",
    "web_search": true
  }'

Responses will include citations from web sources when relevant.

Uncensored Mode

Warning

Uncensored mode disables content safety filters. Use responsibly.

Disable content filters for unrestricted generation:

curl https://oma-ai.com/api/llm \
  -H "Authorization: Bearer oma_your_key" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "venice-uncensored",
    "prompt": "Write an unfiltered analysis...",
    "uncensored": true
  }'

Pricing

Pricing is per 1 million tokens. Input and output tokens are priced separately.

Tier Input Output Example Models
Budget $0.07 - $0.35 $0.20 - $2.50 GLM 4.7 Flash, Gemma 3, Nemotron
Standard $0.40 - $1.10 $1.00 - $3.75 DeepSeek V3.2, Llama 3.3 70B, Kimi K2.5
Premium $2.19 - $6.00 $15.00 - $30.00 GPT-5.2, Claude Opus 4.6, Gemini 3.1 Pro

Every API response includes exact cost calculation in the cost field.

Questions? Join our Discord or check the GitHub repo.