Models

Below are the list of models from Deca:

Latest models (2.5)

Deca 2.5 Pro

deca-2.5-pro

Intelligence:

Cost:

  • Normal Price:

    $1.25 / MTok in, $4 / MTok out

Reasoning:

  • Sometimes
    Deca dynamically choses when to think based on the prompt.

Context Length:

  • 128000 context

Deca 2.5 pro is our flagship, groundbreakingly efficient model.

Snapshots:
  • deca-2.5-pro:flex

    ? / MTok in, ? / MTok out

    Flex tier allows you to spend less by only paying for the GPU resources you use. It is generally cheaper than the normal tier. Requests might be queued for processing. Prompts will be used to improve Deca's AI models
  • deca-2.5-pro:free

    $0 / MTok in, $0 / MTok out

    Due to low capacity, this snapshot is currently unavailable. We are working to improve rate limits and usage for all our customers.
  • deca-2.5-pro
  • deca-2.5-pro-beta-06032025

Deca 2.5 Ultra Preview

deca-2.5-ultra-preview

Intelligence:

Cost:

  • Price:

    $7.5 / MTok in, $30 / MTok out

Reasoning:

  • Always

Context Length:

  • 1000000 context

Deca 2.5 Ultra is our largest, and most intelligent model in the Deca 2.5 series. Preview

Deca 2.5 is currently in beta and is subject to change.
Snapshots:
  • deca-2.5-ultra-preview
  • deca-2.5-ultra-071225

Deca 2.5 mini

deca-2.5-mini

Intelligence:

Cost:

  • Normal Price:

    $0.35 / MTok in, $0.99 / MTok out

Reasoning:

  • No

Context Length:

  • 128000 context

Deca 2.5 mini is our smallest model in the Deca 2.5 series.

Snapshots:
  • deca-2.5-mini:flex

    ? / MTok in, ? / MTok out

    Flex tier allows you to spend less by only paying for the GPU resources you use. It is generally cheaper than the normal tier. Requests might be queued for processing. Prompts will be used to improve Deca's AI models
  • deca-2.5-mini
  • deca-2.5-mini-beta-06032025

Using Deca Models

Once you have selected a model, this guide will provide the necessary steps to begin using the Deca API.

1. Get Your API Key

Sign in to the and navigate to the API Keys page. Click the button to generate a new key. Copy this key and store it securely, as you will need it for all API requests.

Don't have an account? for free.

2. Fire Up a Request

cURL

curl https://api.genlabs.dev/deca/v1/chat/completions \
  -H "Authorization: Bearer <YOUR_API_KEY>" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "deca-2.5-ultra-preview",
    "messages": [
      { "role": "user", "content": "Write acrostic about Deca, a powerful AI model" }
    ]
  }'

Python + OpenAI SDK

                  from openai import OpenAI
client = OpenAI(
  base_url = "https://api.genlabs.dev/deca/v1",
  api_key = "YOUR_API_KEY"
)
completion = client.chat.completions.create(
  model="deca-2.5-ultra-preview",
  messages=[
    {
      "role": "user",
      "content": "Write acrostic about Deca, a powerful AI model"
    }
  ]
)

print(completion.choices[0].message.content)

                  
                

Substitute deca-2.5-ultra-preview with any model or snapshot from the table above.

Tips:

3. Parameters Cheat-Sheet

ParamWhat It DoesDefault
temperatureControls the randomness of the output. A value of 0 makes the model deterministic, while a higher value like 2.0 increases creativity.0.7
max_tokensHow long the model can talk before we stop it.Max*
top_pTop_p controls text diversity by selecting from the smallest set of words whose probabilities add up to a threshold. 1.0
streamSet to true for token-by-token response.false

* The default maximum tokens is calculated by context - input. For example, if you send 10,000 tokens to Deca 2.5 Pro (128,000 context), the default max tokens will be 128,000 - 10,000 = 118,000 tokens.

4. Counting (and Saving) Tokens

Prices are shown as $ / MTok – that’s per million tokens. For a ten-token “What are the best hotels in Italy?”, you’ll pay roughly the cost of finding a penny on the sidewalk. Use the cheaper :free snapshots for testing, then move up when you feel you're ready for production.

Rate Limits (Beta)

Beta Notice: The rate limits below are part of a beta rollout to ensure API stability. These limits are temporary and will be increased as we scale our infrastructure. If you require higher limits for your application, please don't hesitate to reach out to us at [email protected].


Model Tier Rate Limit
Pro & Mini (Normal) 1 request per second & 1000 requests per hour
Pro & Mini (Flex) 1 request per second & 500 requests per hour
Ultra 1 request per second & 500 requests per hour

6. Error Codes That Matter

Provide Feedback

0/500