Oxlo.ai

Text Generation

Generate human-like text using state-of-the-art LLMs via the OpenAI-compatible Chat Completions API.

OpenAI Compatible: Use the standard openai Python library just set base_url to https://api.oxlo.ai/v1. Any existing OpenAI code works without changes.

Available Models

Chat Models

ModelAPI IDTierBest For
Llama-3.2-3Bllama-3.2-3bFreeFast responses, edge deployment
Mistral-7Bmistral-7bFreeChat, summaries, basic reasoning
DeepSeek V3.2deepseek-v3.2FreeGeneral-purpose chat, analysis
Llama-3.1-8Bllama-3.1-8bProInstruction following, reasoning
Qwen 2.5 7Bqwen-2.5-7bProMultilingual, general reasoning
Ministral-14Bministral-14bProMultilingual, vision-capable
Llama-4-Maverick-17Bllama-4-maverick-17bProVersatile MoE, diverse tasks
DeepSeek V3 0324deepseek-v3-0324ProEnhanced general-purpose chat
Kimi-2.5kimi-k2.5PremiumBalanced reasoning, vision
Qwen 3 32Bqwen-3-32bPremiumAdvanced reasoning, enterprise
Llama-3.3-70Bllama-3.3-70bPremiumLong context, instruction following
GPT-OSS 120Bgpt-oss-120bPremiumComplex reasoning, long context

Reasoning Models

ModelAPI IDTierBest For
DeepSeek R1 8Bdeepseek-r1-8bFreeMath, science, chain-of-thought
DeepSeek R1 70Bdeepseek-r1-70bProAdvanced reasoning, analysis
GPT-OSS 20Bgpt-oss-20bProAgentic workflows, complex tasks
Kimi-K2-Thinkingkimi-k2-thinkingPremiumDeep reasoning, long-form analysis
DeepSeek-R1-0528deepseek-r1-0528PremiumFrontier-class reasoning

Coding Models

ModelAPI IDTierBest For
DeepSeek Coder 33Bdeepseek-coder-33bProCode generation, refactoring

Quick Example

Chat with any model using the OpenAI-compatible API:

import openai

client = openai.OpenAI(
    base_url="https://api.oxlo.ai/v1",
    api_key="<YOUR_API_KEY>"
)

response = client.chat.completions.create(
    model="deepseek-r1-8b",
    messages=[{"role": "user", "content": "Hello!"}],
    max_tokens=512
)

print(response.choices[0].message.content)

Multi-Turn Conversation

Pass conversation history in the messages array:

response = client.chat.completions.create(
    model="llama-3.3-70b",
    messages=[
        {"role": "system", "content": "You are a helpful coding assistant."},
        {"role": "user", "content": "Write a Python function to reverse a string"},
        {"role": "assistant", "content": "def reverse(s): return s[::-1]"},
        {"role": "user", "content": "Now make it handle Unicode correctly"}
    ],
    max_tokens=1024,
    temperature=0.3
)

Parameters

ParameterTypeDefaultDescription
modelstringModel ID (required)
messagesarrayConversation messages (required)
max_tokensinteger256Max tokens to generate (1–131,072)
temperaturefloat0.7Randomness (0.0 = deterministic, 2.0 = creative)
top_pfloat1.0Nucleus sampling threshold
frequency_penaltyfloat0Penalize repeated tokens (-2.0 to 2.0)
presence_penaltyfloat0Penalize tokens already present (-2.0 to 2.0)
stopstring[]nullUp to 4 stop sequences
seedintegernullFor reproducible results