GPT-OSS Chat
Powered by OpenAI's GPT models via EigenAI endpoint - engage in intelligent conversations
Backend online
0 characters
•
Input 0.10 /M tokens • Output 0.50 /M tokens
🔌 API Access
Integrate GPT-OSS chat into your applications using our REST API.
🔑 API Keys
Use an active API key with every request. Manage your API keys →
POST
/api/v1/generate
Chat with GPT-OSS
Send conversation messages to GPT-OSS.
Two options for API calls:
- Non-streaming:
/api/v1/generatewith JSON body (recommended for complete responses) - Streaming:
/api/v1/generate/streamwith"stream": true(real-time responses) - Alternative:
/api/v1/generate/streamwith"stream": false(equivalent to /api/v1/generate)
Cost:
0.10 credits / 1M input tokens • 0.50 credits / 1M output tokens
Streaming responses include final usage events with reasoningTokens and reasoningText
so you can inspect the model's chain-of-thought when enabled.
Request (cURL) - Non-Streaming
curl -X POST https://app.eigenai.com/api/v1/generate \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-oss",
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Explain quantum computing in simple terms."}
],
"temperature": 0.7,
"reasoning_effort": "medium",
"max_tokens": 2048
}'
Request (Python) - Non-Streaming
import json
import requests
url = "https://app.eigenai.com/api/v1/generate"
headers = {
"Authorization": "Bearer YOUR_API_KEY",
"Content-Type": "application/json"
}
payload = {
"model": "gpt-oss",
"messages": [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "Explain quantum computing in simple terms."}
],
"temperature": 0.7,
"reasoning_effort": "medium",
"max_tokens": 2048
}
response = requests.post(url, headers=headers, json=payload)
response.raise_for_status()
print(json.dumps(response.json(), indent=2))
Request (JavaScript/Node.js) - Non-Streaming
import fetch from 'node-fetch';
const response = await fetch('https://app.eigenai.com/api/v1/generate', {
method: 'POST',
headers: {
'Authorization': 'Bearer YOUR_API_KEY',
'Content-Type': 'application/json'
},
body: JSON.stringify({
model: 'gpt-oss',
messages: [
{ role: 'system', content: 'You are a helpful assistant.' },
{ role: 'user', content: 'Explain quantum computing in simple terms.' }
],
temperature: 0.7,
reasoning_effort: 'medium',
max_tokens: 2048
})
});
if (!response.ok) {
throw new Error(`Request failed: ${response.status} ${await response.text()}`);
}
const result = await response.json();
console.log(result);
Parameters
| Parameter | Type | Required | Description |
|---|---|---|---|
model |
string | ✅ Yes | Use gpt-oss |
messages |
array | ✅ Yes | Conversation history with {role, content} objects |
temperature |
number | Optional | Controls creativity (0-2) |
max_tokens |
number | Optional | Maximum tokens in the response |
reasoning_effort |
string | Optional | Controls reasoning depth (low, medium, high). Default medium |
stream |
boolean | Optional | Use true for streaming (requires /api/v1/generate/stream). Omit or use false for complete responses |