import OpenAI from 'openai'
const HICAP_API_KEY = process.env.HICAP_API_KEY
const client = new OpenAI({
baseURL: 'https://api.hicap.ai/v1',
apiKey: HICAP_API_KEY,
defaultHeaders: { "api-key": HICAP_API_KEY },
})
const response = await client.chat.completions.create({
model: 'gpt-5.2',
messages: [
{ role: 'user', content: 'Hello, GPT!' }
],
})Hicap works with any OpenAI-compatible tool. Just change your base URL and start saving.
Any OpenAI-compatible app
Claude Code, Aider & more
Cline, Copilot & extensions
Full-time AI-assisted coding across projects
10 developers + product AI usage for rapid development
Large-scale AI development with high-volume usage
Full-time AI-assisted coding across projects
10 developers + product AI usage for rapid development
Large-scale AI development with high-volume usage
Swipe to see more examples
We buy reserved GPU capacity in bulk, then let you tap into it on-demand.
You get the speed of provisioned throughput at a fraction of the cost.
Access the same models at a fraction of pay-as-you-go pricing through reserved GPU capacity.
Provisioned throughput delivers consistent performance for your workloads. No cold starts, no throttling.
Browse the reserved capacity catalog—GPT-5.2, Claude 4 Sonnet, Gemini 3.0 Flash and more. Contact our team to get started.
View catalog →Track token usage, costs, and latency across all models. See exactly where your AI budget goes.
Works with curl, OpenAI SDK, or any OpenAI-compatible tool. Just change the base URL.
Your requests are load-balanced across multiple providers for redundancy and high availability.
Get dedicated GPU throughput, volume discounts, and priority support for your team. We'll tailor a plan to your usage.