How Atlas Cloud Works

Understand how Atlas Cloud connects you to 300+ AI models through a unified API

Architecture Overview

Atlas Cloud is an AI API aggregation platform that sits between your application and multiple AI model providers. Instead of integrating with each provider separately, you use a single Atlas Cloud API key and consistent API endpoints to access 300+ models from dozens of providers.

Your Application


  Atlas Cloud API  ──────  Unified authentication, billing, and monitoring

      ├── DeepSeek (V3, Coder)
      ├── Alibaba (Qwen, Qwen-Image)
      ├── ByteDance (Seedream, Seedance, Kling)
      ├── Black Forest Labs (FLUX)
      ├── MoonshotAI (Kimi)
      ├── MiniMax (Hailuo)
      ├── Luma AI (Video)
      ├── Zhipu AI (GLM)
      └── ... 20+ more providers

How Requests Work

Synchronous APIs (LLM / Chat)

LLM chat completions return responses synchronously, just like the OpenAI API:

  1. You send a POST request to /v1/chat/completions with your prompt
  2. Atlas Cloud routes the request to the selected model provider
  3. You receive the response directly (or via streaming chunks)
Client  →  POST /v1/chat/completions  →  Atlas Cloud  →  Model Provider
Client  ←  Response (text/stream)     ←  Atlas Cloud  ←  Model Provider

Asynchronous APIs (Image / Video Generation)

Image and video generation tasks run asynchronously because they take longer to process:

  1. You send a POST request to submit a generation task
  2. Atlas Cloud returns a predictionId immediately
  3. You poll the /api/v1/model/getResult endpoint to check task status
  4. Once completed, you receive the output URL(s)
Client  →  POST /api/v1/model/generateImage  →  Atlas Cloud
Client  ←  { predictionId: "abc123" }        ←  Atlas Cloud

Client  →  GET /api/v1/model/getResult?predictionId=abc123  →  Atlas Cloud
Client  ←  { status: "processing" }                         ←  Atlas Cloud

Client  →  GET /api/v1/model/getResult?predictionId=abc123  →  Atlas Cloud
Client  ←  { status: "completed", output: "https://..." }   ←  Atlas Cloud

For more details on polling, see Predictions.

API Endpoints Summary

EndpointMethodTypeDescription
/v1/chat/completionsPOSTSynchronousLLM chat (OpenAI-compatible)
/api/v1/model/generateImagePOSTAsynchronousImage generation
/api/v1/model/generateVideoPOSTAsynchronousVideo generation
/api/v1/model/uploadMediaPOSTSynchronousUpload files for generation tasks
/api/v1/model/getResultGETSynchronousGet async task results

Authentication

All API requests require an API key in the Authorization header:

Authorization: Bearer your-api-key

Get your API key from the Atlas Cloud Console. See the API Keys guide for details.

Key Benefits

One API, 300+ Models

No need to manage multiple provider accounts, API keys, or billing relationships. Atlas Cloud handles all provider integrations for you.

OpenAI SDK Compatible

The LLM API is fully compatible with the OpenAI SDK. Switch to Atlas Cloud by changing just two lines of code — the base URL and API key.

Optimized Infrastructure

Atlas Cloud's inference infrastructure is optimized for speed and reliability:

  • Image generation in under 5 seconds
  • Video generation in under 2 minutes
  • 99.9% API uptime

Unified Billing

One account, one balance, one invoice — regardless of how many models or providers you use. Monitor usage and costs in real-time from the Console.

Next Steps