Hailuo-02 Fast
Imagem para Vídeo

Hailuo 02 Fast API by MiniMax

minimax/hailuo-02/fast
Fast

Hailuo 02 is a new AI video generation model from Hailuo AI.

Entrada

Carregando configuração de parâmetros...

Saída

Inativo
Os vídeos gerados serão exibidos aqui
Configure os parâmetros e clique em executar para começar a gerar

Cada execução custará $0.1. Com $10 você pode executar aproximadamente 100 vezes.

Você pode continuar com:

Parâmetros

Exemplo de código

import requests
import time

# Step 1: Start video generation
generate_url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
    "model": "minimax/hailuo-02/fast",
    "prompt": "A beautiful sunset over the ocean with gentle waves",
    "width": 512,
    "height": 512,
    "duration": 3,
    "fps": 24,
}

generate_response = requests.post(generate_url, headers=headers, json=data)
generate_result = generate_response.json()
prediction_id = generate_result["data"]["id"]

# Step 2: Poll for result
poll_url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"

def check_status():
    while True:
        response = requests.get(poll_url, headers={"Authorization": "Bearer $ATLASCLOUD_API_KEY"})
        result = response.json()

        if result["data"]["status"] in ["completed", "succeeded"]:
            print("Generated video:", result["data"]["outputs"][0])
            return result["data"]["outputs"][0]
        elif result["data"]["status"] == "failed":
            raise Exception(result["data"]["error"] or "Generation failed")
        else:
            # Still processing, wait 2 seconds
            time.sleep(2)

video_url = check_status()

Instalar

Instale o pacote necessário para a sua linguagem de programação.

bash
pip install requests

Autenticação

Todas as solicitações de API requerem autenticação por meio de uma chave de API. Você pode obter sua chave de API no painel do Atlas Cloud.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

Cabeçalhos HTTP

python
import os

API_KEY = os.environ.get("ATLASCLOUD_API_KEY")
headers = {
    "Content-Type": "application/json",
    "Authorization": f"Bearer {API_KEY}"
}
Mantenha sua chave de API segura

Nunca exponha sua chave de API em código do lado do cliente ou repositórios públicos. Use variáveis de ambiente ou um proxy de backend.

Enviar uma solicitação

import requests

url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
    "model": "your-model",
    "prompt": "A beautiful landscape"
}

response = requests.post(url, headers=headers, json=data)
print(response.json())

Enviar uma solicitação

Envie uma solicitação de geração assíncrona. A API retorna um ID de predição que você pode usar para verificar o status e obter o resultado.

POST/api/v1/model/generateVideo

Corpo da solicitação

import requests

url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}

data = {
    "model": "minimax/hailuo-02/fast",
    "input": {
        "prompt": "A beautiful sunset over the ocean with gentle waves"
    }
}

response = requests.post(url, headers=headers, json=data)
result = response.json()

print(f"Prediction ID: {result['id']}")
print(f"Status: {result['status']}")

Resposta

{
  "id": "pred_abc123",
  "status": "processing",
  "model": "model-name",
  "created_at": "2025-01-01T00:00:00Z"
}

Verificar status

Consulte o endpoint de predição para verificar o status atual da sua solicitação.

GET/api/v1/model/prediction/{prediction_id}

Exemplo de polling

import requests
import time

prediction_id = "pred_abc123"
url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }

while True:
    response = requests.get(url, headers=headers)
    result = response.json()
    status = result["data"]["status"]
    print(f"Status: {status}")

    if status in ["completed", "succeeded"]:
        output_url = result["data"]["outputs"][0]
        print(f"Output URL: {output_url}")
        break
    elif status == "failed":
        print(f"Error: {result['data'].get('error', 'Unknown')}")
        break

    time.sleep(3)

Valores de status

processingA solicitação ainda está sendo processada.
completedA geração está completa. As saídas estão disponíveis.
succeededA geração foi bem-sucedida. As saídas estão disponíveis.
failedA geração falhou. Verifique o campo de erro.

Resposta concluída

{
  "data": {
    "id": "pred_abc123",
    "status": "completed",
    "outputs": [
      "https://storage.atlascloud.ai/outputs/result.mp4"
    ],
    "metrics": {
      "predict_time": 45.2
    },
    "created_at": "2025-01-01T00:00:00Z",
    "completed_at": "2025-01-01T00:00:10Z"
  }
}

Enviar arquivos

Envie arquivos para o armazenamento do Atlas Cloud e obtenha uma URL que pode ser usada nas suas solicitações de API. Use multipart/form-data para enviar.

POST/api/v1/model/uploadMedia

Exemplo de upload

import requests

url = "https://api.atlascloud.ai/api/v1/model/uploadMedia"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }

with open("image.png", "rb") as f:
    files = {"file": ("image.png", f, "image/png")}
    response = requests.post(url, headers=headers, files=files)

result = response.json()
download_url = result["data"]["download_url"]
print(f"File URL: {download_url}")

Resposta

{
  "data": {
    "download_url": "https://storage.atlascloud.ai/uploads/abc123/image.png",
    "file_name": "image.png",
    "content_type": "image/png",
    "size": 1024000
  }
}

Schema de entrada

Os seguintes parâmetros são aceitos no corpo da solicitação.

Total: 0Obrigatório: 0Opcional: 0

Nenhum parâmetro disponível.

Exemplo de corpo da solicitação

json
{
  "model": "minimax/hailuo-02/fast"
}

Schema de saída

A API retorna uma resposta de predição com as URL de saída geradas.

idstringrequired
Unique identifier for the prediction.
statusstringrequired
Current status of the prediction.
processingcompletedsucceededfailed
modelstringrequired
The model used for generation.
outputsarray[string]
Array of output URLs. Available when status is "completed".
errorstring
Error message if status is "failed".
metricsobject
Performance metrics.
predict_timenumber
Time taken for video generation in seconds.
created_atstringrequired
ISO 8601 timestamp when the prediction was created.
Format: date-time
completed_atstring
ISO 8601 timestamp when the prediction was completed.
Format: date-time

Exemplo de resposta

json
{
  "id": "pred_abc123",
  "status": "completed",
  "model": "model-name",
  "outputs": [
    "https://storage.atlascloud.ai/outputs/result.mp4"
  ],
  "metrics": {
    "predict_time": 45.2
  },
  "created_at": "2025-01-01T00:00:00Z",
  "completed_at": "2025-01-01T00:00:10Z"
}

Atlas Cloud Skills

O Atlas Cloud Skills integra mais de 300 modelos de IA diretamente no seu assistente de codificação com IA. Um comando para instalar e depois use linguagem natural para gerar imagens, vídeos e conversar com LLM.

Clientes compatíveis

Claude Code
OpenAI Codex
Gemini CLI
Cursor
Windsurf
VS Code
Trae
GitHub Copilot
Cline
Roo Code
Amp
Goose
Replit
40+ clientes compatíveis

Instalar

bash
npx skills add AtlasCloudAI/atlas-cloud-skills

Configurar chave de API

Obtenha sua chave de API no painel do Atlas Cloud e defina-a como variável de ambiente.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

Funcionalidades

Após a instalação, você pode usar linguagem natural no seu assistente de IA para acessar todos os modelos do Atlas Cloud.

Geração de imagensGere imagens com modelos como Nano Banana 2, Z-Image e mais.
Criação de vídeosCrie vídeos a partir de texto ou imagens com Kling, Vidu, Veo, etc.
Chat com LLMConverse com Qwen, DeepSeek e outros modelos de linguagem de grande escala.
Upload de mídiaEnvie arquivos locais para fluxos de trabalho de edição de imagens e imagem para vídeo.

MCP Server

O Atlas Cloud MCP Server conecta seu IDE com mais de 300 modelos de IA através do Model Context Protocol. Funciona com qualquer cliente compatível com MCP.

Clientes compatíveis

Cursor
VS Code
Windsurf
Claude Code
OpenAI Codex
Gemini CLI
Cline
Roo Code
100+ clientes compatíveis

Instalar

bash
npx -y atlascloud-mcp

Configuração

Adicione a seguinte configuração ao arquivo de configuração de MCP do seu IDE.

json
{
  "mcpServers": {
    "atlascloud": {
      "command": "npx",
      "args": [
        "-y",
        "atlascloud-mcp"
      ],
      "env": {
        "ATLASCLOUD_API_KEY": "your-api-key-here"
      }
    }
  }
}

Ferramentas disponíveis

atlas_generate_imageGere imagens a partir de prompts de texto.
atlas_generate_videoCrie vídeos a partir de texto ou imagens.
atlas_chatConverse com modelos de linguagem de grande escala.
atlas_list_modelsExplore mais de 300 modelos de IA disponíveis.
atlas_quick_generateCriação de conteúdo em uma etapa com seleção automática de modelo.
atlas_upload_mediaEnvie arquivos locais para fluxos de trabalho de API.

API Schema

Schema não disponível

Faça login para ver o histórico de solicitações

Você precisa fazer login para acessar o histórico de solicitações do modelo.

Fazer Login

MiniMax Hailuo 02 Fast

This is a fast version of Hailuo 02 that can generate videos in 6s and 10s at 512p resolution.

Hailuo 02 is a new AI video generation model, from Hailuo AI, created on MiniMax's evolving framework. It has been fine-tuned to deliver ultra-clear 1080P resolution and unprecedented responsiveness while even handling, the craziest of physics driven scenes.

Indeed, artists have discovered that for highly intricate scenarios, such as gymnastics, MiniMax Hailuo 02 is currently the only model globally capable of delivering such performance. We eagerly invite the community to explore and unlock even more creative possibilities.

Our journey began late last August when we informally launched a demo webpage showcasing an early version of our video generation model. To our surprise, it attracted significant attention and acclaim from talented creators worldwide. This pivotal moment led to the development of Hailuo Video 01, our AI native video generation product, which has since empowered creators to generate over 370 million videos globally.

Returning to our foundational principle of "Intelligence with Everyone," our ambition is to equip global creators to fully unleash their imagination, elevate the quality of their video content, and lower the barriers to video creation. Crucially, we strive to achieve this without imposing prohibitive costs that would limit the widespread accessibility of this technology.

To this end, our team embarked on a quest to develop a more efficient video generation model architecture. This pursuit culminated in the core framework of MiniMax Hailuo 02, which we've named Noise-aware Compute Redistribution (NCR). In essence, the new architecture's central idea is as follows:

At a comparable parameter scale, the new architecture boosts our training and inference efficiency by 2.5 times. This significant gain enables us to implement a much larger parameter model—thereby enhancing its expressive capabilities—without increasing costs for creators. This approach also leaves ample room for inference optimization. We ultimately expanded the model's total parameter count to 3 times that of its predecessor.

A larger parameter count and heightened training efficiency mean our model can learn from a more extensive dataset. The wealth of feedback from Hailuo 01 provided invaluable guidance for our model training strategy. As a result, we expanded our training data volume by 4 times, achieving significant improvements in data quality and diversity.

With this architectural innovation, combined with a threefold increase in parameters and four times the training data, our model has taken a significant leap forward, particularly in its adherence to complex instructions and its rendering of extreme physics. The new model accurately interprets and executes highly detailed prompts, delivering more precise outputs. Furthermore, the efficiency gains from the new architecture also mean we can offer native 1080p video generation at a very affordable price point.

An early iteration of this model was tested by users on the Artificial Analysis Video Arena, where it secured the second position globally. Stay tuned for an upcoming new version!

These model enhancements are now fully integrated into the Hailuo Video web platform, mobile application, and our API platform. We currently offer three distinct versions: 768p-6s, 768p-10s, and 1080p-6s. True to our commitment, and thanks to the aforementioned architectural innovation, we continue to offer creators and developers the most open access and affordable pricing in the industry.

Through sustained technological research and development, coupled with deep collaborations with creators, developers, and artists, our mission and strategic direction have become ever clearer.

MiniMax Hailuo 02 represents a new milestone, and we are poised for rapid advancements in the following areas:

  • Enhancing generation speed

  • Improving alignment, leading to higher generation success rates and improved stability

  • Advancing model features beyond Text-to-Video (T2V) and Image-to-Video (I2V)

And, as always, we remain steadfast in our commitment to relentlessly exploring the upper limits of what technology and art can achieve together.

Mais de 300 Modelos, Comece Agora,

Explorar Todos os Modelos

Join our Discord community

Join the Discord community for the latest model updates, prompts, and support.