bytedance/seedance-v1-lite-t2v-480p

An efficient text-to-video model geared toward fast, cost-effective generation. Ideal for prototyping short narrative clips (5–10 s) with stylistic flexibility and prompt-faithful motion.

TEXT-TO-VIDEONEW
Seedance v1 Lite t2v 480p
text-to-video

An efficient text-to-video model geared toward fast, cost-effective generation. Ideal for prototyping short narrative clips (5–10 s) with stylistic flexibility and prompt-faithful motion.

INPUT

Loading parameter configuration...

OUTPUT

Idle
Your generated videos will appear here
Configure your settings and click Run to get started

Your request will cost $0.014 per run. For $10 you can run this model approximately 714 times.

Here's what you can do next:

Parameters

Code Example

import requests
import time

# Step 1: Start video generation
generate_url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
    "model": "bytedance/seedance-v1-lite-t2v-480p",
    "prompt": "A beautiful sunset over the ocean with gentle waves",
    "width": 512,
    "height": 512,
    "duration": 3,
    "fps": 24,
}

generate_response = requests.post(generate_url, headers=headers, json=data)
generate_result = generate_response.json()
prediction_id = generate_result["data"]["id"]

# Step 2: Poll for result
poll_url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"

def check_status():
    while True:
        response = requests.get(poll_url, headers={"Authorization": "Bearer $ATLASCLOUD_API_KEY"})
        result = response.json()

        if result["data"]["status"] in ["completed", "succeeded"]:
            print("Generated video:", result["data"]["outputs"][0])
            return result["data"]["outputs"][0]
        elif result["data"]["status"] == "failed":
            raise Exception(result["data"]["error"] or "Generation failed")
        else:
            # Still processing, wait 2 seconds
            time.sleep(2)

video_url = check_status()

Install

Install the required package for your language.

bash
pip install requests

Authentication

All API requests require authentication via an API key. You can get your API key from the Atlas Cloud dashboard.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

HTTP Headers

python
import os

API_KEY = os.environ.get("ATLASCLOUD_API_KEY")
headers = {
    "Content-Type": "application/json",
    "Authorization": f"Bearer {API_KEY}"
}
Keep your API key secure

Never expose your API key in client-side code or public repositories. Use environment variables or a backend proxy instead.

Submit a request

import requests

url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
    "model": "your-model",
    "prompt": "A beautiful landscape"
}

response = requests.post(url, headers=headers, json=data)
print(response.json())

Submit a Request

Submit an asynchronous generation request. The API returns a prediction ID that you can use to check the status and retrieve the result.

POST/api/v1/model/generateVideo

Request Body

import requests

url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}

data = {
    "model": "bytedance/seedance-v1-lite-t2v-480p",
    "input": {
        "prompt": "A beautiful sunset over the ocean with gentle waves"
    }
}

response = requests.post(url, headers=headers, json=data)
result = response.json()

print(f"Prediction ID: {result['id']}")
print(f"Status: {result['status']}")

Response

{
  "id": "pred_abc123",
  "status": "processing",
  "model": "model-name",
  "created_at": "2025-01-01T00:00:00Z"
}

Check Status

Poll the prediction endpoint to check the current status of your request.

GET/api/v1/model/prediction/{prediction_id}

Polling Example

import requests
import time

prediction_id = "pred_abc123"
url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }

while True:
    response = requests.get(url, headers=headers)
    result = response.json()
    status = result["data"]["status"]
    print(f"Status: {status}")

    if status in ["completed", "succeeded"]:
        output_url = result["data"]["outputs"][0]
        print(f"Output URL: {output_url}")
        break
    elif status == "failed":
        print(f"Error: {result['data'].get('error', 'Unknown')}")
        break

    time.sleep(3)

Status Values

processingThe request is still being processed.
completedGeneration is complete. Outputs are available.
succeededGeneration succeeded. Outputs are available.
failedGeneration failed. Check the error field.

Completed Response

{
  "data": {
    "id": "pred_abc123",
    "status": "completed",
    "outputs": [
      "https://storage.atlascloud.ai/outputs/result.mp4"
    ],
    "metrics": {
      "predict_time": 45.2
    },
    "created_at": "2025-01-01T00:00:00Z",
    "completed_at": "2025-01-01T00:00:10Z"
  }
}

Upload Files

Upload files to Atlas Cloud storage and get a URL you can use in your API requests. Use multipart/form-data to upload.

POST/api/v1/model/uploadMedia

Upload Example

import requests

url = "https://api.atlascloud.ai/api/v1/model/uploadMedia"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }

with open("image.png", "rb") as f:
    files = {"file": ("image.png", f, "image/png")}
    response = requests.post(url, headers=headers, files=files)

result = response.json()
download_url = result["data"]["download_url"]
print(f"File URL: {download_url}")

Response

{
  "data": {
    "download_url": "https://storage.atlascloud.ai/uploads/abc123/image.png",
    "file_name": "image.png",
    "content_type": "image/png",
    "size": 1024000
  }
}

Input Schema

The following parameters are accepted in the request body.

Total: 0Required: 0Optional: 0

No parameters available.

Example Request Body

json
{
  "model": "bytedance/seedance-v1-lite-t2v-480p"
}

Output Schema

The API returns a prediction response with the generated output URLs.

idstringrequired
Unique identifier for the prediction.
statusstringrequired
Current status of the prediction.
processingcompletedsucceededfailed
modelstringrequired
The model used for generation.
outputsarray[string]
Array of output URLs. Available when status is "completed".
errorstring
Error message if status is "failed".
metricsobject
Performance metrics.
predict_timenumber
Time taken for video generation in seconds.
created_atstringrequired
ISO 8601 timestamp when the prediction was created.
Format: date-time
completed_atstring
ISO 8601 timestamp when the prediction was completed.
Format: date-time

Example Response

json
{
  "id": "pred_abc123",
  "status": "completed",
  "model": "model-name",
  "outputs": [
    "https://storage.atlascloud.ai/outputs/result.mp4"
  ],
  "metrics": {
    "predict_time": 45.2
  },
  "created_at": "2025-01-01T00:00:00Z",
  "completed_at": "2025-01-01T00:00:10Z"
}

Atlas Cloud Skills

Atlas Cloud Skills integrates 300+ AI models directly into your AI coding assistant. One command to install, then use natural language to generate images, videos, and chat with LLMs.

Supported Clients

Claude Code
OpenAI Codex
Gemini CLI
Cursor
Windsurf
VS Code
Trae
GitHub Copilot
Cline
Roo Code
Amp
Goose
Replit
40+ supported clients

Install

bash
npx skills add AtlasCloudAI/atlas-cloud-skills

Setup API Key

Get your API key from the Atlas Cloud dashboard and set it as an environment variable.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

Capabilities

Once installed, you can use natural language in your AI assistant to access all Atlas Cloud models.

Image GenerationGenerate images with models like Nano Banana 2, Z-Image, and more.
Video CreationCreate videos from text or images with Kling, Vidu, Veo, etc.
LLM ChatChat with Qwen, DeepSeek, and other large language models.
Media UploadUpload local files for image editing and image-to-video workflows.

MCP Server

Atlas Cloud MCP Server connects your IDE with 300+ AI models via the Model Context Protocol. Works with any MCP-compatible client.

Supported Clients

Cursor
VS Code
Windsurf
Claude Code
OpenAI Codex
Gemini CLI
Cline
Roo Code
100+ supported clients

Install

bash
npx -y atlascloud-mcp

Configuration

Add the following configuration to your IDE's MCP settings file.

json
{
  "mcpServers": {
    "atlascloud": {
      "command": "npx",
      "args": [
        "-y",
        "atlascloud-mcp"
      ],
      "env": {
        "ATLASCLOUD_API_KEY": "your-api-key-here"
      }
    }
  }
}

Available Tools

atlas_generate_imageGenerate images from text prompts.
atlas_generate_videoCreate videos from text or images.
atlas_chatChat with large language models.
atlas_list_modelsBrowse 300+ available AI models.
atlas_quick_generateOne-step content creation with auto model selection.
atlas_upload_mediaUpload local files for API workflows.

API Schema

Schema not available

Please log in to view request history

You need to be logged in to access your model request history.

Log In

ByteDance Seedance Lite T2V 480p

ByteDance Seedance Lite T2V 480p is an optimized AI text-to-video generation model developed by ByteDance, now available on WaveSpeedAI. This cost-effective model transforms text prompts into dynamic 5-10 second videos at 480p resolution with efficient processing speed, offering quality visual outputs with enhanced motion and semantic understanding. Part of the Seedance Lite model family, this model delivers excellent performance in text-to-video synthesis at an affordable price point.

Key Features

  • Fast Video Generation: Efficient processing creates 5-10 second videos at 480p resolution with vivid details and smooth motion.
  • Quality Motion Rendering: Advanced dynamic rendering techniques create natural and realistic movements that bring text descriptions to life.
  • Enhanced Semantic Understanding: AI excels in interpreting text prompts to generate coherent and dynamic scenes with professional quality.
  • Realistic Physical Simulation: Physics engine simulates realistic physical properties and movements for lifelike video generation.
  • Optimized Processing: Balanced for speed and quality, allowing efficient creation of high-quality videos.
  • Flexible Parameter Control: Customizable settings offer duration (5-10s), aspect ratio and style adjustments for creative control.
  • Professional Quality Output: Lite technology ensures quality results with smooth temporal consistency.
  • Cost-Effective Creation: Transform text descriptions into engaging, dynamic content at an affordable price.

Technical Excellence & Optimization

  • Lite Architecture: Built on optimized research delivering excellent text-to-video generation performance at reduced cost.
  • Efficient Processing: ByteDance technology delivers superior generation efficiency compared to standard T2V solutions.
  • Balanced Performance: Optimized video synthesis enables efficient creative workflows and content creation.
  • High-Performance Computing: Optimized infrastructure supports efficient, high-volume video generation.

Perfect for Budget-Conscious Creative Work

  • Content Creators: Transform text ideas into engaging video content for social platforms at affordable rates.
  • Marketing Professionals: Create dynamic promotional videos from text descriptions cost-effectively.
  • Social Media Managers: Convert text concepts into shareable, dynamic content that captures attention without breaking budgets.
  • E-commerce Teams: Generate product demonstration videos from text descriptions using cost-effective T2V technology.
  • Small Businesses: Deliver professional video content with budget-friendly video generation capabilities.

Performance & Efficiency Advantages

  • Quality Results: Generate professional videos efficiently with optimized processing speed.
  • Cost-Effective Workflow: Advanced AI enables creative iteration and content production at reduced costs.
  • Scalable Efficiency: Handle multiple text-to-video conversions efficiently without performance degradation.
  • Optimized Value: Advanced algorithms maximize quality while maintaining cost-effectiveness.

Limitations

  • Creative Focus: Designed primarily for creative video synthesis; not intended for generating factually accurate content.
  • Inherent Biases: Outputs may reflect biases present in the training data, typical of current models.
  • Input Sensitivity: The quality and consistency of generated videos depend significantly on the quality of the input text prompt; subtle variations may lead to output variability.
  • Resolution Limitation: This model is optimized for 480p video generation and does not support higher resolutions.
  • Lite Performance: While cost-effective, may have slightly reduced capabilities compared to Pro versions.

Out-of-Scope Use

The model and its derivatives may not be used in any way that violates applicable national, federal, state, local, or international law or regulation, including but not limited to:

  • Exploiting, harming, or attempting to exploit or harm minors, including solicitation, creation, acquisition, or dissemination of child exploitative content.
  • Generating or disseminating verifiably false information with the intent to harm others.
  • Creating or distributing personal identifiable information that could be used to harm an individual.
  • Harassing, abusing, threatening, stalking, or bullying individuals or groups.
  • Producing non-consensual nudity or illegal pornographic content.
  • Making fully automated decisions that adversely affect an individual's legal rights or create binding obligations.
  • Facilitating large-scale disinformation campaigns.

Start From 300+ Models,

Explore all models

Join our Discord community

Join the Discord community for the latest model updates, prompts, and support.