
Wan 2.6 Spicy Image-to-Video API by Atlas Cloud
AtlasCloud Wan 2.6 Spicy Image-to-Video turns a reference image into a short motion clip with expressive character movement and stable temporal detail.
INPUT
OUTPUT
IdleYour request will cost $0.07 per run. For $10 you can run this model approximately 142 times.
Here's what you can do next:
Code Example
import requests
import time
# Step 1: Start video generation
generate_url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
"Content-Type": "application/json",
"Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
"model": "atlascloud/wan-2.6-spicy/image-to-video",
"prompt": "A beautiful sunset over the ocean with gentle waves",
"width": 512,
"height": 512,
"duration": 3,
"fps": 24,
}
generate_response = requests.post(generate_url, headers=headers, json=data)
generate_result = generate_response.json()
prediction_id = generate_result["data"]["id"]
# Step 2: Poll for result
poll_url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"
def check_status():
while True:
response = requests.get(poll_url, headers={"Authorization": "Bearer $ATLASCLOUD_API_KEY"})
result = response.json()
if result["data"]["status"] in ["completed", "succeeded"]:
print("Generated video:", result["data"]["outputs"][0])
return result["data"]["outputs"][0]
elif result["data"]["status"] == "failed":
raise Exception(result["data"]["error"] or "Generation failed")
else:
# Still processing, wait 2 seconds
time.sleep(2)
video_url = check_status()Install
Install the required package for your language.
pip install requestsAuthentication
All API requests require authentication via an API key. You can get your API key from the Atlas Cloud dashboard.
export ATLASCLOUD_API_KEY="your-api-key-here"HTTP Headers
import os
API_KEY = os.environ.get("ATLASCLOUD_API_KEY")
headers = {
"Content-Type": "application/json",
"Authorization": f"Bearer {API_KEY}"
}Never expose your API key in client-side code or public repositories. Use environment variables or a backend proxy instead.
Submit a request
import requests
url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
"Content-Type": "application/json",
"Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
"model": "your-model",
"prompt": "A beautiful landscape"
}
response = requests.post(url, headers=headers, json=data)
print(response.json())Submit a Request
Submit an asynchronous generation request. The API returns a prediction ID that you can use to check the status and retrieve the result.
/api/v1/model/generateVideoRequest Body
import requests
url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
"Content-Type": "application/json",
"Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
"model": "atlascloud/wan-2.6-spicy/image-to-video",
"input": {
"prompt": "A beautiful sunset over the ocean with gentle waves"
}
}
response = requests.post(url, headers=headers, json=data)
result = response.json()
print(f"Prediction ID: {result['id']}")
print(f"Status: {result['status']}")Response
{
"id": "pred_abc123",
"status": "processing",
"model": "model-name",
"created_at": "2025-01-01T00:00:00Z"
}Check Status
Poll the prediction endpoint to check the current status of your request.
/api/v1/model/prediction/{prediction_id}Polling Example
import requests
import time
prediction_id = "pred_abc123"
url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }
while True:
response = requests.get(url, headers=headers)
result = response.json()
status = result["data"]["status"]
print(f"Status: {status}")
if status in ["completed", "succeeded"]:
output_url = result["data"]["outputs"][0]
print(f"Output URL: {output_url}")
break
elif status == "failed":
print(f"Error: {result['data'].get('error', 'Unknown')}")
break
time.sleep(3)Status Values
processingThe request is still being processed.completedGeneration is complete. Outputs are available.succeededGeneration succeeded. Outputs are available.failedGeneration failed. Check the error field.Completed Response
{
"data": {
"id": "pred_abc123",
"status": "completed",
"outputs": [
"https://storage.atlascloud.ai/outputs/result.mp4"
],
"metrics": {
"predict_time": 45.2
},
"created_at": "2025-01-01T00:00:00Z",
"completed_at": "2025-01-01T00:00:10Z"
}
}Upload Files
Upload files to Atlas Cloud storage and get a URL you can use in your API requests. Use multipart/form-data to upload.
/api/v1/model/uploadMediaUpload Example
import requests
url = "https://api.atlascloud.ai/api/v1/model/uploadMedia"
headers = { "Authorization": "Bearer $ATLASCLOUD_API_KEY" }
with open("image.png", "rb") as f:
files = {"file": ("image.png", f, "image/png")}
response = requests.post(url, headers=headers, files=files)
result = response.json()
download_url = result["data"]["download_url"]
print(f"File URL: {download_url}")Response
{
"data": {
"download_url": "https://storage.atlascloud.ai/uploads/abc123/image.png",
"file_name": "image.png",
"content_type": "image/png",
"size": 1024000
}
}Input Schema
The following parameters are accepted in the request body.
No parameters available.
Example Request Body
{
"model": "atlascloud/wan-2.6-spicy/image-to-video"
}Output Schema
The API returns a prediction response with the generated output URLs.
Example Response
{
"id": "pred_abc123",
"status": "completed",
"model": "model-name",
"outputs": [
"https://storage.atlascloud.ai/outputs/result.mp4"
],
"metrics": {
"predict_time": 45.2
},
"created_at": "2025-01-01T00:00:00Z",
"completed_at": "2025-01-01T00:00:10Z"
}Atlas Cloud Skills
Atlas Cloud Skills integrates 300+ AI models directly into your AI coding assistant. One command to install, then use natural language to generate images, videos, and chat with LLMs.
Supported Clients
Install
npx skills add AtlasCloudAI/atlas-cloud-skillsSetup API Key
Get your API key from the Atlas Cloud dashboard and set it as an environment variable.
export ATLASCLOUD_API_KEY="your-api-key-here"Capabilities
Once installed, you can use natural language in your AI assistant to access all Atlas Cloud models.
MCP Server
Atlas Cloud MCP Server connects your IDE with 300+ AI models via the Model Context Protocol. Works with any MCP-compatible client.
Supported Clients
Install
npx -y atlascloud-mcpConfiguration
Add the following configuration to your IDE's MCP settings file.
{
"mcpServers": {
"atlascloud": {
"command": "npx",
"args": [
"-y",
"atlascloud-mcp"
],
"env": {
"ATLASCLOUD_API_KEY": "your-api-key-here"
}
}
}
}Available Tools
API Schema
Schema not availablePlease log in to view request history
You need to be logged in to access your model request history.
Log InWan 2.6Professional Multi-Shot AI Video Creation
Alibaba's latest breakthrough in AI video generation. Create up to 15-second 1080p videos with multi-shot storytelling, reference-driven character consistency, and native audio-visual synchronization. The first model to truly understand storyboard logic for cinematic narratives.
Revolutionary Breakthroughs
What makes Wan 2.6 the game-changer in AI video generation
Multi-Shot Storytelling
First model to understand storyboard logic. Automatically generates sequential shots with coherent transitions, maintaining character appearance and environment consistency across scene changes—enabling complete story arcs in a single 15-second generation.
Reference-to-Video (R2V)
Upload a 2-30 second reference video to extract and preserve character appearance, movement patterns, and voice characteristics. Create consistent character performances across multiple videos with unprecedented accuracy.
Accurate Text Rendering
Industry-leading text rendering capabilities for product packaging, signage, and branded content. Generate clear, readable text within video frames—essential for marketing and commercial applications.
Core Capabilities
Extended 15-Second Duration
Generate up to 15 seconds per video with complete "Three Act" structure (Setup → Action → Resolution)
Professional 1080p Quality
Native 1080p output at 24fps with cinematic quality and enhanced visual stability
Native Audio Sync
Dialogue matches lip movements, background music aligns with pacing, sound effects trigger perfectly
Character Consistency
Maintain character appearance, costumes, and identity across shots and multiple videos
Cinematic Camera Control
Professional camera movements including pans, zooms, tracking shots, and dolly movements
Flexible Aspect Ratios
16:9 (YouTube), 9:16 (Reels), 1:1 (Square) - platform-optimized without post-production cropping
Wan 2.6 vs Wan 2.5: Major Improvements
See what's new in the latest release
Three Specialized Generation Modes
Choose the right mode for your creative workflow
Text-to-Video (T2V)
Most PopularGenerate complete videos from text prompts with enhanced multi-shot segmentation and improved prompt handling. Perfect for storytelling and creative exploration.
- Automatic shot segmentation from single prompt
- Multi-character interaction understanding
- Camera movement and emotional cues
- Environmental detail preservation
Image-to-Video (I2V)
EnhancedTransform still images into motion videos with improved motion coherence. Ideal for product showcases, photo animation, and visual storytelling.
- Precise text rendering for products
- Style consistency across frames
- Natural motion from static images
- Narrative-driven visual optimization
Reference-to-Video (R2V)
NEWUpload a reference video (2-30s) to preserve character appearance, movement patterns, and voice. Strongest consistency guarantee for character-driven content.
- Full character identity preservation
- Voice characteristics extraction
- Movement pattern replication
- Multi-character co-acting scenes
Perfect For
Marketing & Advertising
Product demos with text rendering, brand campaigns with character consistency, and promotional videos
Content Creation
YouTube videos, social media reels, multi-shot storytelling, and video editing workflows
E-commerce
Product showcases with accurate text, tutorial videos, and customer testimonial recreation
Education & Training
Instructional content, course materials, and multi-scene educational narratives
Entertainment
Short films, character-driven stories, cinematic sequences, and creative experiments
Pre-visualization
Film concept development, storyboard creation, and scene planning for productions
Wan 2.6 T2V, I2V, and R2V API Integration
Complete API suite for Text-to-Video, Image-to-Video, and Reference-to-Video generation
Text-to-Video API (T2V API)
Our Wan 2.6 T2V API transforms text prompts into multi-shot cinematic videos with automatic scene segmentation. Generate professional 1080p videos up to 15 seconds with native audio sync.
Image-to-Video API (I2V API)
Our Wan 2.6 I2V API brings still images to life with precise motion control and text rendering. Perfect for product videos, photo animation, and branded content creation.
Reference-to-Video API (R2V API)
Our Wan 2.6 R2V API preserves character identity from reference videos. Upload 2-30 second clips to extract appearance, voice, and movement patterns for consistent character generation.
Complete API Suite
All three Wan 2.6 API modes (T2V API, I2V API, R2V API) support RESTful architecture with comprehensive documentation. Get started with SDKs for Python, Node.js, and more. Each endpoint includes native audio-visual synchronization and full commercial usage rights.
How to Get Started with Wan 2.6
Start creating professional videos in minutes with two simple paths
API Integration
For developers building applications
Sign Up & Login
Create your Atlas Cloud account or login to access the console
Add Payment Method
Bind your credit card in the Billing section to fund your account
Generate API Key
Navigate to Console → API Keys and create your authentication key
Start Building
Use T2V, I2V, or R2V API endpoints to integrate Wan 2.6 into your application
Playground Experience
For quick testing and experimentation
Sign Up & Login
Create your Atlas Cloud account or login to access the platform
Add Payment Method
Bind your credit card in the Billing section to get started
Use Playground
Go to the Wan 2.6 playground, choose T2V/I2V/R2V mode, and generate videos instantly
Frequently Asked Questions
What makes Wan 2.6's multi-shot capability unique?
Wan 2.6 is the first model to truly understand storyboard logic. Unlike Wan 2.5 which created messy "morphing" effects, Wan 2.6 can automatically segment a single prompt into multiple distinct shots with coherent transitions, maintaining character consistency across scene changes.
How does Reference-to-Video (R2V) work?
Upload a 2-30 second reference video, and Wan 2.6 extracts the character's appearance, movement patterns, and voice characteristics. You can then generate new videos featuring the same character with consistent identity—ideal for creating character-driven content series.
What video formats and durations are supported?
Wan 2.6 generates 1080p videos at 24fps with durations from 5 to 15 seconds. Supported aspect ratios include 16:9 (YouTube), 9:16 (Instagram Reels/TikTok), and 1:1 (square format), optimized for each platform without requiring post-production cropping.
Can Wan 2.6 render text in videos?
Yes! Wan 2.6 features industry-leading text rendering for product packaging, signage, and branded content. The model can generate clear, readable text within video frames—a critical feature that Seedance and most competitors lack.
What's the difference between T2V, I2V, and R2V modes?
T2V (Text-to-Video) generates from text prompts with multi-shot capability. I2V (Image-to-Video) animates still images with precise text rendering. R2V (Reference-to-Video) uses video references to preserve character identity across generations. Choose based on your input type and consistency needs.
Do I have commercial rights to generated videos?
Yes! Every Wan 2.6 creation comes with full commercial usage rights. Videos are production-ready for marketing campaigns, client deliverables, branded content, and commercial applications without additional licensing requirements.
Why Use Wan 2.6 on Atlas Cloud?
Leverage enterprise-grade infrastructure for your professional video generation workflows
Purpose-Built Infrastructure
Deploy Wan 2.6's multi-shot generation and R2V capabilities on infrastructure specifically optimized for demanding AI video workloads. Maximum performance for 1080p 15-second generation.
Unified API for All Models
Access Wan 2.6 (T2V, I2V, R2V) alongside 300+ AI models (LLMs, image, video, audio) through one unified API. Single integration for all your generative AI needs with consistent auth.
Competitive Pricing
Save up to 70% compared to AWS with transparent, pay-as-you-go pricing. No hidden fees, no commitments—scale from prototype to production without breaking the bank.
SOC I & II Certified Security
Your reference videos and generated content protected with SOC I & II certifications and HIPAA compliance. Enterprise-grade security with encrypted transmission and storage.
99.9% Uptime SLA
Enterprise-grade reliability with guaranteed 99.9% uptime. Your Wan 2.6 multi-shot video generation is always available for production campaigns and critical content workflows.
Easy Integration
Complete integration in minutes with REST API and multi-language SDKs (Python, Node.js, Go). Switch between T2V, I2V, and R2V modes seamlessly with unified endpoint structure.
Technical Specifications
Experience Professional Multi-Shot Video Generation
Join content creators, marketers, and filmmakers worldwide who are revolutionizing video production with Wan 2.6's groundbreaking multi-shot storytelling and character consistency capabilities.
Wan 2.6 Spicy Image-to-Video
Wan 2.6 Spicy Image-to-Video turns a first-frame image into a short motion clip with expressive character movement and stable temporal detail. This AtlasCloud variant uses a dedicated Wan 2.6 image-to-video LoRA deployment for a more stylized motion profile.
Highlights
- First-frame image-to-video: Use one starting image plus a text prompt to control movement and camera direction.
- 720p and 1080p output: 480p is not exposed because the underlying deployment does not support it.
- Short-form generation: Supports 5s, 10s, and 15s clips.
- Optional audio control: Provide an audio URL to guide motion, or disable generated audio for silent output.
- Negative prompt support: Add optional constraints to reduce blur, distortion, or unwanted artifacts.
Parameters
| Parameter | Required | Description |
|---|---|---|
model | Yes | atlascloud/wan-2.6-spicy/image-to-video |
prompt | Yes | Text prompt describing the desired motion. |
image | Yes | First-frame image URL or Base64 image. |
audio | No | Audio URL to guide the generated motion. |
negative_prompt | No | Text describing what to avoid. |
resolution | Yes | 720p or 1080p. |
duration | No | 5, 10, or 15 seconds. Defaults to 5. |
enable_prompt_expansion | No | Enable upstream prompt expansion. Defaults to false. |
shot_type | No | single or multi. Multi-shot mode requires prompt expansion. Defaults to single. |
generate_audio | No | Whether to include generated audio. Defaults to true; set false for silent output. |
seed | No | Random seed. -1 means random. |
How To Use
curl -X POST "https://api.atlascloud.ai/api/v1/model/generateVideo" \ -H "Authorization: Bearer $AIP_API_KEY" \ -H "Content-Type: application/json" \ --data-raw '{ "model": "atlascloud/wan-2.6-spicy/image-to-video", "prompt": "The woman turns toward the camera with a confident smile, hair moving naturally as the camera slowly pushes in.", "image": "https://static.atlascloud.ai/media/images/db548fe3bd5cafa4ef7e0141d69c8566.jpeg", "negative_prompt": "blurry, low quality, distorted hands, extra limbs", "duration": 5, "resolution": "720p", "generate_audio": true, "seed": -1 }'
Pricing
Pricing matches Wan 2.6 Image-to-Video final user pricing.
| Resolution | Multiplier | 5s Price | 10s Price | 15s Price |
|---|---|---|---|---|
| 720p | 1.0x | $0.35 | $0.70 | $1.05 |
| 1080p | 1.5x | $0.525 | $1.05 | $1.575 |
Formula:
effective_rate * max(5, duration) * (resolution == "1080p" || resolution == "1080P" ? 1.5 : 1)
effective_rate = $0.0700/s for 720p. 1080p uses a 1.5x multiplier.
Notes
- This model is allowlist-enabled. Contact AtlasCloud if it is not visible or callable from your account.
- 480p is not exposed for this model.
- This endpoint uses the input image as the first frame of the generated video.
shot_type: "multi"requiresenable_prompt_expansion: true.- The underlying deployment accepts
720Pand1080P; AtlasCloud normalizes API input automatically. - Generation is asynchronous. Poll
/api/v1/model/prediction/{request_id}for the final video URL.






