Kling Omni Video O3 Video-Edit (Standard) enables natural-language video edits: remove or replace objects, change backgrounds, add effects, and more. Video duration limited to 10s.
Kling Omni Video O3 Video-Edit (Standard) enables natural-language video edits: remove or replace objects, change backgrounds, add effects, and more. Video duration limited to 10s.
Your request will cost 0.238 per run. For $10 you can run this model approximately 42 times.
Here's what you can do next:
import requests
import time
# Step 1: Start video generation
generate_url = "https://api.atlascloud.ai/api/v1/model/generateVideo"
headers = {
"Content-Type": "application/json",
"Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
"model": "kwaivgi/kling-video-o3-std/video-edit",
"prompt": "A beautiful sunset over the ocean with gentle waves",
"width": 512,
"height": 512,
"duration": 3,
"fps": 24
}
generate_response = requests.post(generate_url, headers=headers, json=data)
generate_result = generate_response.json()
prediction_id = generate_result["data"]["id"]
# Step 2: Poll for result
poll_url = f"https://api.atlascloud.ai/api/v1/model/prediction/{prediction_id}"
def check_status():
while True:
response = requests.get(poll_url, headers={"Authorization": "Bearer $ATLASCLOUD_API_KEY"})
result = response.json()
if result["data"]["status"] in ["completed", "succeeded"]:
print("Generated video:", result["data"]["outputs"][0])
return result["data"]["outputs"][0]
elif result["data"]["status"] == "failed":
raise Exception(result["data"]["error"] or "Generation failed")
else:
# Still processing, wait 2 seconds
time.sleep(2)
video_url = check_status()You need to be logged in to access your model request history.
Log InKling Video O3 Standard Video Edit enables natural-language video edits: remove or replace objects, swap backgrounds, restyle scenes, change weather/lighting, and apply localized 3-10s transformations with strong temporal consistency. Built for stable production use with a ready-to-use REST API and predictable performance.
Prompt-driven editing Describe your edits in plain language — no timeline, no masks, no manual keyframing required.
Reference image support Attach up to 4 reference images to guide the target element, scene, or style in the output.
Audio preservation Keep the original soundtrack intact with the keep_original_sound option.
Scene-level understanding The model recognizes objects, backgrounds, and context within the video to apply accurate, context-aware edits.
Motion-consistent output Edits blend naturally across frames with strong temporal coherence — minimal flicker or ghosting.
| Parameter | Required | Description |
|---|---|---|
| prompt | Yes | Text description of the desired edit |
| video | Yes | Input video to edit (URL or upload) |
| images | No | Up to 4 reference images for element, scene, or style guidance |
| keep_original_sound | No | Whether to keep the original sound from the video (default: enabled) |