跳至主要内容
小龙虾小龙虾AI
🤖

Seedance 2.0 API

Out-of-the-box Seedance 2.0 API skill — just one API key to generate AI videos. Builds storyboards, generates reference images with Seedream 4.5, submits vid...

下载463
星标0
版本1.1.0
设计媒体
安全通过
⚙️脚本

技能说明


name: seedance2-api description: Out-of-the-box Seedance 2.0 API skill — just one API key to generate AI videos. Builds storyboards, generates reference images with Seedream 4.5, submits video tasks, and polls results. Supports both MCP and standalone Python script mode. Use when the user mentions seedance, AI video, storyboard, or video generation. license: MIT compatibility: Requires Python 3.8+ with requests. Works with Cursor, Claude Code, or any SKILL.md-compatible agent. metadata: author: hexiaochun version: "1.1" tags: video-generation ai-video seedance storyboard bytedance seedream

Seedance 2.0 Storyboard & Video Generation

End-to-end workflow from concept to final video: Storyboard → Reference images → Submit video task → Get results.

Step 0: Determine Execution Mode (MCP or Script)

Check MCP availability first:

  1. Check xskill-ai MCP service status (read mcps/user-xskill-ai/STATUS.md)
  2. If MCP is available → use submit_task / get_task and other MCP tools
  3. If MCP is unavailable or returns errors → switch to Script Mode

Script mode prerequisites:

  1. Verify XSKILL_API_KEY environment variable is set (run echo $XSKILL_API_KEY | head -c 10)
  2. If not set, prompt the user:
    export XSKILL_API_KEY=sk-your-api-key
    Get your API Key: https://www.xskill.ai/#/v2/api-keys
    
  3. Verify requests is installed (pip install requests)

Script path: Located under this skill's directory at scripts/seedance_api.py:

# Find via Glob tool
glob: .cursor/skills/seedance2-api/scripts/seedance_api.py

In the following steps, each API call provides both MCP method and Script method. Choose one based on the Step 0 result.

Step 1: Understand the User's Idea

Collect the following information (proactively ask if anything is missing):

  • Story concept: one-sentence summary of the video
  • Duration: 4–15 seconds
  • Aspect ratio: 16:9 / 9:16 / 1:1 / 21:9 / 4:3 / 3:4
  • Visual style: realistic / animation / ink wash / sci-fi / cyberpunk, etc.
  • Assets: existing images/videos/audio, or need AI generation
  • Function mode: first & last frame control (first_last_frames) or default omni mode (omni_reference)

Step 2: Deep Dive (5 Dimensions)

Guide the user through each dimension for richer detail:

  1. Content – Who is the subject? What are they doing? Where?
  2. Visuals – Lighting, color palette, texture, mood
  3. Camera – Push in / pull out / pan / tilt / track / orbit / crane
  4. Motion – Subject actions and pacing
  5. Audio – Music style, sound effects, dialogue

Step 3: Build Storyboard Structure

Break down shots along the timeline using this formula:

[Style] _____ style, _____ seconds, _____ ratio, _____ mood

0-Xs: [Camera movement] + [Visual content] + [Action description]
X-Ys: [Camera movement] + [Visual content] + [Action description]
...

[Audio] _____ music + _____ SFX + _____ dialogue
[References] @image_file_1 _____, @video_file_1 _____

See reference.md for detailed templates and examples.

Step 4: Generate Reference Images (If Needed)

If the user has no existing assets, use Seedream 4.5 to generate character art, scenes, first/last frames, etc.

Text-to-Image

<details> <summary><b>MCP Method</b></summary>

Call submit_task tool:

  • model_id: fal-ai/bytedance/seedream/v4.5/text-to-image
  • parameters:
    • prompt: detailed image description (English works best)
    • image_size: choose based on video aspect ratio
    • num_images: number needed (1–6)
</details> <details> <summary><b>Script Method</b></summary>
python .cursor/skills/seedance2-api/scripts/seedance_api.py submit \
  --model "fal-ai/bytedance/seedream/v4.5/text-to-image" \
  --params '{"prompt":"An astronaut in a white spacesuit...","image_size":"landscape_16_9","num_images":1}'
</details>

Image Editing (Modify Existing Images)

<details> <summary><b>MCP Method</b></summary>

Call submit_task tool:

  • model_id: fal-ai/bytedance/seedream/v4.5/edit
  • parameters:
    • prompt: editing instructions (use Figure 1/2/3 to reference images)
    • image_urls: array of input image URLs
    • image_size: output size
</details> <details> <summary><b>Script Method</b></summary>
python .cursor/skills/seedance2-api/scripts/seedance_api.py submit \
  --model "fal-ai/bytedance/seedream/v4.5/edit" \
  --params '{"prompt":"Change the background to a forest","image_urls":["https://..."],"image_size":"landscape_16_9"}'
</details>

Poll Image Results

Images typically complete in 1–2 minutes.

<details> <summary><b>MCP Method</b></summary>

Call get_task tool to check status:

  • First query after 30 seconds
  • Then every 30 seconds
  • Extract image URL when status is completed
</details> <details> <summary><b>Script Method</b></summary>

Single query:

python .cursor/skills/seedance2-api/scripts/seedance_api.py query \
  --task-id "TASK_ID_HERE"

Auto-poll (recommended for images, interval 10s, timeout 180s):

python .cursor/skills/seedance2-api/scripts/seedance_api.py poll \
  --task-id "TASK_ID_HERE" --interval 10 --timeout 180
</details>

image_size Reference

Aspect RatioRecommended image_sizeNote
16:9landscape_16_9Landscape
9:16portrait_16_9Portrait
4:3landscape_4_3Landscape
3:4portrait_4_3Portrait
1:1square_hdSquare
21:9landscape_16_9Approximate ultrawide

Step 5: Compose the Final Prompt

Merge the storyboard structure and reference images into the final prompt:

  • Use @image_file_1, @image_file_2, etc. to reference images in the image_files array
  • Use @video_file_1, etc. to reference videos in the video_files array
  • Use @audio_file_1, etc. to reference audio in the audio_files array

Reference syntax example:

@image_file_1 as character reference, follow @video_file_1 camera movement, with @audio_file_1 as background music

Important: The Nth URL in image_files maps to @image_file_N. video_files and audio_files are independently numbered.

Step 6: Submit Video Task

Handle asset URLs:

  • Seedream-generated images: URL already available, use directly
  • User-provided web images: use directly
  • User-provided local images: upload first to get URL (see upload methods below)

Upload Local Images

<details> <summary><b>MCP Method</b></summary>

Call upload_image tool: image_url or image_data

</details> <details> <summary><b>Script Method</b></summary>
# Upload from URL
python .cursor/skills/seedance2-api/scripts/seedance_api.py upload \
  --image-url "https://example.com/image.png"

# Upload local file
python .cursor/skills/seedance2-api/scripts/seedance_api.py upload \
  --image-path "/path/to/local/image.png"
</details>

Submit Seedance 2.0 Task (Omni Reference Mode)

<details> <summary><b>MCP Method</b></summary>

Call submit_task tool:

  • model_id: st-ai/super-seed2
  • parameters:
    • prompt: the full prompt from Step 5
    • functionMode: omni_reference (default, can be omitted)
    • image_files: reference image URL array (up to 9, order matches @image_file_1/2/3...)
    • video_files: reference video URL array (up to 3, total duration ≤ 15s)
    • audio_files: reference audio URL array (up to 3)
    • ratio: aspect ratio (16:9 / 9:16 / 1:1 / 21:9 / 4:3 / 3:4)
    • duration: integer length (415)
    • model: seedance_2.0_fast (default, faster) or seedance_2.0 (standard quality)
</details> <details> <summary><b>Script Method</b></summary>
python .cursor/skills/seedance2-api/scripts/seedance_api.py submit \
  --model "st-ai/super-seed2" \
  --params '{
    "prompt": "Cinematic realistic sci-fi style, 15 seconds, 16:9...",
    "functionMode": "omni_reference",
    "image_files": ["https://img1.png", "https://img2.png"],
    "ratio": "16:9",
    "duration": 15,
    "model": "seedance_2.0_fast"
  }'
</details>

Submit Seedance 2.0 Task (First & Last Frames Mode)

<details> <summary><b>MCP Method</b></summary>

Call submit_task tool:

  • model_id: st-ai/super-seed2
  • parameters:
    • prompt: video description prompt
    • functionMode: first_last_frames
    • filePaths: image URL array (0 = text-to-video, 1 = first frame, 2 = first & last frames)
    • ratio: aspect ratio
    • duration: integer length
    • model: seedance_2.0_fast or seedance_2.0
</details> <details> <summary><b>Script Method</b></summary>
python .cursor/skills/seedance2-api/scripts/seedance_api.py submit \
  --model "st-ai/super-seed2" \
  --params '{
    "prompt": "Camera smoothly transitions from first frame to last frame, fluid motion",
    "functionMode": "first_last_frames",
    "filePaths": ["https://first-frame.png", "https://last-frame.png"],
    "ratio": "16:9",
    "duration": 5,
    "model": "seedance_2.0_fast"
  }'
</details>

Step 7: Poll for Video Results

Video generation takes approximately 10 minutes.

<details> <summary><b>MCP Method</b></summary>

Polling strategy:

  1. After submission, inform the user: "Video is generating, estimated ~10 minutes"
  2. First query after 60 seconds via get_task
  3. Then every 90 seconds
  4. Report status to the user after each query
</details> <details> <summary><b>Script Method</b></summary>

Recommended: auto-poll (runs in foreground, interval 30s, timeout 600s):

python .cursor/skills/seedance2-api/scripts/seedance_api.py poll \
  --task-id "TASK_ID_HERE" --interval 30 --timeout 600

Progress is printed to stderr; final JSON result is printed to stdout when complete.

Manual single query:

python .cursor/skills/seedance2-api/scripts/seedance_api.py query \
  --task-id "TASK_ID_HERE"
</details>

Status reference:

  • pending → "Queued..."
  • processing → "Generating..."
  • completed → Extract the video URL and present to the user
  • failed → Report the error; suggest adjusting the prompt and retrying

Full Workflow Example

User says: "Make a video of an astronaut walking on Mars"

When MCP Is Available

1. Gather info → 15s, 16:9, cinematic sci-fi style, no existing assets

2. Generate astronaut + Mars scene images with Seedream 4.5
   submit_task("fal-ai/bytedance/seedream/v4.5/text-to-image", {...})
   → poll get_task → get image URLs

3. Compose prompt → submit video task
   submit_task("st-ai/super-seed2", {...})

4. Poll get_task, ~10 min later → get video URL

When MCP Is Unavailable (Script Mode)

1. Gather info → 15s, 16:9, cinematic sci-fi style

2. Generate reference images:
   python scripts/seedance_api.py submit \
     --model "fal-ai/bytedance/seedream/v4.5/text-to-image" \
     --params '{"prompt":"An astronaut in white spacesuit on Mars...","image_size":"landscape_16_9"}'
   → get task_id

3. Poll for image results:
   python scripts/seedance_api.py poll --task-id "xxx" --interval 10 --timeout 180
   → get image URL

4. Submit video task:
   python scripts/seedance_api.py submit \
     --model "st-ai/super-seed2" \
     --params '{"prompt":"...storyboard prompt...","functionMode":"omni_reference","image_files":["IMAGE_URL"],"ratio":"16:9","duration":15,"model":"seedance_2.0_fast"}'
   → get task_id

5. Poll for video results:
   python scripts/seedance_api.py poll --task-id "xxx" --interval 30 --timeout 600
   → get video URL

Model Parameters Quick Reference

Seedream 4.5 Text-to-Image

ParameterTypeRequiredDescription
promptstringYesImage description
image_sizestringNoauto_2K / auto_4K / square_hd / portrait_4_3 / portrait_16_9 / landscape_4_3 / landscape_16_9
num_imagesintNo1–6, default 1

Seedream 4.5 Image Editing

ParameterTypeRequiredDescription
promptstringYesEditing instructions, reference images as Figure 1/2/3
image_urlsarrayYesInput image URL list
image_sizestringNoSame as above
num_imagesintNo1–6, default 1

Seedance 2.0 Video (Omni Reference Mode)

ParameterTypeRequiredDescription
promptstringYesStoryboard prompt, use @image_file_N/@video_file_N/@audio_file_N
functionModestringNoomni_reference (default)
image_filesarrayNoReference image URL array (up to 9)
video_filesarrayNoReference video URL array (up to 3, total ≤ 15s)
audio_filesarrayNoReference audio URL array (up to 3)
ratiostringNo21:9 / 16:9 / 4:3 / 1:1 / 3:4 / 9:16
durationintegerNo4–15, default 5
modelstringNoseedance_2.0_fast (default) / seedance_2.0

Seedance 2.0 Video (First & Last Frames Mode)

ParameterTypeRequiredDescription
promptstringYesVideo description prompt
functionModestringYesfirst_last_frames
filePathsarrayNoImage URL array (0 = text-to-video, 1 = first frame, 2 = first & last)
ratiostringNo21:9 / 16:9 / 4:3 / 1:1 / 3:4 / 9:16
durationintegerNo4–15, default 5
modelstringNoseedance_2.0_fast (default) / seedance_2.0

Tools Quick Reference

MCP Tools

ActionToolKey Parameters
Submit tasksubmit_taskmodel_id, parameters
Query resultget_tasktask_id
Upload imageupload_imageimage_url or image_data
Check balanceget_balance(none)

Script Commands (When MCP Is Unavailable)

ActionCommandDescription
Submit taskpython scripts/seedance_api.py submit --model MODEL --params '{...}'Returns task_id
Single querypython scripts/seedance_api.py query --task-id IDReturns current status
Auto-pollpython scripts/seedance_api.py poll --task-id ID --interval N --timeout NBlocks until done
Check balancepython scripts/seedance_api.py balanceReturns account balance
Upload imagepython scripts/seedance_api.py upload --image-url URL or --image-path PATHReturns image URL

Script path note: The scripts/seedance_api.py path above is relative to .cursor/skills/seedance2-api/. Use the full path .cursor/skills/seedance2-api/scripts/seedance_api.py when executing, or cd into the skill directory first.

Seedance 2.0 Limitations

  • Realistic human face uploads are not supported
  • Maximum 12 files: images ≤ 9 + videos ≤ 3 + audio ≤ 3
  • Total video/audio reference duration ≤ 15 seconds
  • Video references consume more credits

More Resources

See reference.md for detailed storyboard templates, full examples, and camera movement glossary.

如何使用「Seedance 2.0 API」?

  1. 打开小龙虾AI(Web 或 iOS App)
  2. 点击上方「立即使用」按钮,或在对话框中输入任务描述
  3. 小龙虾AI 会自动匹配并调用「Seedance 2.0 API」技能完成任务
  4. 结果即时呈现,支持继续对话优化

相关技能