wan-2-2-animate-v2v
Creates animated videos by transferring movement from one video to a static character image, or replaces a character in an existing video with a new image while preserving the original motion and environment.
api reference
about
creates animated videos by transferring movement from one video to a static character image, or replaces a character in an existing video with a new image while preserving the original motion and environment.
1. calling the api
install the client
the client provides a convenient way to interact with the api.
1pip install inferenceshsetup your api key
set INFERENCE_API_KEY as an environment variable. get your key from settings → api keys.
1export INFERENCE_API_KEY="inf_your_key"run and get result
submit a request and wait for the final result. best for batch processing or when you don't need progress updates.
1from inferencesh import inference23client = inference()456result = client.run({7 "app": "infsh/wan-2-2-animate-v2v",8 "input": {}9 })1011print(result["output"])stream live updates
get real-time progress updates as the task runs. ideal for showing progress bars, partial results, or long-running tasks.
1from inferencesh import inference23client = inference()456# stream=True yields updates as they arrive7for update in client.run({8 "app": "infsh/wan-2-2-animate-v2v",9 "input": {}10 }, stream=True):11 if update.get("progress"):12 print(f"progress: {update['progress']}%")13 if update.get("output"):14 print(f"output: {update['output']}")2. authentication
the api uses api keys for authentication. see the authentication docs for detailed setup instructions.
3. files
file inputs are automatically handled by the sdk. you can pass local paths, urls, or base64 data.
automatic upload
the python sdk automatically detects local file paths and uploads them. urls are passed through as-is.
1# local file paths are automatically uploaded2result = client.run({3 "app": "infsh/wan-2-2-animate-v2v",4 "input": {5 "image": "/path/to/local/image.png", # detected & uploaded6 "audio": "https://example.com/audio.mp3", # url passed through7 }8})4. webhooks
get notified when a task completes by providing a webhook url. when the task reaches a terminal state (completed, failed, or cancelled), a POST request is sent to your url with the task result.
1result = client.run({2 "app": "infsh/wan-2-2-animate-v2v",3 "input": {},4 "webhook": "https://your-server.com/webhook"5}, wait=False)webhook payload
your endpoint receives a JSON POST with the task result:
1{2 "id": "task_abc123",3 "status": 9,4 "output": { ... },5 "error": "",6 "session_id": null,7 "created_at": "2024-01-15T10:30:00Z",8 "updated_at": "2024-01-15T10:30:05Z"9}5. schema
input
frames per clip (must be 4n+1, e.g., 77, 81, 121)
driving_video
target fps (-1 to use original video fps)
classifier-free guidance scale for expression control
text prompt (leave empty for default: '视频中的人在做动作')
mask height subdivisions for detail (replacement mode)
mask dilation iterations (replacement mode)
mask dilation kernel size (replacement mode)
mask width subdivisions for detail (replacement mode)
mode: 'animation' (character animation) or 'replacement' (replace character in video)
negative prompt
offload models to cpu to save vram
reference_image
temporal guidance frames (1 or 5 recommended)
target height for processing
target width for processing
enable pose retargeting (recommended for different body proportions)
sampling solver (dpm++ or unipc)
number of diffusion steps
random seed (-1 for random)
noise schedule shift parameter
use flux image editing for better pose retargeting (slower but better quality)
ready to run wan-2-2-animate-v2v?
we use cookies
we use cookies to ensure you get the best experience on our website. for more information on how we use cookies, please see our cookie policy.
by clicking "accept", you agree to our use of cookies.
learn more.