Multi-Input Turns Generators into Precise Video Editors
Sirio Berati, founder of Enhancor, positions Seedance V2 as the first widely accessible model supporting true multi-input generation: up to two images, two videos, and one audio file in a single prompt. This shifts it from mere video creation to sophisticated editing. In the first demo, Sirio takes a green-screen AI-generated video with two characters and swaps both for new references while replacing the background—all in one 60-second generation. Motion from the original is preserved exactly, controlled via natural language like "keep the motion of the original video exactly the same."
Greg Isenberg notes the motion control's impressiveness, and Sirio emphasizes: "Cense 2 it's not only a video generator it is a video editor that's how I see it. It's almost like nano banana pro whereby the use cases are unlimited." This capability rivals Kling 3 but surpasses it in quality, enabling production studios to iterate landing page demos or social clips without costly reshoots.
Specificity in Prompts and References Drives Quality
Seedance V2 demands detailed prompts unlike shorter ones suiting Kling 3. Sirio starts drafts manually, then optimizes with Claude Opus 4.6, which excels at vision model prompting. For high-fidelity outputs—preserving character identity, motions, or transitions—specificity is key: describe exact actions, textures, and references.
Source references are the biggest quality lever. Sirio likens models to human assistants: "Everything starts with a very good idea a very good source reference source image. What is your vision? ... they're able to understand your taste and they're able to mimic uh um that that reference image." In demos, strong references ensure tasteful outputs, like matching pant patterns or boot cuts. Greg praises Sirio's style in references, highlighting how they elevate results beyond model capabilities.
"You have to be highly specific if you want to get very high quality output," Sirio advises, especially for identity preservation.
E-Commerce Try-Ons and Scalable Ad Localization
For e-commerce, Sirio shot himself in -30°C Montreal wearing shorts, then prompted Seedance V2 to swap into a winter outfit with a bear walking by. Face identity holds perfectly—no distortions Greg could spot—while outfit details (boot patterns, pant cuts) match references exactly. The model even tracks the bear with eyes and head turns, adding footprints dynamically.
Sirio sees this for ecom shoots: reuse actor motions across outfits for consistent assets. Commercial angle: generate brand-specific visuals rapidly.
Ad translation demo swaps a Chinese glasses ad model for an English-speaking AI-generated one. Same wink, hand-on-glasses motion, camera blur, and focus. Audio translates Mandarin to English: "This one's amazing. It's flattering and versatile. Must have." Greg calls it A/B testing gold: "creating ads and just creating content spec in in like a hundred languages... Cheaper ads, higher conversion, continuous optimization."
Another: Populate 3D product templates. Sirio textures a generic package render with a branded image (yellow background, consistent logo), keeping all else identical. Source from stock like Freepik, extend to video via prompts referencing inputs.
Video Extension and Lifelike AI Influencers
Seedance fills longstanding gaps in extension. Sirio extends a 3-second clip seamlessly, recreating the storyline from the last frame per prompt, maintaining consistency. A variant fills middles between two clips, ideal for ads or films needing extra seconds without reshooting—a personal pain point for Greg.
For AI influencers, it's unmatched for lip-sync realism. Using a Midjourney-like source image (nano banana pro), Sirio prompts hyper-specific actions: muscle movements, emotional transitions over labels like "happy." Influencers perform any scripted dialogue fluidly. Sirio: "This is the best model for you to generate AI influencers and they can do anything you want them to do."
Enhancor integrates this across models, but Seedance V2 is default for editing/generation.
Trade-Offs: Seedance Leads Editing, Others Niche Wins
Sirio crowns Seedance V2 best overall for realism, motion, quality—at 720p now, 1080p soon a game-changer. Kling 3 for cinematic feel, Enhancer V4 for talking-heads. Greg probes Adobe's future: Sirio predicts disruption as prompt-based editing scales creative assets.
Business playbook emerges: Build apps productizing workflows (Enhancor-style), create converting ads/influencers/movies, faceless accounts. Avoid hype—focus prompts, references for production use.
"Is is Cance 2 the best video model to ever exist ... for now? Yes. ... by far, uh, it is the best out there," Sirio affirms.
Key Takeaways
- Use multi-input (2 images/videos + audio) for complex edits like character/background swaps in one prompt, preserving original motion.
- Optimize prompts with Claude Opus 4.6 after manual drafts; prioritize hyper-specific details on identity, motions, transitions.
- Leverage strong source references to instill taste—mimic human inspiration for tangible, high-quality outputs.
- For ecom: Virtual try-ons preserve face/motion while swapping outfits; add dynamic elements like animals seamlessly.
- Scale ads via translation/character swaps: A/B test languages/demographics holding visuals constant for optimization.
- Extend videos by filling ends/middles or populate 3D templates with brand textures for evergreen assets.
- Craft AI influencers with muscle/emotion descriptions (not labels) for realistic lip-sync performances.
- Default to Seedance V2 for editing/generation; pair with Kling 3 (cinematic), Enhancer V4 (talking heads).
- Productize workflows in platforms like Enhancor to monetize: ads, influencers, ecom assets at scale.