SAMA is a new 14B open model for instruction-guided video editing that separates semantic anchoring from motion alignment and claims state-of-the-art open results. Track it if you need edits that change objects or style without wrecking motion.

SAMA is pitched as a general video editor for object replacement, addition, removal, and style transfer. The core idea is sparse anchor frames: the model predicts semantic tokens and video latents at key frames, then uses a separate motion-focused module to carry those edits through time without the usual flicker or drift.
The paper describes a two-stage training setup. First, SAMA learns motion from raw video with restoration-style pretext tasks including cube inpainting, speed perturbation, and tube shuffling. Then it is fine-tuned on paired editing data for instruction following. That separation is the creative hook: it is designed for prompts that swap subjects or restyle scenes while keeping camera movement and scene dynamics intact, which is also what the supporting writeup highlights in its demo overview.
Topview added Seedance 2.0 to Agent V2, pairing multi-scene generation with a storyboard timeline and Business Annual access billed as 365 days of unlimited generations. That moves longform video workflows toward editable sequences instead of stitched clips.
releaseTopview added Seedance 2.0 to Agent V2, pairing multi-scene generation with a storyboard timeline and Business Annual access billed as 365 days of unlimited generations. That moves longform video workflows toward editable sequences instead of stitched clips.
workflowCreators are moving from V8 calibration complaints to darker film-still scenes, fashion shots, and worldbuilding tests, with ECLIPTIC remakes showing stronger depth and lighting. Retest saved SREF recipes if you rely on V8 for cinematic ideation.
workflowA shared workflow converts GTA-style stills into photoreal images with Nano Banana 2, then animates them in LTX-2.3 Pro 4K using detailed material, skin, vehicle, and camera prompts. Try it for trailer-style previsualization if you want more control at lower cost.
workflowShared Nano Banana 2 workflows now cover turnaround sheets, distinctive facial traits, and photoreal rerenders that keep the framing of a reference image. Use one prompt grammar for concept art, editorial portraits, and animation prep.
Baidu, Tsinghua, and Zhejiang University release SAMA. Factorized semantic anchoring and motion alignment for instruction-guided video editing. Balances semantic modification and motion preservation through sparse anchor frames and motion-centric pretext tasks. Two-stageย Show more
SAMA Factorized Semantic Anchoring and Motion Alignment for Instruction-Guided Video Editing paper: huggingface.co/papers/2603.19โฆ