How Omni-Effects Enables Spatially Controllable Multi‑VFX Generation with LoRA‑MoE

Omni-Effects introduces a unified framework that combines LoRA‑based expert mixture models and spatially aware prompts to generate multiple, precisely placed visual effects in video, supported by the new Omni‑VFX dataset and evaluation suite, demonstrating superior spatial control and diversity over prior single‑effect methods.

Amap Tech
Amap Tech
Amap Tech
How Omni-Effects Enables Spatially Controllable Multi‑VFX Generation with LoRA‑MoE

Overview

Visual effects (VFX) are essential in modern film production, but existing video‑generation models are limited to single‑effect generation because each effect requires a dedicated LoRA adapter, preventing spatially controllable composite effects. This limitation hampers applications that need multiple effects simultaneously at specific locations.

To address these challenges, the authors propose Omni-Effects , the first unified framework that can generate prompt‑guided, spatially controllable composite VFX. The framework’s core innovations are:

LoRA‑MoE (LoRA expert mixture model) : integrates a set of expert LoRAs within a single model to handle multiple effects while reducing cross‑task interference.

Spatial‑Aware Prompt (SAP) with an Independent Information Flow (IIF) module: embeds spatial mask information into text tokens for precise pixel‑level control and isolates control signals for each effect to avoid unwanted mixing.

To support this research, a novel data‑collection pipeline combining image editing with first‑last‑frame‑to‑video (FLF2V) synthesis was used to build the comprehensive VFX dataset Omni‑VFX , accompanied by a dedicated VFX evaluation framework.

Approach

The strength of Omni‑Effects stems from the two core technologies:

LoRA‑MoE : By coupling LoRA with a mixture‑of‑experts architecture, each “effect expert” specializes in a particular visual effect, mitigating the interference that occurs when multiple LoRAs are jointly activated.

SAP‑IIF : The Spatial‑Aware Prompt injects spatial mask data into the Transformer’s attention mechanism, achieving pixel‑accurate effect placement without the heavy parameter overhead of methods like ControlNet. The IIF acts as a firewall, ensuring that control signals for different effects remain independent.

Experiments

Quantitative comparisons show that LoRA‑MoE outperforms baseline LoRA setups, and the controllable VFX results demonstrate precise spatial control and higher visual fidelity. Visual examples illustrate the diversity and quality of generated effects.

Conclusion

Omni‑Effects provides a unified framework for generating customized VFX videos, supporting single, multiple, and spatially controllable multi‑VFX generation. By integrating LoRA‑MoE and SAP‑IIF, the system mitigates cross‑condition interference and ensures precise spatial control. The accompanying Omni‑VFX dataset and evaluation suite validate the method’s robustness, opening new possibilities for film, game, and advertising industries.

Paper: Omni-Effects: Unified and Spatially Controllable Visual Effects Generation

Project: https://github.com/AMAP-ML/Omni-Effects

AIvideo generationLoRAcomputer graphicsvisual effectsspatial control
Amap Tech
Written by

Amap Tech

Official Amap technology account showcasing all of Amap's technical innovations.

0 followers
Reader feedback

How this landed with the community

Sign in to like

Rate this article

Was this worth your time?

Sign in to rate
Discussion

0 Comments

Thoughtful readers leave field notes, pushback, and hard-won operational detail here.