MotionBricks: Scalable Real-Time Motions with Modular Latent Generative Model and Smart Primitives

Transactions on Graphics (Proc. ACM SIGGRAPH 2026)

Tingwu Wang* (1)    Olivier Dionne* (1)    Michael De Ruyter (1)    David Minor (1)    Davis Rempe (1)    Kaifeng Zhao (1, 2)    Mathis Petrovich (1)    Ye Yuan (1)    Chenran Li (1)    Zhengyi Luo (1)    Brian Robison (1)    Xavier Blackwell (1)    Bernardo Antoniazzi (1)    Xue Bin Peng (1, 3)    Yuke Zhu (1, 4)    Simon Yuen (1)

(1) NVIDIA    (2) ETH Zürich    (3) Simon Fraser University    (4) The University of Texas at Austin

*Joint first authors.



Abstract

Despite transformative advances in generative motion synthesis, real-time interactive motion control remains dominated by traditional techniques. In this work, we identify two key challenges in bridging research and production: 1) Real-time scalability: Industry applications demand real-time generation of a vast repertoire of motion skills, while generative methods exhibit significant degradation in quality and scalability under real-time computation constraints, and 2) Integration: Industry applications demand fine-grained multi-modal control involving velocity commands, style selection, and precise keyframes, a need largely unmet by existing text- or tag-driven models. To overcome these limitations, we introduce MotionBricks: a large-scale, real-time generative framework with a two-fold solution. First, we propose a large-scale modular latent generative backbone tailored for robust real-time motion generation, effectively modeling a dataset of over 350,000 motion clips with a single model. Second, we introduce smart primitives that provide a unified, robust, and intuitive interface for authoring both navigation and object interaction. Applications can be designed in a plug-and-play manner like assembling bricks without expert animation knowledge. Quantitatively, we show that MotionBricks produces state-of-the-art motion quality on open-source and proprietary datasets of various scales, while also achieving a real-time throughput of 15,000 FPS with 2ms latency. We demonstrate the flexibility and robustness of MotionBricks in a complete production-level animation demo, covering navigation and object-scene interaction across various styles with a unified model. To showcase our framework's application beyond animation, we deploy MotionBricks on the Unitree G1 humanoid robot to demonstrate its flexibility and generalization for real-time robotic control.

Paper: [PDF]       Webpage: [Link]       Preprint: [arXiv]

Video



Bibtex

@misc{
	wang2026motionbricksscalablerealtimemotions,
	title={MotionBricks: Scalable Real-Time Motions with Modular Latent Generative Model and Smart Primitives},
	author={Tingwu Wang and Olivier Dionne and Michael De Ruyter and David Minor and Davis Rempe and Kaifeng Zhao and Mathis Petrovich and Ye Yuan and Chenran Li and Zhengyi Luo and Brian Robison and Xavier Blackwell and Bernardo Antoniazzi and Xue Bin Peng and Yuke Zhu and Simon Yuen},
	year={2026},
	eprint={2604.24833},
	archivePrefix={arXiv},
	primaryClass={cs.RO},
	url={https://arxiv.org/abs/2604.24833},
}