Lists (12)
Sort Name ascending (A-Z)
Stars
[ICLR 2025] Official lmplementation of SPM-Diff: Incorporating Visual Correspondence into Diffusion Model for Virtual Try-On
HunyuanCustom: A Multimodal-Driven Architecture for Customized Video Generation
[ARXIV'25] ReCamMaster: Camera-Controlled Generative Rendering from A Single Video
Official implementation of Inductive Moment Matching
Pytorch implementation for the paper titled "SimpleAR: Pushing the Frontier of Autoregressive Visual Generation"
This is a repo to track the latest autoregressive visual generation papers.
[CVPR 2025] Official Implementation of MotionPro: A Precise Motion Controller for Image-to-Video Generation
MAGI-1: Autoregressive Video Generation at Scale
SkyReels-V2: Infinite-length Film Generative model
πππA curated list of papers on controllable video generation.
SkyReels-A2: Compose anything in video diffusion transformers
STeP: a general and scalable framework for solving video inverse problems with spatiotemporal diffusion priors
Stable Virtual Camera: Generative View Synthesis with Diffusion Models
SpatialLM: Large Language Model for Spatial Understanding
D-FINE: Redefine Regression Task of DETRs as Fine-grained Distribution Refinement [ICLR 2025 Spotlight]
Official implementations for paper: VACE: All-in-One Video Creation and Editing
[SIGGRAPH2025] Official repo for paper "Any-length Video Inpainting and Editing with Plug-and-Play Context Control"
Wan: Open and Advanced Large-Scale Video Generative Models
AC3D: Analyzing and Improving 3D Camera Control in Video Diffusion Transformers
[CVPR2025 Highlight] Video Generation Foundation Models: https://saiyan-world.github.io/goku/
Official code of "EVF-SAM: Early Vision-Language Fusion for Text-Prompted Segment Anything Model"
Official repository of "SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory"
π Cross attention map tools for huggingface/diffusers