![bot logo](https://raw.githubusercontent.com/github/explore/0a84ca418425da147e4e43b1c74aa169d3265870/topics/bot/bot.png)
-
Tsinghua University
- Nanshan, Shenzhen, China
-
11:17
(UTC +08:00) - https://yisuanwang.github.io/
Lists (1)
Sort Name descending (Z-A)
Starred repositories
MagicArticulate: Make Your 3D Models Articulation-Ready
Automate the process of making money online.
Light-A-Video: Training-free Video Relighting via Progressive Light Fusion
[NeurIPS 2024] Direct3D: Scalable Image-to-3D Generation via 3D Latent Diffusion Transformer
Investigating CoT Reasoning in Autoregressive Image Generation
VILA is a family of state-of-the-art vision language models (VLMs) for diverse multimodal AI tasks across the edge, data center, and cloud.
MatAnyone: Stable Video Matting with Consistent Memory Propagation
[ICLR 2025] Animate-X: Universal Character Image Animation with Enhanced Motion Representation
Character Animation (AnimateAnyone, Face Reenactment)
A fork to add multimodal model training to open-r1
Janus-Series: Unified Multimodal Understanding and Generation Models
Fully open reproduction of DeepSeek-R1
Video Depth Anything: Consistent Depth Estimation for Super-Long Videos
High-Resolution 3D Assets Generation with Large Scale Hunyuan3D Diffusion Models.
AniGS: Animatable Gaussian Avatar from a Single Image with Inconsistent Gaussian Reconstruction
An aggregation of human motion understanding research.
[arXiv 2025] Official pytorch implementation of "FramePainter: Endowing Interactive Image Editing with Video Diffusion Priors"
Official repository of "SAMURAI: Adapting Segment Anything Model for Zero-Shot Visual Tracking with Motion-Aware Memory"
🔥 Sa2VA: Marrying SAM2 with LLaVA for Dense Grounded Understanding of Images and Videos
SPAR3D: Stable Point-Aware Reconstruction of 3D Objects from Single Images
Blending Custom Photos with Video Diffusion Transformers
Agent Laboratory is an end-to-end autonomous research workflow meant to assist you as the human researcher toward implementing your research ideas
Bringing Characters to Life with Computer Brains in Unity
Official repository for "MMM: Generative Masked Motion Model" (CVPR 2024 -- Highlight)