-
Tencent, YouTu lab
- Shanghai
- linchuming.github.io
Stars
Evaluation code for "Efficient Emotional Adaptation for Audio-Driven Talking-Head Generation"
Dynamic and static models for real-time facial emotion recognition
A Lightweight Face Recognition and Facial Attribute Analysis (Age, Gender, Emotion and Race) Library for Python
ICIP 2019: Frame Attention Networks for Facial Expression Recognition in Videos
FaceVid-1K: A Large-Scale High-Quality Multiracial Human Face Video Dataset
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
Faster Whisper transcription with CTranslate2
Multilingual Automatic Speech Recognition with word-level timestamps and confidence
Whisper realtime streaming for long speech-to-text transcription and translation
Official Pytorch Implementation of FLOAT: Generative Motion Latent Flow Matching for Audio-driven Talking Portrait.
Official implementation of "Sonic: Shifting Focus to Global Audio Perception in Portrait Animation"
Memory-optimized training library for diffusion models
Official inference repo for FLUX.1 models
HunyuanVideo: A Systematic Framework For Large Video Generation Model
[CVPR 2025] EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation
Human Motion Video Generation: A Survey (https://www.techrxiv.org/users/836049/articles/1228135-human-motion-video-generation-a-survey)
Hallo2: Long-Duration and High-Resolution Audio-driven Portrait Image Animation
The minimal opencv for Android, iOS, ARM Linux, Windows, Linux, MacOS, HarmonyOS, WebAssembly, watchOS, tvOS, visionOS
Code for paper 'EAMM: One-Shot Emotional Talking Face via Audio-Based Emotion-Aware Motion Model'
Project Page of 'GANFIT: Generative Adversarial Network Fitting for High Fidelity 3D Face Reconstruction' [CVPR2019]
Official implementation of EMOPortraits: Emotion-enhanced Multimodal One-shot Head Avatars
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)