98 skills found · Page 1 of 4
swapagarwal / JARVIS On Messenger:speech_balloon: A community-driven python bot that aims to be as simple as possible to serve humans with their everyday tasks
DinoMan / Speech Driven AnimationNo description available
EvelynFan / FaceFormer[CVPR 2022] FaceFormer: Speech-Driven 3D Facial Animation with Transformers
yiranran / Audio Driven TalkingFace HeadPoseCode for "Audio-driven Talking Face Video Generation with Learning-based Personalized Head Pose" (Arxiv 2020) and "Predicting Personalized Head Movement From Short Video and Speech Signal" (TMM 2022)
Doubiiu / CodeTalker[CVPR 2023] CodeTalker: Speech-Driven 3D Facial Animation with Discrete Motion Prior
psyai-net / EmoTalk ReleaseThis is the official source for our ICCV 2023 paper "EmoTalk: Speech-Driven Emotional Disentanglement for 3D Face Animation"
DiffPoseTalk / DiffPoseTalkDiffPoseTalk: Speech-Driven Stylistic 3D Facial Animation and Head Pose Generation via Diffusion Models
Aratako / Irodori TTSA Flow Matching-based Text-to-Speech Model with Emoji-driven Style Control
Olney1 / ChatGPT OpenAI Smart SpeakerThis AI Smart Speaker uses speech recognition, TTS (text-to-speech), and STT (speech-to-text) to enable voice and vision-driven conversations, with additional web search capabilities via OpenAI and Langchain agents.
Advocate99 / DiffGesture[CVPR'2023] Taming Diffusion Models for Audio-Driven Co-Speech Gesture Generation
shivangi-aneja / FaceTalk[CVPR 2024] FaceTalk: Audio-Driven Motion Diffusion for Neural Parametric Head Models
DanBigioi / DiffusionVideoEditingOfficial project repo for paper "Speech Driven Video Editing via an Audio-Conditioned Diffusion Model"
YoungSeng / DiffuseStyleGestureDiffuseStyleGesture: Stylized Audio-Driven Co-Speech Gesture Generation with Diffusion Models (IJCAI 2023) | The DiffuseStyleGesture+ entry to the GENEA Challenge 2023 (ICMI 2023, Reproducibility Award)
JeremyCJM / DiffSHEG[CVPR'24] DiffSHEG: A Diffusion-Based Approach for Real-Time Speech-driven Holistic 3D Expression and Gesture Generation
eeskimez / EmotalkingfaceThe code for the paper "Speech Driven Talking Face Generation from a Single Image and an Emotion Condition"
theEricMa / DiffSpeakerDiffSpeaker: Speech-Driven 3D Facial Animation with Diffusion Transformer
kiranchhatre / Amuse[CVPR 2024] AMUSE: Emotional Speech-driven 3D Body Animation via Disentangled Latent Diffusion
ZiqiaoPeng / EmoTalkThis is the repository for EmoTalk: Speech-Driven Emotional Disentanglement for 3D Face Animation
yunik1004 / SAiDSAiD: Blendshape-based Audio-Driven Speech Animation with Diffusion
Svito-zar / GesticulatorThe official implementation for ICMI 2020 Best Paper Award "Gesticulator: A framework for semantically-aware speech-driven gesture generation"