AI & ML interests

None defined yet.

Recent Activity

DmitryRyuminย 
posted an update 2 days ago
view post
Post
2615
๐Ÿš€๐Ÿ‘Œ๐ŸŒŸ New Research Alert - ICCV 2025 (Oral)! ๐ŸŒŸ๐ŸคŒ๐Ÿš€
๐Ÿ“„ Title: Understanding Co-speech Gestures in-the-wild ๐Ÿ”

๐Ÿ“ Description: JEGAL is a tri-modal model that learns from gestures, speech and text simultaneously, enabling devices to interpret co-speech gestures in the wild.

๐Ÿ‘ฅ Authors: @sindhuhegde , K R Prajwal, Taein Kwon, and Andrew Zisserman

๐Ÿ“… Conference: ICCV, 19 โ€“ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: Understanding Co-speech Gestures in-the-wild (2503.22668)

๐ŸŒ Web Page: https://www.robots.ox.ac.uk/~vgg/research/jegal
๐Ÿ“ Repository: https://github.com/Sindhu-Hegde/jegal
๐Ÿ“บ Video: https://www.youtube.com/watch?v=TYFOLKfM-rM

๐Ÿš€ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers

๐Ÿš€ Added to the Human Modeling Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/human-modeling.md

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #CoSpeechGestures #GestureUnderstanding #TriModalRepresentation #MultimodalLearning #AI #ICCV2025 #ResearchHighlight
DmitryRyuminย 
posted an update 5 days ago
view post
Post
3856
๐Ÿš€๐Ÿ’ก๐ŸŒŸ New Research Alert - ICCV 2025 (Oral)! ๐ŸŒŸ๐Ÿช„๐Ÿš€
๐Ÿ“„ Title: LoftUp: Learning a Coordinate-based Feature Upsampler for Vision Foundation Models ๐Ÿ”

๐Ÿ“ Description: LoftUp is a coordinate-based transformer that upscales the low-resolution features of VFMs (e.g. DINOv2 and CLIP) using cross-attention and self-distilled pseudo-ground truth (pseudo-GT) from SAM.

๐Ÿ‘ฅ Authors: Haiwen Huang, Anpei Chen, Volodymyr Havrylov, Andreas Geiger, and Dan Zhang

๐Ÿ“… Conference: ICCV, 19 โ€“ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: LoftUp: Learning a Coordinate-Based Feature Upsampler for Vision Foundation Models (2504.14032)

๐ŸŒ Github Page: https://andrehuang.github.io/loftup-site
๐Ÿ“ Repository: https://github.com/andrehuang/loftup

๐Ÿš€ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers

๐Ÿš€ Added to the Foundation Models and Representation Learning Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/foundation-models-and-representation-learning.md

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #LoftUp #VisionFoundationModels #FeatureUpsampling #Cross-AttentionTransformer #CoordinateBasedLearning #SelfDistillation #PseudoGroundTruth #RepresentationLearning #AI #ICCV2025 #ResearchHighlight
DmitryRyuminย 
posted an update 6 days ago
view post
Post
1891
๐Ÿš€๐Ÿท๏ธ๐ŸŒŸ New Research Alert - ICCV 2025 (Oral)! ๐ŸŒŸ๐Ÿงฉ๐Ÿš€
๐Ÿ“„ Title: Heavy Labels Out! Dataset Distillation with Label Space Lightening ๐Ÿ”

๐Ÿ“ Description: The HeLlO framework is a new corpus distillation method that removes the need for large soft labels. It uses a lightweight, online image-to-label projector based on CLIP. This projector has been adapted using LoRA-style, parameter-efficient tuning. It has also been initialized with text embeddings.

๐Ÿ‘ฅ Authors: @roseannelexie , @Huage001 , Zigeng Chen, Jingwen Ye, and Xinchao Wang

๐Ÿ“… Conference: ICCV, 19 โ€“ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: Heavy Labels Out! Dataset Distillation with Label Space Lightening (2408.08201)

๐Ÿ“บ Video: https://www.youtube.com/watch?v=kAyK_3wskgA

๐Ÿš€ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers

๐Ÿš€ Added to the Efficient Learning Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/efficient-learning.md

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #DatasetDistillation #LabelCompression #CLIP #LoRA #EfficientAI #FoundationModels #AI #ICCV2025 #ResearchHighlight
  • 2 replies
ยท
DmitryRyuminย 
posted an update 7 days ago
view post
Post
4727
๐Ÿš€๐Ÿค–๐ŸŒŸ New Research Alert - ICCV 2025 (Oral)! ๐ŸŒŸ๐Ÿค–๐Ÿš€
๐Ÿ“„ Title: Variance-based Pruning for Accelerating and Compressing Trained Networks ๐Ÿ”

๐Ÿ“ Description: The one-shot pruning method efficiently compresses networks, reducing computation and memory usage while retaining almost full performance and requiring minimal fine-tuning.

๐Ÿ‘ฅ Authors: Uranik Berisha, Jens Mehnert, and Alexandru Paul Condurache

๐Ÿ“… Conference: ICCV, 19 โ€“ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: Variance-Based Pruning for Accelerating and Compressing Trained Networks (2507.12988)

๐Ÿš€ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers

๐Ÿš€ Added to the Efficient Learning Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/efficient-learning.md

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #VarianceBasedPruning #NetworkCompression #ModelAcceleration #EfficientDeepLearning #VisionTransformers #AI #ICCV2025 #ResearchHighlight
DmitryRyuminย 
posted an update 9 days ago
view post
Post
2926
๐Ÿš€๐Ÿ‘๏ธ๐ŸŒŸ New Research Alert - ICCV 2025 (Oral)! ๐ŸŒŸ๐Ÿ‘๏ธ๐Ÿš€
๐Ÿ“„ Title: Token Activation Map to Visually Explain Multimodal LLMs ๐Ÿ”

๐Ÿ“ Description: The Token Activation Map (TAM) is an advanced explainability method for multimodal LLMs. Using causal inference and a Rank Gaussian Filter, TAM reveals token-level interactions and eliminates redundant activations. The result is clearer, high-quality visualizations that enhance understanding of object localization, reasoning and multimodal alignment across models.

๐Ÿ‘ฅ Authors: Yi Li, Hualiang Wang, Xinpeng Ding, Haonan Wang, and Xiaomeng Li

๐Ÿ“… Conference: ICCV, 19 โ€“ 23 Oct, 2025 | Honolulu, Hawai'i, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: Token Activation Map to Visually Explain Multimodal LLMs (2506.23270)

๐Ÿ“ Repository: https://github.com/xmed-lab/TAM

๐Ÿš€ ICCV-2023-25-Papers: https://github.com/DmitryRyumin/ICCV-2023-25-Papers

๐Ÿš€ Added to the Multi-Modal Learning Section: https://github.com/DmitryRyumin/ICCV-2023-25-Papers/blob/main/sections/2025/main/multi-modal-learning.md

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #TokenActivationMap #TAM #CausalInference #VisualReasoning #Multimodal #Explainability #VisionLanguage #LLM #XAI #AI #ICCV2025 #ResearchHighlight
  • 2 replies
ยท
DmitryRyuminย 
posted an update 9 months ago
view post
Post
4084
๐Ÿš€๐ŸŽญ๐ŸŒŸ New Research Alert - WACV 2025 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿš€
๐Ÿ“„ Title: EmoVOCA: Speech-Driven Emotional 3D Talking Heads ๐Ÿ”

๐Ÿ“ Description: EmoVOCA is a data-driven method for generating emotional 3D talking heads by combining speech-driven lip movements with expressive facial dynamics. This method has been developed to overcome the limitations of corpora and to achieve state-of-the-art animation quality.

๐Ÿ‘ฅ Authors: @FedeNoce , Claudio Ferrari, and Stefano Berretti

๐Ÿ“… Conference: WACV, 28 Feb โ€“ 4 Mar, 2025 | Arizona, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: https://arxiv.org/abs/2403.12886

๐ŸŒ Github Page: https://fedenoce.github.io/emovoca/
๐Ÿ“ Repository: https://github.com/miccunifi/EmoVOCA

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #EmoVOCA #3DAnimation #TalkingHeads #SpeechDriven #FacialExpressions #MachineLearning #ComputerVision #ComputerGraphics #DeepLearning #AI #WACV2024
  • 1 reply
ยท
DmitryRyuminย 
posted an update about 1 year ago
view post
Post
3074
๐Ÿ”ฅ๐ŸŽญ๐ŸŒŸ New Research Alert - HeadGAP (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿ”ฅ
๐Ÿ“„ Title: HeadGAP: Few-shot 3D Head Avatar via Generalizable Gaussian Priors ๐Ÿ”

๐Ÿ“ Description: HeadGAP introduces a novel method for generating high-fidelity, animatable 3D head avatars from few-shot data, using Gaussian priors and dynamic part-based modelling for personalized and generalizable results.

๐Ÿ‘ฅ Authors: @zxz267 , @walsvid , @zhaohu2 , Weiyi Zhang, @hellozhuo , Xu Chang, Yang Zhao, Zheng Lv, Xiaoyuan Zhang, @yongjie-zhang-mail , Guidong Wang, and Lan Xu

๐Ÿ“„ Paper: HeadGAP: Few-shot 3D Head Avatar via Generalizable Gaussian Priors (2408.06019)

๐ŸŒ Github Page: https://headgap.github.io

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #HeadGAP #3DAvatar #FewShotLearning #GaussianPriors #AvatarCreation #3DModeling #MachineLearning #ComputerVision #ComputerGraphics #GenerativeAI #DeepLearning #AI
DmitryRyuminย 
posted an update about 1 year ago
view post
Post
2145
๐Ÿš€๐Ÿ•บ๐ŸŒŸ New Research Alert - ECCV 2024 (Avatars Collection)! ๐ŸŒŸ๐Ÿ’ƒ๐Ÿš€
๐Ÿ“„ Title: Expressive Whole-Body 3D Gaussian Avatar ๐Ÿ”

๐Ÿ“ Description: ExAvatar is a model that generates animatable 3D human avatars with facial expressions and hand movements from short monocular videos using a hybrid mesh and 3D Gaussian representation.

๐Ÿ‘ฅ Authors: Gyeongsik Moon, Takaaki Shiratori, and @psyth

๐Ÿ“… Conference: ECCV, 29 Sep โ€“ 4 Oct, 2024 | Milano, Italy ๐Ÿ‡ฎ๐Ÿ‡น

๐Ÿ“„ Paper: MeshAvatar: Learning High-quality Triangular Human Avatars from Multi-view Videos (2407.08414)

๐Ÿ“„ Paper: Expressive Whole-Body 3D Gaussian Avatar (2407.21686)

๐ŸŒ Github Page: https://mks0601.github.io/ExAvatar
๐Ÿ“ Repository: https://github.com/mks0601/ExAvatar_RELEASE

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #ExAvatar #3DAvatar #FacialExpressions #HandMotions #MonocularVideo #3DModeling #GaussianSplatting #MachineLearning #ComputerVision #ComputerGraphics #DeepLearning #AI #ECCV2024
DmitryRyuminย 
posted an update about 1 year ago
view post
Post
1867
๐Ÿ”ฅ๐ŸŽญ๐ŸŒŸ New Research Alert - ECCV 2024 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿ”ฅ
๐Ÿ“„ Title: MeshAvatar: Learning High-quality Triangular Human Avatars from Multi-view Videos ๐Ÿ”

๐Ÿ“ Description: MeshAvatar is a novel pipeline that generates high-quality triangular human avatars from multi-view videos, enabling realistic editing and rendering through a mesh-based approach with physics-based decomposition.

๐Ÿ‘ฅ Authors: Yushuo Chen, Zerong Zheng, Zhe Li, Chao Xu, and Yebin Liu

๐Ÿ“… Conference: ECCV, 29 Sep โ€“ 4 Oct, 2024 | Milano, Italy ๐Ÿ‡ฎ๐Ÿ‡น

๐Ÿ“„ Paper: MeshAvatar: Learning High-quality Triangular Human Avatars from Multi-view Videos (2407.08414)

๐ŸŒ Github Page: https://shad0wta9.github.io/meshavatar-page
๐Ÿ“ Repository: https://github.com/shad0wta9/meshavatar

๐Ÿ“บ Video: https://www.youtube.com/watch?v=Kpbpujkh2iI

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #MeshAvatar #3DAvatars #MultiViewVideo #PhysicsBasedRendering #TriangularMesh #AvatarCreation #3DModeling #NeuralRendering #Relighting #AvatarEditing #MachineLearning #ComputerVision #ComputerGraphics #DeepLearning #AI #ECCV2024
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
2330
๐Ÿš€๐Ÿ•บ๐ŸŒŸ New Research Alert - CVPR 2024 (Avatars Collection)! ๐ŸŒŸ๐Ÿ’ƒ๐Ÿš€
๐Ÿ“„ Title: IntrinsicAvatar: Physically Based Inverse Rendering of Dynamic Humans from Monocular Videos via Explicit Ray Tracing ๐Ÿ”

๐Ÿ“ Description: IntrinsicAvatar is a method for extracting high-quality geometry, albedo, material, and lighting properties of clothed human avatars from monocular videos using explicit ray tracing and volumetric scattering, enabling realistic animations under varying lighting conditions.

๐Ÿ‘ฅ Authors: Shaofei Wang, Boลพidar Antiฤ‡, Andreas Geiger, and Siyu Tang

๐Ÿ“… Conference: CVPR, Jun 17-21, 2024 | Seattle WA, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ”— Paper: IntrinsicAvatar: Physically Based Inverse Rendering of Dynamic Humans from Monocular Videos via Explicit Ray Tracing (2312.05210)

๐ŸŒ Github Page: https://neuralbodies.github.io/IntrinsicAvatar/
๐Ÿ“ Repository: https://github.com/taconite/IntrinsicAvatar

๐Ÿ“บ Video: https://www.youtube.com/watch?v=aS8AIxgVXzI

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #IntrinsicAvatar #InverseRendering #MonocularVideos #RayTracing #VolumetricScattering #3DReconstruction #MachineLearning #ComputerVision #DeepLearning #AI #CVPR2024
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
3151
๐Ÿ”ฅ๐ŸŽญ๐ŸŒŸ New Research Alert - ECCV 2024 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿ”ฅ
๐Ÿ“„ Title: RodinHD: High-Fidelity 3D Avatar Generation with Diffusion Models ๐Ÿ”

๐Ÿ“ Description: RodinHD generates high-fidelity 3D avatars from portrait images using a novel data scheduling strategy and weight consolidation regularization to capture intricate details such as hairstyles.

๐Ÿ‘ฅ Authors: Bowen Zhang, @yiji , @chunyuwang , Ting Zhang, @jiaolong , Yansong Tang, Feng Zhao, Dong Chen, and Baining Guo

๐Ÿ“… Conference: ECCV, 29 Sep โ€“ 4 Oct, 2024 | Milano, Italy ๐Ÿ‡ฎ๐Ÿ‡น

๐Ÿ“„ Paper: RodinHD: High-Fidelity 3D Avatar Generation with Diffusion Models (2407.06938)

๐ŸŒ Github Page: https://rodinhd.github.io/
๐Ÿ“ Repository: https://github.com/RodinHD/RodinHD

๐Ÿ“บ Video: https://www.youtube.com/watch?v=ULvHt7dZx-Q

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #RodinHD #3DAvatars #DiffusionModels #HighFidelity #PortraitTo3D #MachineLearning #ComputerVision #DeepLearning #AI #ECCV2024
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
2440
๐Ÿ”ฅ๐ŸŽญ๐ŸŒŸ New Research Alert - LivePortrait (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿ”ฅ
๐Ÿ“„ Title: LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control ๐Ÿ”

๐Ÿ“ Description: LivePortrait is an efficient video-driven portrait animation framework that uses implicit keypoints and stitching/retargeting modules to generate high-quality, controllable animations from a single source image.

๐Ÿ‘ฅ Authors: @cleardusk , Dingyun Zhang, Xiaoqiang Liu, Zhizhou Zhong, Yuan Zhang, Pengfei Wan, and Di Zhang

๐Ÿค— Demo: KwaiVGI/LivePortrait

๐Ÿ“„ Paper: LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control (2407.03168)

๐ŸŒ Github Page: https://liveportrait.github.io/
๐Ÿ“ Repository: https://github.com/KwaiVGI/LivePortrait

๐Ÿ”ฅ Model ๐Ÿค–: KwaiVGI/LivePortrait

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #LivePortrait #PortraitAnimation #ComputerVision #MachineLearning #DeepLearning #ComputerGraphics #FacialAnimation #GenerativeAI #RealTimeRendering #AI
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
2732
๐Ÿš€๐Ÿ•บ๐ŸŒŸ New Research Alert (Avatars Collection)! ๐ŸŒŸ๐Ÿ’ƒ๐Ÿš€
๐Ÿ“„ Title: Expressive Gaussian Human Avatars from Monocular RGB Video ๐Ÿ”

๐Ÿ“ Description: The new EVA model enhances the expressiveness of digital avatars by using 3D Gaussians and SMPL-X to capture fine-grained hand and face details from monocular RGB video.

๐Ÿ‘ฅ Authors: Hezhen Hu, Zhiwen Fan, Tianhao Wu, Yihan Xi, Seoyoung Lee, Georgios Pavlakos, and Zhangyang Wang

๐Ÿ“„ Paper: Expressive Gaussian Human Avatars from Monocular RGB Video (2407.03204)

๐ŸŒ Github Page: https://evahuman.github.io/
๐Ÿ“ Repository: https://github.com/evahuman/EVA

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #DigitalAvatars #3DModeling #ComputerVision #MonocularVideo #SMPLX #3DGaussians #AvatarExpressiveness #HandTracking #FacialExpressions #AI #MachineLearning
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
2069
๐Ÿ”ฅ๐ŸŽญ๐ŸŒŸ New Research Alert - ECCV 2024 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿ”ฅ
๐Ÿ“„ Title: Topo4D: Topology-Preserving Gaussian Splatting for High-Fidelity 4D Head Capture ๐Ÿ”

๐Ÿ“ Description: Topo4D is a novel method for automated, high-fidelity 4D head tracking that optimizes dynamic topological meshes and 8K texture maps from multi-view time-series images.

๐Ÿ‘ฅ Authors: @Dazz1e , Y. Cheng, @Ryan-sjtu , H. Jia, D. Xu, W. Zhu, Y. Yan

๐Ÿ“… Conference: ECCV, 29 Sep โ€“ 4 Oct, 2024 | Milano, Italy ๐Ÿ‡ฎ๐Ÿ‡น

๐Ÿ“„ Paper: Topo4D: Topology-Preserving Gaussian Splatting for High-Fidelity 4D Head Capture (2406.00440)

๐ŸŒ Github Page: https://xuanchenli.github.io/Topo4D/
๐Ÿ“ Repository: https://github.com/XuanchenLi/Topo4D

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿš€ WACV-2024-Papers: https://github.com/DmitryRyumin/WACV-2024-Papers

๐Ÿš€ ICCV-2023-Papers: https://github.com/DmitryRyumin/ICCV-2023-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #Topo4D #4DHead #3DModeling #4DCapture #FacialAnimation #ComputerGraphics #MachineLearning #HighFidelity #TextureMapping #DynamicMeshes #GaussianSplatting #VisualEffects #ECCV2024
  • 1 reply
ยท
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
3720
๐Ÿš€๐ŸŽญ๐ŸŒŸ New Research Alert - Portrait4D-v2 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿš€
๐Ÿ“„ Title: Portrait4D-v2: Pseudo Multi-View Data Creates Better 4D Head Synthesizer ๐Ÿ”

๐Ÿ“ Description: Portrait4D-v2 is a novel method for one-shot 4D head avatar synthesis using pseudo multi-view videos and a vision transformer backbone, achieving superior performance without relying on 3DMM reconstruction.

๐Ÿ‘ฅ Authors: Yu Deng, Duomin Wang, and Baoyuan Wang

๐Ÿ“„ Paper: Portrait4D-v2: Pseudo Multi-View Data Creates Better 4D Head Synthesizer (2403.13570)

๐ŸŒ GitHub Page: https://yudeng.github.io/Portrait4D-v2/
๐Ÿ“ Repository: https://github.com/YuDeng/Portrait-4D

๐Ÿ“บ Video: https://www.youtube.com/watch?v=5YJY6-wcOJo

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: Portrait4D #4DAvatar #HeadSynthesis #3DModeling #TechInnovation #DeepLearning #ComputerGraphics #ComputerVision #Innovation
  • 1 reply
ยท
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
2390
๐Ÿ˜€๐Ÿ˜ฒ๐Ÿ˜๐Ÿ˜ก New Research Alert - CVPRW 2024 (Facial Expressions Recognition Collection)! ๐Ÿ˜ก๐Ÿ˜ฅ๐Ÿฅด๐Ÿ˜ฑ
๐Ÿ“„ Title: Zero-Shot Audio-Visual Compound Expression Recognition Method based on Emotion Probability Fusion ๐Ÿ”

๐Ÿ“ Description: AVCER is a novel audio-visual method for compound expression recognition based on pair-wise sum of emotion probability, evaluated in multi- and cross-corpus setups without task-specific training data, demonstrating its potential for intelligent emotion annotation tools.

๐Ÿ‘ฅ Authors: @ElenaRyumina , Maxim Markitantov, @DmitryRyumin , Heysem Kaya, and Alexey Karpov

๐Ÿ“… Conference: CVPRW, Jun 17-21, 2024 | Seattle WA, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿค— Demo: ElenaRyumina/AVCER

๐Ÿ“„ Paper: Audio-Visual Compound Expression Recognition Method based on Late Modality Fusion and Rule-based Decision (2403.12687)

๐ŸŒ Github Page: https://elenaryumina.github.io/AVCER
๐Ÿ“ Repository: https://github.com/ElenaRyumina/AVCER/tree/main/src

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Facial Expressions Recognition Collection: DmitryRyumin/facial-expressions-recognition-65f22574e0724601636ddaf7

๐Ÿ” Keywords: #AVCER #AudioVisual #CompoundExpressions #EmotionRecognition #ModalityFusion #RuleBasedAI #ABAWCompetition #AIResearch #HumanEmotion #IntelligentTools #MachineLearning #DeepLearning #MultiCorpus #CrossCorpus #CVPR2024
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
1877
๐Ÿš€๐ŸŽญ๐ŸŒŸ New Research Alert - CVPR 2024 (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿš€
๐Ÿ“„ Title: Relightable Gaussian Codec Avatars ๐Ÿ”

๐Ÿ“ Description: Relightable Gaussian Codec Avatars is a method for creating highly detailed and relightable 3D head avatars that can animate expressions in real time and support complex features such as hair and skin with efficient rendering suitable for VR.

๐Ÿ‘ฅ Authors: @psyth , @GBielXONE02 , Tomas Simon, Junxuan Li, and @giljoonam

๐Ÿ“… Conference: CVPR, Jun 17-21, 2024 | Seattle WA, USA ๐Ÿ‡บ๐Ÿ‡ธ

๐Ÿ“„ Paper: Relightable Gaussian Codec Avatars (2312.03704)

๐ŸŒ GitHub Page: https://shunsukesaito.github.io/rgca/

๐Ÿš€ CVPR-2023-24-Papers: https://github.com/DmitryRyumin/CVPR-2023-24-Papers

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #3DAvatars #RealTimeRendering #RelightableAvatars #3DModeling #VirtualReality #CVPR2024 #DeepLearning #ComputerGraphics #ComputerVision #Innovation #VR
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
870
๐Ÿš€๐ŸŽญ๐ŸŒŸ New Research Alert - InstructAvatar (Avatars Collection)! ๐ŸŒŸ๐ŸŽญ๐Ÿš€
๐Ÿ“„ Title: InstructAvatar: Text-Guided Emotion and Motion Control for Avatar Generation ๐Ÿ”

๐Ÿ“ Description: InstructAvatar is a novel method for generating emotionally expressive 2D avatars using text-guided instructions, offering improved emotion control, lip-sync quality, and naturalness. It uses a two-branch diffusion-based generator to predict avatars based on both audio and text input.

๐Ÿ‘ฅ Authors: Yuchi Wang et al.

๐Ÿ“„ Paper: InstructAvatar: Text-Guided Emotion and Motion Control for Avatar Generation (2405.15758)

๐ŸŒ Github Page: https://wangyuchi369.github.io/InstructAvatar/
๐Ÿ“ Repository: https://github.com/wangyuchi369/InstructAvatar

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿš€ Added to the Avatars Collection: DmitryRyumin/avatars-65df37cdf81fec13d4dbac36

๐Ÿ” Keywords: #InstructAvatar #AvatarGeneration #EmotionControl #FacialMotion #LipSynchronization #NaturalLanguageInterface #DiffusionBasedGenerator #TextGuidedInstructions #2DAvatars #VideoSynthesis #Interactivity #ComputerGraphics #DeepLearning #ComputerVision #Innovation
DmitryRyuminย 
posted an update over 1 year ago
view post
Post
1521
๐Ÿ”ฅ๐Ÿš€๐ŸŒŸ New Research Alert - YOLOv10! ๐ŸŒŸ๐Ÿš€๐Ÿ”ฅ
๐Ÿ“„ Title: YOLOv10: Real-Time End-to-End Object Detection ๐Ÿ”

๐Ÿ“ Description: YOLOv10 improves real-time object recognition by eliminating non-maximum suppression and optimizing the model architecture to achieve state-of-the-art performance with lower latency and computational overhead.

๐Ÿ‘ฅ Authors: Ao Wang et al.

๐Ÿ“„ Paper: YOLOv10: Real-Time End-to-End Object Detection (2405.14458)

๐Ÿค— Demo: kadirnar/Yolov10 curated by @kadirnar
๐Ÿ”ฅ Model ๐Ÿค–: kadirnar/Yolov10

๐Ÿ“ Repository: https://github.com/THU-MIG/yolov10

๐Ÿ“ฎ Post about YOLOv9 - https://huggingface.co/posts/DmitryRyumin/519784698531054

๐Ÿ“š More Papers: more cutting-edge research presented at other conferences in the DmitryRyumin/NewEraAI-Papers curated by @DmitryRyumin

๐Ÿ” Keywords: #YOLOv10 #ObjectDetection #RealTimeAI #ModelOptimization #MachineLearning #DeepLearning #ComputerVision #Innovation
  • 1 reply
ยท