Investigating Identity Signals in Conversational Facial Dynamics via Disentangled Expression Features
By: Masoumeh Chapariniya , Pierre Vuillecard , Jean-Marc Odobez and more
Potential Business Impact:
Recognizes people by how their faces move.
This work investigates whether individuals can be identified solely through the pure dynamical components of their facial expressions, independent of static facial appearance. We leverage the FLAME 3D morphable model to achieve explicit disentanglement between facial shape and expression dynamics, extracting frame-by-frame parameters from conversational videos while retaining only expression and jaw coefficients. On the CANDOR dataset of 1,429 speakers in naturalistic conversations, our Conformer model with supervised contrastive learning achieves 61.14\%accuracy on 1,429-way classification -- 458 times above chance -- demonstrating that facial dynamics carry strong identity signatures. We introduce a drift-to-noise ratio (DNR) that quantifies the reliability of shape expression separation by measuring across-session shape changes relative to within-session variability. DNR strongly negatively correlates with recognition performance, confirming that unstable shape estimation compromises dynamic identification. Our findings reveal person-specific signatures in conversational facial dynamics, with implications for social perception and clinical assessment.
Similar Papers
Beyond Appearance: Transformer-based Person Identification from Conversational Dynamics
CV and Pattern Recognition
Identifies people by how they move and stand.
Learning Disentangled Speech- and Expression-Driven Blendshapes for 3D Talking Face Animation
CV and Pattern Recognition
Makes computer faces show real feelings when talking.
Disentangle Identity, Cooperate Emotion: Correlation-Aware Emotional Talking Portrait Generation
CV and Pattern Recognition
Makes talking videos show real feelings, not fake ones.