F-Actor: Controllable Conversational Behaviour in Full-Duplex Models
By: Maike Züfle , Ondrej Klejch , Nicholas Sanders and more
Potential Business Impact:
Makes talking computers act more like real people.
Spoken conversational systems require more than accurate speech generation to have human-like conversations: to feel natural and engaging, they must produce conversational behaviour that adapts dynamically to the context. Current spoken conversational systems, however, rarely allow such customization, limiting their naturalness and usability. In this work, we present the first open, instruction-following full-duplex conversational speech model that can be trained efficiently under typical academic resource constraints. By keeping the audio encoder frozen and finetuning only the language model, our model requires just 2,000 hours of data, without relying on large-scale pretraining or multi-stage optimization. The model can follow explicit instructions to control speaker voice, conversation topic, conversational behaviour (e.g., backchanneling and interruptions), and dialogue initiation. We propose a single-stage training protocol and systematically analyze design choices. Both the model and training code will be released to enable reproducible research on controllable full-duplex speech systems.
Similar Papers
Enabling Conversational Behavior Reasoning Capabilities in Full-Duplex Speech
Computation and Language
Helps computers understand conversations like people.
From Turn-Taking to Synchronous Dialogue: A Survey of Full-Duplex Spoken Language Models
Computation and Language
Lets AI talk and listen at the same time.
FLM-Audio: Natural Monologues Improves Native Full-Duplex Chatbots via Dual Training
Sound
Lets chatbots talk and listen at once.