Language-Grounded Multi-Domain Image Translation via Semantic Difference Guidance
By: Jongwon Ryu , Joonhyung Park , Jaeho Han and more
Multi-domain image-to-image translation re quires grounding semantic differences ex pressed in natural language prompts into corresponding visual transformations, while preserving unrelated structural and seman tic content. Existing methods struggle to maintain structural integrity and provide fine grained, attribute-specific control, especially when multiple domains are involved. We propose LACE (Language-grounded Attribute Controllable Translation), built on two compo nents: (1) a GLIP-Adapter that fuses global semantics with local structural features to pre serve consistency, and (2) a Multi-Domain Control Guidance mechanism that explicitly grounds the semantic delta between source and target prompts into per-attribute translation vec tors, aligning linguistic semantics with domain level visual changes. Together, these modules enable compositional multi-domain control with independent strength modulation for each attribute. Experiments on CelebA(Dialog) and BDD100K demonstrate that LACE achieves high visual fidelity, structural preservation, and interpretable domain-specific control, surpass ing prior baselines. This positions LACE as a cross-modal content generation framework bridging language semantics and controllable visual translation.
Similar Papers
Language as an Anchor: Preserving Relative Visual Geometry for Domain Incremental Learning
CV and Pattern Recognition
Keeps AI smart when learning new things.
Exploiting Domain Properties in Language-Driven Domain Generalization for Semantic Segmentation
CV and Pattern Recognition
Helps computers see objects in new places.
Constrained Prompt Enhancement for Improving Zero-Shot Generalization of Vision-Language Models
CV and Pattern Recognition
Helps computers understand pictures and words better.