Salience-SGG: Enhancing Unbiased Scene Graph Generation with Iterative Salience Estimation
By: Runfeng Qu , Ole Hall , Pia K Bideau and more
Scene Graph Generation (SGG) suffers from a long-tailed distribution, where a few predicate classes dominate while many others are underrepresented, leading to biased models that underperform on rare relations. Unbiased-SGG methods address this issue by implementing debiasing strategies, but often at the cost of spatial understanding, resulting in an over-reliance on semantic priors. We introduce Salience-SGG, a novel framework featuring an Iterative Salience Decoder (ISD) that emphasizes triplets with salient spatial structures. To support this, we propose semantic-agnostic salience labels guiding ISD. Evaluations on Visual Genome, Open Images V6, and GQA-200 show that Salience-SGG achieves state-of-the-art performance and improves existing Unbiased-SGG methods in their spatial understanding as demonstrated by the Pairwise Localization Average Precision
Similar Papers
Unbiased Video Scene Graph Generation via Visual and Semantic Dual Debiasing
CV and Pattern Recognition
Helps videos show what's really happening.
SAGE: Saliency-Guided Contrastive Embeddings
CV and Pattern Recognition
Teaches computers to see what humans see.
VOST-SGG: VLM-Aided One-Stage Spatio-Temporal Scene Graph Generation
CV and Pattern Recognition
Helps videos understand what's happening and why.