Generalized Visual Relation Detection with Diffusion Models
Visual relation detection (VRD) aims to identify relationships (or interactions) between object pairs in an image. Although recent VRD models have achieved impressive performance, they are all restricted to pre-defined relation categories, while failing to consider the semantic ambiguity characteris...
Saved in:
Main Authors | , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
16.04.2025
|
Subjects | |
Online Access | Get full text |
DOI | 10.48550/arxiv.2504.12100 |
Cover
Loading…
Summary: | Visual relation detection (VRD) aims to identify relationships (or
interactions) between object pairs in an image. Although recent VRD models have
achieved impressive performance, they are all restricted to pre-defined
relation categories, while failing to consider the semantic ambiguity
characteristic of visual relations. Unlike objects, the appearance of visual
relations is always subtle and can be described by multiple predicate words
from different perspectives, e.g., ``ride'' can be depicted as ``race'' and
``sit on'', from the sports and spatial position views, respectively. To this
end, we propose to model visual relations as continuous embeddings, and design
diffusion models to achieve generalized VRD in a conditional generative manner,
termed Diff-VRD. We model the diffusion process in a latent space and generate
all possible relations in the image as an embedding sequence. During the
generation, the visual and text embeddings of subject-object pairs serve as
conditional signals and are injected via cross-attention. After the generation,
we design a subsequent matching stage to assign the relation words to
subject-object pairs by considering their semantic similarities. Benefiting
from the diffusion-based generative process, our Diff-VRD is able to generate
visual relations beyond the pre-defined category labels of datasets. To
properly evaluate this generalized VRD task, we introduce two evaluation
metrics, i.e., text-to-image retrieval and SPICE PR Curve inspired by image
captioning. Extensive experiments in both human-object interaction (HOI)
detection and scene graph generation (SGG) benchmarks attest to the superiority
and effectiveness of Diff-VRD. |
---|---|
DOI: | 10.48550/arxiv.2504.12100 |