GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives

Challenges in learning scene graphs from natural language description.

Abstract

Training Scene Graph Generation (SGG) models with natural language captions has become increasingly popular due to the abundant, cost-effective, and open-world generalization supervision signals that natural language offers. However, such unstructured caption data and its processing pose significant challenges in learning accurate and comprehensive scene graphs. The challenges can be summarized as three aspects:

Aiming to address these problems, we propose a divide-and-conquer strategy with a novel framework named GPT4SGG, to obtain more accurate and comprehensive scene graph signals. This framework decomposes a complex scene into a bunch of simple regions, resulting in a set of region-specific narratives. With these region-specific narratives (partial observations) and a holistic narrative (global observation) for an image, a large language model (LLM) performs the relationship reasoning to synthesize an accurate and comprehensive scene graph.

Method

Overview of <b>GPT4SGG</b>.

Textual representation of image data: localised objects, holistic & region-specific narratives.

Task-specific (SGG-aware) Prompt: synthesize scene graphs based on the textual input for image data.

Comparison with state-of-the-arts on VG150 test set, diamond symbol marks fully supervised methods

comparison with sota

Example of GPT4SGG

gpt4sgg-example

Samples of COCO-SG@GPT

gpt4sgg-samples

BibTeX

Please cite GPT4SGG in your publications if it helps your research:


@misc{chen2023gpt4sgg,
      title={GPT4SGG: Synthesizing Scene Graphs from Holistic and Region-specific Narratives}, 
      author={Zuyao Chen and Jinlin Wu and Zhen Lei and Zhaoxiang Zhang and Changwen Chen},
      year={2023},
      eprint={2312.04314},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}