Draw Me Like My Triples: Leveraging Generative AI for Wikidata Image Completion

Raia Abu Ahmad, Martin Critelli, Sefika Efeoglu, Eleonora Mancini, Célian Ringwald, Xinyue Zhang, Albert Meroño-Peñuela

Research output: Contribution to journalArticlepeer-review

Abstract

Humans are critical for the creation and maintenance of high-quality Knowledge Graphs (KGs). However, creating and maintaining large KGs only with humans does not scale, especially for contributions based on multimedia (e.g. images) that are hard to find and reuse on the Web and expensive to generate by humans from scratch. Therefore, we leverage generative AI for the task of creating images for Wikidata items that do not have them. Our approach uses knowledge contained in Wikidata triples of items describing fictional characters and uses the fine-tuned T5 model based on the WDV dataset to generate natural text descriptions of items about fictional characters with missing images. We use those natural text descriptions as prompts for a transformer-based text-to-image model, Stable Diffusion v2.1, to generate plausible candidate images for Wikidata image completion. We design and implement quantitative and qualitative approaches to evaluate the plausibility of our methods, which include conducting a survey to assess the quality of the generated images.
Original languageEnglish
JournalCEUR Workshop Proceedings
Volume3640
Publication statusPublished - 1 Jan 2023

Keywords

  • Automated Prompt Generation
  • Generative AI
  • Image Generation

Fingerprint

Dive into the research topics of 'Draw Me Like My Triples: Leveraging Generative AI for Wikidata Image Completion'. Together they form a unique fingerprint.

Cite this