Datasets:

Modalities:
Image
Text
Formats:
json
ArXiv:
Libraries:
Datasets
Dask
License:
vqa / README.md
gentaiscool's picture
Update README.md
de41a4a verified
metadata
license: cc-by-sa-4.0
language:
  - eng
  - ind
  - ind
  - zho
  - kor
  - kor
  - jpn
  - jpn
  - sun
  - jav
  - jav
  - ces
  - spa
  - fra
  - ara
  - hin
  - ben
  - mar
  - sin
  - yor
  - yue
  - nan
  - nan
  - tgl
  - tha
  - aze
  - rus
  - rus
  - ita
  - srd
multilinguality:
  - multilingual
language_details: >-
  en, id_formal, id_casual, zh_cn, ko_formal, ko_casual, ja_formal, ja_casual,
  su_loma, jv_krama, jv_ngoko, cs, es, fr, ar, hi, bn, mr, si_formal_spoken, yo,
  yue, nan, nan_spoken, tl, th, az, ru_formal, ru_casual, it, sc
configs:
  - config_name: task1
    data_files:
      - split: test_large
        path: hf_prompt/large_eval_task1/*
      - split: test_small
        path: hf_prompt/small_eval_task1/*
      - split: train
        path: hf_prompt/train_task1/*
  - config_name: task2
    data_files:
      - split: test_large
        path: hf_prompt/large_eval_task2/*
      - split: test_small
        path: hf_prompt/small_eval_task2/*
      - split: train
        path: hf_prompt/train_task2/*

WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines

Overview

WorldCuisines is a massive-scale visual question answering (VQA) benchmark for multilingual and multicultural understanding through global cuisines. The dataset contains text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark as of 17 October 2024.

Overview

We develop both a VQA dataset (WC-VQA) and a curated Knowledge Base (KB) for world cuisines (WC-KB). The WC-VQA dataset is constructed using WC-KB, which serves as the primary data source. We design two tasks as follows:

Task 1: Dish Name Prediction

This task requires predicting the name of a dish based on its image, a question, and contextual information. It is divided into three subtasks, each with a distinct query type:

  • (a) No-context question: Predict the dish name without additional context.
  • (b) Contextualized question: Predict the dish name with provided contextual information.
  • (c) Adversarial contextualized question: Predict the dish name with misleading or adversarial context.

Task 2: Location Prediction

This task involves predicting the location where the dish is commonly consumed or originated, based on the dish image, a question, and contextual information.

WC-KB encompasses 2,414 dishes worldwide, including 6,045 images and metadata, covering both coarse-grained (e.g., stew) and fine-grained categories (e.g., beef stew), locations, and regional cuisines. It also features multilingual translations of 90 crowd-sourced prompt templates and 401 parallel data entries (i.e., multilingual information) for location and regional cuisine information. From WC-KB, we construct WC-VQA, a multilingual parallel VQA dataset with 1 million samples encompassing over 30 languages and dialects, including various varieties and registers, such as formal and casual styles, with high-quality human annotations. The VQA is designed to evaluate models' ability to understand cultural food names and their origins.

We provide WC-VQA evaluation datasets in two sizes (12,000 and 60,000 instances) alongside a training dataset (1,080,000 instances). The table below provides more detailed statistics regarding the number of VQA instances and images for each data split.

StatisticsTable

Dataset Construction

Our data sources are gathered from Wikipedia and Wikimedia Commons to ensure they can be easily redistributed under an accepted open-source license. The data construction process involves four key steps:

  1. Dish selection
  2. Metadata annotation
  3. Quality assurance
  4. Data compilation.

Dish Selection

We compile a comprehensive list of dish names sourced from Wikipedia. We manually review pages that feature lists of dishes to determine whether each dish is a specialty unique to a specific culture, as we aim to focus on dishes that have distinct cultural significance. We exclude generic categories, such as ice cream, which lacks a specific cultural association. We ensure that each dish on our list has its own dedicated Wikipedia page. If a dish does not have a Wikipedia page, it is also excluded from our compilation. This meticulous approach ensures that our dataset is both culturally relevant and well-documented.

Metadata Annotation

Given a dish name and its corresponding Wikipedia page link, annotators manually compile metadata based on the provided information. This metadata includes:

  • Visual Representation: Images sourced from Wikimedia Commons are included, along with their license information.
  • Categorization: Dishes are classified into both coarse-grained (e.g., rice, bread) and fine-grained (e.g., fried rice, flatbread) categories.
  • Description: Annotators provide a description of each dish based on the content from its Wikipedia page, avoiding the use of the dish's name, origin, or any distinctive keywords that uniquely identify the dish.
  • Cuisine: The dish's origin cuisine and any cuisines with which it is strongly associated.
  • Geographic Distribution: This includes the dish's associated countries, area (city or region), and broader continental region.

Quality Assurance

Before beginning the quality assurance process, we first identify common issues that arise during annotation and develop automated rules to detect easily identifiable annotation errors, such as incorrect string formatting. Annotators are then asked to correct these errors. To further ensure data quality and validity, we conduct several rounds of quality assurance:

  1. Image Quality: We remove instances where images are blurry, dark, or contain distracting elements such as people or other dishes. We also verify image licenses by cross-referencing them with information on Wikimedia Commons.
  2. Categorization and Descriptions: We refine dish categorization and descriptions, ensuring consistency in category assignments and keeping descriptions free from "information breaches" (e.g., excluding regional details from the description).
  3. Cuisine Names and Geographic Accuracy: We standardize cuisine names and meticulously review all country and area information for accuracy.

This comprehensive approach guarantees the integrity and reliability of our dataset.

Data Compilation

In this phase, we verify the overall quality check done by annotators, identifying any potential inconsistencies missed during quality assurance. We then compile the dataset by collecting the metadata into a single file.

VQA Generation

In this phase, we generate VQA data by sampling from WC-KB. A VQA data entry comprises a visual image, question text, and answer text. This process involves four stages:

  1. Conducting a similarity search for dish names
  2. Constructing questions and contexts
  3. Translating these elements into multiple languages
  4. Generating the VQA triplets.

Dish Names Similarity Search

To identify similar dishes in our dataset, we follow the approach from Winata et al. (2024) to employ a multilingual model E5$\text{LARGE}$ Instruct (Wang et al. 2024) for computing text embeddings. Formally, given a dish $x$ with name $x{\text{name}}$ and text description $x_{\text{desc}}$, we use a multilingual model $\theta$ to compute the embedding vector $v_x = \theta({x_\text{name};x_\text{desc}})$, and then apply cosine similarity to compute a score $s = \text{similarity}(v_i, v_j)$ between dish $i$ and dish $j$. For each dish, we consider the top-$k$ most similar dishes to generate distractors in the multiple choice questions.

Question and Context Construction

Dish name prediction (Task 1) is divided into three question variations depending on the context:

  1. No-context question: Simply asks for the name of the dish without any provided context.
  2. Contextualized question: Provides additional information related to cuisine or location.
  3. Adversarial contextualized question: Similar to contextualized questions, but may include misleading location information to assess the model's robustness to irrelevant details.

For regional cuisine prediction (Task 2), only a basic question without any provided context is available.

Multiple Language Translation

Question and Context

All questions and contexts are initially collected in English and are then carefully translated by native speakers into 30 language varieties: 23 different languages, with 7 languages having two varieties. Translators prioritize naturalness, followed by diversity in translations when duplication occurs.

Food Name Alias

Using Wikipedia pages as our primary source, we verify if the English page has translations available in other languages. This enables us to extract dish names in multiple languages and compile them as translations for each dish. We utilize both the Wikipedia page titles in various languages and alias text from the English page. These translations enhance cultural relevance and accuracy for multilingual prompts. When translation is unavailable, we use the English name as the default.

Locations and Cuisines

With over 400 unique locations, including countries, cities, and areas, we first translate the English locations into other languages using GPT-4, followed by proofreading by native speakers. Regional cuisine names (the adjective form of the location in English) are translated in the same manner.

Morphological Inflections

In languages with rich inflectional morphology (e.g., Czech or Spanish), words are modified to express different grammatical categories (e.g., number, gender, or case). We provide a framework for human translators to use natural inflections in the prompt template while keeping the inflections as few as possible.

Generating VQA Triplets

To ensure no overlap between training and testing subsets, we split the dishes and multilingual questions into two subsets. For each subset, we randomly sample dish and question pairs. We use the dish entry in our KB dataset to select the image and inject the location into the context, if applicable. Answer candidates for multiple-choice questions are picked using the similarity search process. We repeat this process until the desired number of training or test samples is reached, discarding duplicates.

Ethical Considerations

Our research focuses on evaluating VLMs within the context of multilingual and multicultural VQA, a field that holds significant implications for diverse multilingual communities. We are committed to conducting our data collection and evaluations with the highest standards of transparency and fairness. To achieve this, we have adopted a crowd-sourcing approach for the annotation process, inviting volunteers to contribute and become co-authors if they provide significant contributions. We follow the guidelines from ACL for authorship eligibility as shown in https://www.aclweb.org/adminwiki/index.php/Authorship_Changes_Policy_for_ACL_Conference_Papers. In line with our commitment to openness and collaboration, we will release our dataset under an open-source license, CC-BY-SA 4.0.

Contact

E-mail: Genta Indra Winata and Frederikus Hudi

Citation

If you find this dataset useful, please cite the following works

@misc{winata2024worldcuisinesmassivescalebenchmarkmultilingual,
      title={WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines}, 
      author={Genta Indra Winata and Frederikus Hudi and Patrick Amadeus Irawan and David Anugraha and Rifki Afina Putri and Yutong Wang and Adam Nohejl and Ubaidillah Ariq Prathama and Nedjma Ousidhoum and Afifa Amriani and Anar Rzayev and Anirban Das and Ashmari Pramodya and Aulia Adila and Bryan Wilie and Candy Olivia Mawalim and Ching Lam Cheng and Daud Abolade and Emmanuele Chersoni and Enrico Santus and Fariz Ikhwantri and Garry Kuwanto and Hanyang Zhao and Haryo Akbarianto Wibowo and Holy Lovenia and Jan Christian Blaise Cruz and Jan Wira Gotama Putra and Junho Myung and Lucky Susanto and Maria Angelica Riera Machin and Marina Zhukova and Michael Anugraha and Muhammad Farid Adilazuarda and Natasha Santosa and Peerat Limkonchotiwat and Raj Dabre and Rio Alexander Audino and Samuel Cahyawijaya and Shi-Xiong Zhang and Stephanie Yulia Salim and Yi Zhou and Yinxuan Gui and David Ifeoluwa Adelani and En-Shiun Annie Lee and Shogo Okada and Ayu Purwarianti and Alham Fikri Aji and Taro Watanabe and Derry Tanti Wijaya and Alice Oh and Chong-Wah Ngo},
      year={2024},
      eprint={2410.12705},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2410.12705}, 
}