×
A new multilingual evaluation benchmark for the visual question answering task. We extend the established English GQA dataset to 7 typologically diverse ...
Sep 13, 2021 · A new multilingual evaluation benchmark for the visual question answering task. We extend the established English GQA dataset to 7 typologically diverse ...
A new dataset for real-world visual reasoning and compositional question answering. The training data can be downloaded here.
May 22, 2022 · Aiming to address this gap, in this paper we pro- pose xGQA, a multilingual evaluation benchmark for the visual question answering task, ...
The current state-of-the-art on xGQA is UC2. See a full comparison of 4 papers with code.
People also ask
This paper proposes Regularized Contrastive Cross-lingual Cross-modal (RC^3) pre-training, which further exploits more abundant weakly-aligned multilingual ...
Jun 4, 2024 · We use xGQA (Pfeiffer et al., 2022) , a representative benchmark for the cross-lingual VQA task. Each sample in the dataset consists of an image ...
xGQA [22] is a multi-lingual evaluation benchmark for the visual question answering (VQA) task, extending the test-dev set of the English GQA dataset [11] to 7 ...
This is the implementation of the approaches described in the paper: "Improving the Cross-Lingual Generalisation in Visual Question Answering" ...
Sep 10, 2024 · In this work, we address this gap and provide xGQA, a new multilingual evaluation benchmark for the visual question answering task. We extend ...