GrailQA

The Strongly Generalizable Question Answering Dataset

What is GrailQA?

Strongly Generalizable Question Answering (GrailQA) is a new large-scale, high-quality dataset for question answering on knowledge bases (KBQA) on Freebase with 64,331 questions annotated with both answers and corresponding logical forms in different syntax (i.e., SPARQL, S-expression, etc.). It can be used to test three levels of generalization in KBQA: i.i.d., compositional, and zero-shot.

Explore GrailQAGrailQA paper (Gu et al. '20)Code

Why GrailQA?

GrailQA is by far the largest crowdsourced KBQA dataset with questions of high diversity (i.e., questions in GrailQA can have up to 4 relations and optionally have a function from counting, superlatives and comparatives). It also has the highest coverage over Freebase; it widely covers 3,720 relations and 86 domains from Freebase. Last but not least, our meticulous data split allows GrailQA to test not only i.i.d. generalization, but also compositional generalization and zero-shot generalization, which are critical for practical KBQA systems.

News

  • 01/24/2021We provide instructions on Freebase setup.
  • 01/22/2021We have updated our baseline performance based on a new entity linker with a recall of around 0.77 on the dev set, compared with the previous recall of around 0.46. More details about entity linking can be found in the updated paper.
  • 01/18/2021 We will have some major update on our baseline model results with a new entity linker in this week. The numbers will be higher. Please stay tuned!
  • 11/30/2020 We fix some minor error in the sparql_queries provided in our dataset.

Getting Started

We've built a few resources to help you get started with the dataset.

Download a copy of the dataset (distributed under the CC BY-SA 4.0 license):


To work with our dataset, we recommend you setting up a Virtuoso server for Freebase (feel free to choose your own way to index Freebase). Please find both a clean version of Freebase dump and instructions on setting up the server via:


To evaluate your models, we have also made available the evaluation script we will use for official evaluation, along with a sample prediction file that the script will take as input. Evaluating semantic-level exact match also depends on several preprocessed ontology files of Freebase. You can find all of them here. To run the evaluation, use python evaluate.py <path_to_dev> <path_to_predictions> --fb_roles <path_to_fb_roles> --fb_types <path_to_fb_types> --reverse_properties <path_to_reverse_properties>.

Once you have a built a model that works to your expectations on the dev set, you submit it to get official scores on the dev and a hidden test set. To preserve the integrity of test results, we do not release the labels of test set to the public. Here's a tutorial walking you through official evaluation of your model:

Submission Tutorial

Have Questions?

Send an email to gu.826@osu.edu, or create an issue in github.

Acknowledgement

We thank Pranav Rajpurkar and Robin Jia for giving us the permission to build this website based on SQuAD.

Leaderboard: Overall

Here are the overall Exact Match (EM) and F1 scores evaluated on GrailQA test set. To get the EM score on GrailQA, please submit your results with logical forms in S-expression. Note that, submissions are ranked only based on F1, so feel free to choose your own meaning representation as EM won't affect your ranking.

RankModelEMF1

1

Jan 11, 2023
Rank-KBQA (ensemble)

HIK RESEARCH INST NLP

73.41881.869

2

Dec 12, 2022
Pangu (T5-3B)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
75.37681.701

3

Dec 27, 2022
GAIN (T5-3B)

Microsoft

https://arxiv.org/abs/2309.08345
76.30681.522

4

Nov 25, 2022
GAIN (T5-base)

Microsoft

https://arxiv.org/abs/2309.08345
75.07480.632

5

Dec 02, 2022
Pangu (BERT-base)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
73.73679.930

6

Mar 17, 2024
RetinaQA (single model)

Indian Institute of Technology, Delhi

74.12979.519

7

Nov 09, 2022
FC-KBQA (single model)

Anonymous

73.19978.765

8

Sep 06, 2022
DecAF (single model)

AWS AI Labs

https://arxiv.org/abs/2210.00063
68.36278.758

9

Feb 11, 2023
MGC-KBQA (single model)

Anonymous

72.84478.522

10

May 31, 2022
TIARA (single model)

Microsoft Research

https://arxiv.org/abs/2210.12925
73.04178.520

11

Jul 12, 2022
DeCC (single model)

Anonymous

72.11977.648

12

Dec 28, 2022
Rank-KBQA (single model)

HIK RESEARCH INST NLP

64.58377.125

13

Aug 31, 2022
UniParser (base model)

Salesforce

https://arxiv.org/pdf/2211.05165.pdf
69.45174.649

14

Aug 20, 2021
RnG-KBQA (single model)

Salesforce Research

https://aclanthology.org/2022.acl-long.417/
68.77874.422

15

Apr 19, 2022
ArcaneQA V2 (single model)

The Ohio State University

https://aclanthology.org/2022.coling-1.148/
63.77473.713

16

Aug 12, 2021
S2QL (single model)

UCAS & Meituan Inc.

https://dl.acm.org/doi/abs/10.1007/978-3-031-05981-0_18
57.45666.186

17

Apr 05, 2021
ReTraCk (single model)

Microsoft Research

https://aclanthology.org/2021.acl-demo.39/
58.13665.285

18

Feb 04, 2021
ArcaneQA V1 (single model)

The Ohio State University

57.87264.924

19

Jan 22, 2021
BERT+Ranking (single model)

The Ohio State University

50.57857.988

20

Jan 22, 2021
GloVe+Ranking (single model)

The Ohio State University

39.52145.136

21

Jan 22, 2021
BERT+Transduction (single model)

The Ohio State University

33.25536.803

22

Jan 22, 2021
GloVe+Transduction (single model)

The Ohio State University

17.58718.432

Leaderboard: Compositional Generalization

Here are the Exact Match (EM) and F1 scores evaluated on the subset of GrailQA test set that tests compositional generalization.

RankModelEMF1

1

Jan 11, 2023
Rank-KBQA (ensemble)

HIK RESEARCH INST NLP

74.80682.293

2

Sep 06, 2022
DecAF (single model)

AWS AI Labs

https://arxiv.org/abs/2210.00063
73.38581.807

3

Dec 12, 2022
Pangu (T5-3B)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
74.58081.518

4

Dec 02, 2022
Pangu (BERT-base)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
74.93581.216

5

Dec 27, 2022
GAIN (T5-3B)

Microsoft

https://arxiv.org/abs/2309.08345
73.70880.004

6

Nov 25, 2022
GAIN (T5-base)

Microsoft

https://arxiv.org/abs/2309.08345
72.99779.561

7

Dec 28, 2022
Rank-KBQA (single model)

HIK RESEARCH INST NLP

70.25279.392

8

Mar 17, 2024
RetinaQA (single model)

Indian Institute of Technology, Delhi

71.99678.903

9

Feb 11, 2023
MGC-KBQA (single model)

Anonymous

71.38278.790

10

Nov 09, 2022
FC-KBQA (single model)

Anonymous

69.99476.677

11

May 31, 2022
TIARA (single model)

Microsoft Research

https://arxiv.org/abs/2210.12925
69.18676.546

12

Jul 12, 2022
DeCC (single model)

Anonymous

68.54075.862

13

Apr 19, 2022
ArcaneQA V2 (single model)

The Ohio State University

https://aclanthology.org/2022.coling-1.148/
65.79575.302

14

Aug 20, 2021
RnG-KBQA (single model)

Salesforce Research

https://aclanthology.org/2022.acl-long.417/
63.79271.156

15

Aug 31, 2022
UniParser (base model)

Salesforce

https://arxiv.org/pdf/2211.05165.pdf
65.14971.102

16

Apr 05, 2021
ReTraCk (single model)

Microsoft Research

https://aclanthology.org/2021.acl-demo.39/
61.49970.911

17

Aug 12, 2021
S2QL (single model)

UCAS & Meituan Inc.

https://dl.acm.org/doi/abs/10.1007/978-3-031-05981-0_18
54.71664.679

18

Feb 04, 2021
ArcaneQA V1 (single model)

The Ohio State University

56.39563.533

19

Jan 22, 2021
BERT+Ranking (single model)

The Ohio State University

45.51053.890

20

Jan 22, 2021
GloVe+Ranking (single model)

The Ohio State University

39.95547.753

21

Jan 22, 2021
BERT+Transduction (single model)

The Ohio State University

31.04035.985

22

Jan 22, 2021
GloVe+Transduction (single model)

The Ohio State University

16.44118.507

Leaderboard: Zero-shot Generalization

Here are the Exact Match (EM) and F1 scores evaluated on the subset of GrailQA test set that tests zero-shot generalization.

RankModelEMF1

1

Dec 12, 2022
Pangu (T5-3B)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
71.57578.536

2

Dec 27, 2022
GAIN (T5-3B)

Microsoft

https://arxiv.org/abs/2309.08345
71.84877.751

3

Nov 25, 2022
GAIN (T5-base)

Microsoft

https://arxiv.org/abs/2309.08345
69.91876.357

4

Dec 02, 2022
Pangu (BERT-base)

The Ohio State University

https://aclanthology.org/2023.acl-long.270/
69.12576.059

5

Jan 11, 2023
Rank-KBQA (ensemble)

HIK RESEARCH INST NLP

63.65175.514

6

Mar 17, 2024
RetinaQA (single model)

Indian Institute of Technology, Delhi

68.85274.772

7

Nov 09, 2022
FC-KBQA (single model)

Anonymous

67.58473.986

8

May 31, 2022
TIARA (single model)

Microsoft Research

https://arxiv.org/abs/2210.12925
67.98773.864

9

Feb 11, 2023
MGC-KBQA (single model)

Anonymous

67.25373.355

10

Jul 12, 2022
DeCC (single model)

Anonymous

66.54772.545

11

Sep 06, 2022
DecAF (single model)

AWS AI Labs

https://arxiv.org/abs/2210.00063
58.55172.270

12

Aug 31, 2022
UniParser (base model)

Salesforce

https://arxiv.org/pdf/2211.05165.pdf
63.96869.847

13

Aug 20, 2021
RnG-KBQA (single model)

Salesforce Research

https://aclanthology.org/2022.acl-long.417/
62.98869.182

14

Dec 28, 2022
Rank-KBQA (single model)

HIK RESEARCH INST NLP

50.16668.489

15

Apr 19, 2022
ArcaneQA V2 (single model)

The Ohio State University

https://aclanthology.org/2022.coling-1.148/
52.86066.029

16

Aug 12, 2021
S2QL (single model)

UCAS & Meituan Inc.

https://dl.acm.org/doi/abs/10.1007/978-3-031-05981-0_18
55.12263.598

17

Feb 04, 2021
ArcaneQA V1 (single model)

The Ohio State University

49.96458.844

18

Jan 22, 2021
BERT+Ranking (single model)

The Ohio State University

48.56655.660

19

Apr 05, 2021
ReTraCk (single model)

Microsoft Research

https://aclanthology.org/2021.acl-demo.39/
44.56152.539

20

Jan 22, 2021
GloVe+Ranking (single model)

The Ohio State University

28.88633.792

21

Jan 22, 2021
BERT+Transduction (single model)

The Ohio State University

25.70229.300

22

Jan 22, 2021
GloVe+Transduction (single model)

The Ohio State University

2.9683.123