[Task Submission] Quantifier Understanding (quantifier_understanding)#20
[Task Submission] Quantifier Understanding (quantifier_understanding)#20lerow wants to merge 7 commits intoGenBench:mainfrom
quantifier_understanding)#20Conversation
|
Thanks for submitting a task to GenBench. Please be aware that you're submitting the data files of the tasks in your PR. For the final submission, you will need to host the dataset files somewhere else (preferably as a Huggingface dataset). Additionally, in your current submission, the |
|
Hello! We are getting quite close to the deadline (September 1, 11:59PM anywhere on earth), which is why I wanted to remind you of the fact that your PR still needs some attention: please double-check the automated checks that failed, and ensure that the dataset files are hosted somewhere else; see Amir's message above. Please don't forget to submit your accompanying paper to Openreview via https://openreview.net/group?id=GenBench.org/2023/Workshop by September 1. Good luck finalising your PR and paper, feel free to tag us if you have questions. |
|
Hi, |
Quantifier Understanding
The task evaluates generalization in the understanding of quantifiers. It aims to measure
how well can language models capture the semantics of logical quantifiers in natural language.
Authors
lryw@uw.edushanest@uw.eduImplementation
The task re-implements the evaluation function to compute accuracy scores.
Usage
Given predictions and gold labels,
evaluate_predictions()outputs the accuracy score.Checklist