OpenAlex Citation Counts

OpenAlex Citations Logo

OpenAlex is a bibliographic catalogue of scientific papers, authors and institutions accessible in open access mode, named after the Library of Alexandria. It's citation coverage is excellent and I hope you will find utility in this listing of citing articles!

If you click the article title, you'll navigate to the article, as listed in CrossRef. If you click the Open Access links, you'll navigate to the "best Open Access location". Clicking the citation count will open this listing for that article. Lastly at the bottom of the page, you'll find basic pagination options.

Requested Article:

TyDi QA: A Benchmark for Information-Seeking Question Answering in Typologically Diverse Languages
Jonathan H. Clark, Eunsol Choi, Michael Collins, et al.
Transactions of the Association for Computational Linguistics (2020) Vol. 8, pp. 454-470
Open Access | Times Cited: 295

Showing 1-25 of 295 citing articles:

Large language models encode clinical knowledge
Karan Singhal, Shekoofeh Azizi, Tao Tu, et al.
Nature (2023) Vol. 620, Iss. 7972, pp. 172-180
Open Access | Times Cited: 1432

mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
Linting Xue, Noah Constant, Adam P. Roberts, et al.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (2021)
Open Access | Times Cited: 1262

XTREME: A Massively Multilingual Multi-task Benchmark for Evaluating Cross-lingual Generalization
Junjie Hu, Sebastian Ruder, Aditya Siddhant, et al.
arXiv (Cornell University) (2020)
Open Access | Times Cited: 324

ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models
Linting Xue, Aditya Barua, Noah Constant, et al.
Transactions of the Association for Computational Linguistics (2022) Vol. 10, pp. 291-306
Open Access | Times Cited: 156

How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models
Phillip Rust, Jonas Pfeiffer, Ivan Vulić, et al.
(2021)
Open Access | Times Cited: 133

Going Full-TILT Boogie on Document Understanding with Text-Image-Layout Transformer
Rafał Powalski, Łukasz Borchmann, Dawid Jurkiewicz, et al.
Lecture notes in computer science (2021), pp. 732-747
Closed Access | Times Cited: 115

QA Dataset Explosion: A Taxonomy of NLP Resources for Question Answering and Reading Comprehension
Anna Rogers, Matt Gardner, Isabelle Augenstein
ACM Computing Surveys (2022) Vol. 55, Iss. 10, pp. 1-45
Open Access | Times Cited: 113

Canine: Pre-training an Efficient Tokenization-Free Encoder for Language Representation
Jonathan H. Clark, Dan Garrette, Iulia Turc, et al.
Transactions of the Association for Computational Linguistics (2022) Vol. 10, pp. 73-91
Open Access | Times Cited: 102

BanglaBERT: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in Bangla
Abhik Bhattacharjee, Tahmid Hasan, Wasi Uddin Ahmad, et al.
Findings of the Association for Computational Linguistics: NAACL 2022 (2022)
Open Access | Times Cited: 98

A review on big data based on deep neural network approaches
M Rithani, R Prasanna Kumar, Srinath Doss
Artificial Intelligence Review (2023) Vol. 56, Iss. 12, pp. 14765-14801
Closed Access | Times Cited: 51

Datasets for Large Language Models: A Comprehensive Survey
Yang Liu, Jiahuan Cao, Chongyu Liu, et al.
Research Square (Research Square) (2024)
Open Access | Times Cited: 22

A Primer on Pretrained Multilingual Language Models
Sumanth Doddapaneni, G. Ramesh, Mitesh M. Khapra, et al.
ACM Computing Surveys (2025)
Open Access | Times Cited: 3

Pretrained Transformers for Text Ranking: BERT and Beyond
Jimmy Lin, Rodrigo Nogueira, Andrew Yates
Synthesis lectures on human language technologies (2021) Vol. 14, Iss. 4, pp. 1-325
Closed Access | Times Cited: 104

Did Aristotle Use a Laptop?A Question Answering Benchmark with Implicit Reasoning Strategies
Mor Geva, Daniel Khashabi, Elad Segal, et al.
Transactions of the Association for Computational Linguistics (2021) Vol. 9, pp. 346-361
Open Access | Times Cited: 102

XTREME-R: Towards More Challenging and Nuanced Multilingual Evaluation
Sebastian Ruder, Noah Constant, Jan A. Botha, et al.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (2021), pp. 10215-10245
Open Access | Times Cited: 98

mT5: A massively multilingual pre-trained text-to-text transformer
Linting Xue, Noah Constant, Adam P. Roberts, et al.
arXiv (Cornell University) (2020)
Open Access | Times Cited: 94

MKQA: A Linguistically Diverse Benchmark for Multilingual Open Domain Question Answering
Shayne Longpre, Yi Lu, Joachim Daiber
Transactions of the Association for Computational Linguistics (2021) Vol. 9, pp. 1389-1406
Open Access | Times Cited: 89

X-FACTR: Multilingual Factual Knowledge Retrieval from Pretrained Language Models
Zhengbao Jiang, Antonios Anastasopoulos, Jun Araki, et al.
(2020)
Open Access | Times Cited: 83

XOR QA: Cross-lingual Open-Retrieval Question Answering
Akari Asai, Jungo Kasai, Jonathan H. Clark, et al.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (2021)
Open Access | Times Cited: 82

Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models
Nora Kassner, Philipp Dufter, Hinrich Schütze
(2021), pp. 3250-3258
Open Access | Times Cited: 70

Charformer: Fast Character Transformers via Gradient-based Subword Tokenization
Yi Tay, Vinh Q. Tran, Sebastian Ruder, et al.
arXiv (Cornell University) (2021)
Open Access | Times Cited: 69

Few-Shot Question Answering by Pretraining Span Selection
Ori Ram, Yuval Kirstain, Jonathan Berant, et al.
(2021), pp. 3066-3079
Open Access | Times Cited: 68

Page 1 - Next Page

Scroll to top