OpenAlex Citation Counts

OpenAlex Citations Logo

OpenAlex is a bibliographic catalogue of scientific papers, authors and institutions accessible in open access mode, named after the Library of Alexandria. It's citation coverage is excellent and I hope you will find utility in this listing of citing articles!

If you click the article title, you'll navigate to the article, as listed in CrossRef. If you click the Open Access links, you'll navigate to the "best Open Access location". Clicking the citation count will open this listing for that article. Lastly at the bottom of the page, you'll find basic pagination options.

Requested Article:

mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer
Linting Xue, Noah Constant, Adam P. Roberts, et al.
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (2021)
Open Access | Times Cited: 1262

Showing 1-25 of 1262 citing articles:

On the Opportunities and Risks of Foundation Models
Rishi Bommasani, Drew A. Hudson, Ehsan Adeli, et al.
arXiv (Cornell University) (2021)
Open Access | Times Cited: 1565

DeBERTaV3: Improving DeBERTa using ELECTRA-Style Pre-Training with Gradient-Disentangled Embedding Sharing
Pengcheng He, Jianfeng Gao, Weizhu Chen
arXiv (Cornell University) (2021)
Open Access | Times Cited: 338

Unifying Large Language Models and Knowledge Graphs: A Roadmap
Shirui Pan, Linhao Luo, Yufei Wang, et al.
IEEE Transactions on Knowledge and Data Engineering (2024) Vol. 36, Iss. 7, pp. 3580-3599
Closed Access | Times Cited: 321

What Makes Good In-Context Examples for GPT-3?
Jiachang Liu, Dinghan Shen, Yizhe Zhang, et al.
(2022)
Open Access | Times Cited: 307

XLS-R: Self-supervised Cross-lingual Speech Representation Learning at Scale
Arun Babu, Changhan Wang, Andros Tjandra, et al.
Interspeech 2022 (2022)
Open Access | Times Cited: 283

GPT-NeoX-20B: An Open-Source Autoregressive Language Model
Sidney Black, Stella Biderman, Eric Hallahan, et al.
(2022)
Open Access | Times Cited: 281

Unified Structure Generation for Universal Information Extraction
Yaojie Lu, Q. Liu, Dai Dai, et al.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (2022)
Open Access | Times Cited: 261

ARBERT & MARBERT: Deep Bidirectional Transformers for Arabic
Muhammad Abdul-Mageed, AbdelRahim Elmadany, El Moatez Billah Nagoudi
(2021)
Open Access | Times Cited: 238

Crosslingual Generalization through Multitask Finetuning
Niklas Muennighoff, Thomas J. Wang, Lintang Sutawika, et al.
(2023)
Open Access | Times Cited: 236

LiT: Zero-Shot Transfer with Locked-image text Tuning
Xiaohua Zhai, Xiao Wang, Basil Mustafa, et al.
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2022), pp. 18102-18112
Open Access | Times Cited: 227

Pre-Trained Language Models and Their Applications
Haifeng Wang, Jiwei Li, Hua Wu, et al.
Engineering (2022) Vol. 25, pp. 51-65
Open Access | Times Cited: 180

PPT: Pre-trained Prompt Tuning for Few-shot Learning
Yuxian Gu, Xu Han, Zhiyuan Liu, et al.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (2022)
Open Access | Times Cited: 176

Super-NaturalInstructions: Generalization via Declarative Instructions on 1600+ NLP Tasks
Yizhong Wang, Swaroop Mishra, Pegah Alipoormolabashi, et al.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (2022)
Open Access | Times Cited: 163

Deduplicating Training Data Makes Language Models Better
Katherine Lee, Daphne Ippolito, Andrew Nystrom, et al.
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (2022)
Open Access | Times Cited: 161

Documenting Large Webtext Corpora: A Case Study on the Colossal Clean Crawled Corpus
Jesse Dodge, Maarten Sap, Ana Marasović, et al.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (2021)
Open Access | Times Cited: 157

ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models
Linting Xue, Aditya Barua, Noah Constant, et al.
Transactions of the Association for Computational Linguistics (2022) Vol. 10, pp. 291-306
Open Access | Times Cited: 155

The GEM Benchmark: Natural Language Generation, its Evaluation and Metrics
Sebastian Gehrmann, Tosin Adewumi, Karmanya Aggarwal, et al.
(2021)
Open Access | Times Cited: 151

XL-Sum: Large-Scale Multilingual Abstractive Summarization for 44 Languages
Tahmid Hasan, Abhik Bhattacharjee, Md. Saiful Islam, et al.
(2021)
Open Access | Times Cited: 150

Quality at a Glance: An Audit of Web-Crawled Multilingual Datasets
Julia Kreutzer, Isaac Caswell, Lisa Wang, et al.
Transactions of the Association for Computational Linguistics (2022) Vol. 10, pp. 50-72
Open Access | Times Cited: 143

A Review on Large Language Models: Architectures, Applications, Taxonomies, Open Issues and Challenges
Mohaimenul Azam Khan Raiaan, Md. Saddam Hossain Mukta, Kaniz Fatema, et al.
IEEE Access (2024) Vol. 12, pp. 26839-26874
Open Access | Times Cited: 143

How Good is Your Tokenizer? On the Monolingual Performance of Multilingual Language Models
Phillip Rust, Jonas Pfeiffer, Ivan Vulić, et al.
(2021)
Open Access | Times Cited: 133

Biases in Large Language Models: Origins, Inventory, and Discussion
Roberto Navigli, Simone Conia, Björn Roß
Journal of Data and Information Quality (2023) Vol. 15, Iss. 2, pp. 1-21
Open Access | Times Cited: 133

A survey of GPT-3 family large language models including ChatGPT and GPT-4
Katikapalli Subramanyam Kalyan
Natural Language Processing Journal (2023) Vol. 6, pp. 100048-100048
Open Access | Times Cited: 129

Large Dual Encoders Are Generalizable Retrievers
Jianmo Ni, Chen Qu, Jing Lü, et al.
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (2022)
Open Access | Times Cited: 104

Page 1 - Next Page

Scroll to top