OpenAlex Citation Counts

OpenAlex Citations Logo

OpenAlex is a bibliographic catalogue of scientific papers, authors and institutions accessible in open access mode, named after the Library of Alexandria. It's citation coverage is excellent and I hope you will find utility in this listing of citing articles!

If you click the article title, you'll navigate to the article, as listed in CrossRef. If you click the Open Access links, you'll navigate to the "best Open Access location". Clicking the citation count will open this listing for that article. Lastly at the bottom of the page, you'll find basic pagination options.

Requested Article:

D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Justin Fu, Aviral Kumar, Ofir Nachum, et al.
arXiv (Cornell University) (2020)
Open Access | Times Cited: 319

Showing 26-50 of 319 citing articles:

Hierarchical Planning Through Goal-Conditioned Offline Reinforcement Learning
Jinning Li, Chen Tang, Masayoshi Tomizuka, et al.
IEEE Robotics and Automation Letters (2022) Vol. 7, Iss. 4, pp. 10216-10223
Open Access | Times Cited: 25

Learning to Control Autonomous Fleets from Observation via Offline Reinforcement Learning
Carolin Schmidt, Daniele Gammelli, Francisco C. Pereira, et al.
2022 European Control Conference (ECC) (2024), pp. 1399-1406
Open Access | Times Cited: 5

RL Unplugged: A Suite of Benchmarks for Offline Reinforcement Learning
Çağlar Gülçehre, Ziyu Wang, Alexander Novikov, et al.
arXiv (Cornell University) (2020)
Open Access | Times Cited: 38

Uncertainty-Based Offline Reinforcement Learning with Diversified Q-Ensemble
Gaon An, Seungyong Moon, Jang-Hyun Kim, et al.
arXiv (Cornell University) (2021)
Open Access | Times Cited: 30

Actor–Critic Reinforcement Learning and Application in Developing Computer-Vision-Based Interface Tracking
Oguzhan Dogru, Kirubakaran Velswamy, Biao Huang
Engineering (2021) Vol. 7, Iss. 9, pp. 1248-1261
Open Access | Times Cited: 27

Offline prompt reinforcement learning method based on feature extraction
Ting Yao, Xiliang Chen, Yi Yao, et al.
PeerJ Computer Science (2025) Vol. 11, pp. e2490-e2490
Open Access

Mild evaluation policy via dataset constraint for offline reinforcement learning
Xue Li, Xinghong Ling
Expert Systems with Applications (2025), pp. 126842-126842
Closed Access

ELAPSE: Expand Latent Action Projection Space for policy optimization in Offline Reinforcement Learning
Xinchen Han, Hossam Afifi, Michel Marot
Neurocomputing (2025), pp. 129665-129665
Closed Access

Conservative reward enhancement through the nearest neighbor integration in model-based Offline Policy Optimization
Xue Li, Bangjun Wang, Xinghong Ling
Expert Systems with Applications (2025), pp. 126888-126888
Closed Access

Offline Robustness of Distributional Actor-Critic Ensemble Reinforcement Learning
Zihang Ma, Daphne Teck Ching Lai, Zhu Jian-xiang, et al.
Advances in Pure Mathematics (2025) Vol. 15, Iss. 04, pp. 269-290
Open Access

Weighted Policy Constraints for Offline Reinforcement Learning
Zhiyong Peng, Changlin Han, Yadong Liu, et al.
Proceedings of the AAAI Conference on Artificial Intelligence (2023) Vol. 37, Iss. 8, pp. 9435-9443
Open Access | Times Cited: 9

Learning to View: Decision Transformers for Active Object Detection
Wenhao Ding, Nathalie Majcherczyk, Mohit Deshpande, et al.
(2023)
Open Access | Times Cited: 9

Federated Offline Reinforcement Learning With Multimodal Data
Jiabao Wen, Huiao Dai, Jingyi He, et al.
IEEE Transactions on Consumer Electronics (2023) Vol. 70, Iss. 1, pp. 4266-4276
Closed Access | Times Cited: 9

Efficient Offline Reinforcement Learning With Relaxed Conservatism
Longyang Huang, Botao Dong, Weidong Zhang
IEEE Transactions on Pattern Analysis and Machine Intelligence (2024) Vol. 46, Iss. 8, pp. 5260-5272
Closed Access | Times Cited: 3

Benchmarks for Deep Off-Policy Evaluation
Justin Fu, Mohammad Norouzi, Ofir Nachum, et al.
arXiv (Cornell University) (2021)
Open Access | Times Cited: 22

COMBO: Conservative Offline Model-Based Policy Optimization
Tianhe Yu, Aviral Kumar, Rafael Rafailov, et al.
arXiv (Cornell University) (2021)
Closed Access | Times Cited: 21

ManiSkill: Generalizable Manipulation Skill Benchmark with Large-Scale Demonstrations
Tongzhou Mu, Zhan Ling, Fanbo Xiang, et al.
arXiv (Cornell University) (2021)
Open Access | Times Cited: 21

Dataset Distillation by Matching Training Trajectories
George Cazenavette, Tongzhou Wang, Antonio Torralba, et al.
2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (2022), pp. 4749-4758
Open Access | Times Cited: 16

Value Penalized Q-Learning for Recommender Systems
Chengqian Gao, Ke Xu, Kuangqi Zhou, et al.
Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval (2022), pp. 2008-2012
Open Access | Times Cited: 14

Offline Decentralized Multi-Agent Reinforcement Learning
Jiechuan Jiang, Zongqing Lu
Frontiers in artificial intelligence and applications (2023)
Open Access | Times Cited: 8

Doubly constrained offline reinforcement learning for learning path recommendation
Yun Yue, Huan Dai, Rui An, et al.
Knowledge-Based Systems (2023) Vol. 284, pp. 111242-111242
Closed Access | Times Cited: 8

Bi-DexHands: Towards Human-Level Bimanual Dexterous Manipulation
Yuanpei Chen, Yiran Geng, Fangwei Zhong, et al.
IEEE Transactions on Pattern Analysis and Machine Intelligence (2023) Vol. 46, Iss. 5, pp. 2804-2818
Closed Access | Times Cited: 8

Hierarchical Decision Transformer
André Correia, Luı́s A. Alexandre
2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (2023), pp. 1661-1666
Open Access | Times Cited: 8

BAIL: Best-Action Imitation Learning for Batch Deep Reinforcement Learning
Xinyue Chen, Zijian Zhou, Zheng Wang, et al.
arXiv (Cornell University) (2019)
Closed Access | Times Cited: 25

Is Pessimism Provably Efficient for Offline RL
Ying Jin, Zhuoran Yang, Zhaoran Wang
arXiv (Cornell University) (2020)
Closed Access | Times Cited: 23

Scroll to top