OpenAlex Citation Counts

OpenAlex Citations Logo

OpenAlex is a bibliographic catalogue of scientific papers, authors and institutions accessible in open access mode, named after the Library of Alexandria. It's citation coverage is excellent and I hope you will find utility in this listing of citing articles!

If you click the article title, you'll navigate to the article, as listed in CrossRef. If you click the Open Access links, you'll navigate to the "best Open Access location". Clicking the citation count will open this listing for that article. Lastly at the bottom of the page, you'll find basic pagination options.

Requested Article:

SpeechFormer++: A Hierarchical Efficient Framework for Paralinguistic Speech Processing
Weidong Chen, Xiaofen Xing, Xiangmin Xu, et al.
IEEE/ACM Transactions on Audio Speech and Language Processing (2023) Vol. 31, pp. 775-788
Open Access | Times Cited: 34

Showing 1-25 of 34 citing articles:

Vesper: A Compact and Effective Pretrained Model for Speech Emotion Recognition
Weidong Chen, Xiaofen Xing, Peihao Chen, et al.
IEEE Transactions on Affective Computing (2024) Vol. 15, Iss. 3, pp. 1711-1724
Open Access | Times Cited: 15

Speech Swin-Transformer: Exploring a Hierarchical Transformer with Shifted Windows for Speech Emotion Recognition
Yong Wang, Cheng Lu, Hailun Lian, et al.
ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2024), pp. 11646-11650
Open Access | Times Cited: 12

Hidformer: Hierarchical dual-tower transformer using multi-scale mergence for long-term time series forecasting
Zhaoran Liu, Yizhi Cao, Hu Xu, et al.
Expert Systems with Applications (2023) Vol. 239, pp. 122412-122412
Closed Access | Times Cited: 19

Speech Emotion Recognition Using Dual-Stream Representation and Cross-Attention Fusion
Shaode Yu, Jiajian Meng, Wenqing Fan, et al.
Electronics (2024) Vol. 13, Iss. 11, pp. 2191-2191
Open Access | Times Cited: 4

A Joint Network Based on Interactive Attention for Speech Emotion Recognition
Ying Hu, Shijing Hou, Huamin Yang, et al.
2022 IEEE International Conference on Multimedia and Expo (ICME) (2023), pp. 1715-1720
Closed Access | Times Cited: 10

Automatic recognition of depression based on audio and video: A review
Mengmeng Han, LI Xing-yun, Xinyu Yi, et al.
World Journal of Psychiatry (2024) Vol. 14, Iss. 2, pp. 225-233
Open Access | Times Cited: 3

TS-MEFM: A New Multimodal Speech Emotion Recognition Network Based on Speech and Text Fusion
Wei Wei, Bingkun Zhang, Yibing Wang
Lecture notes in computer science (2025), pp. 454-467
Closed Access

Can Large Language Models Aid in Annotating Speech Emotional Data? Uncovering New Frontiers [Research Frontier]
Siddique Latif, Muhammad Usama, Muhammad Ibrahim Malik, et al.
IEEE Computational Intelligence Magazine (2025) Vol. 20, Iss. 1, pp. 66-77
Closed Access

Machine Learning and Neural Networks for IT-Diagnostics of Neurological Diseases
U. А. Vishniakou, Yudong Xia, C. Yu
Doklady BGUIR (2025) Vol. 23, Iss. 1, pp. 68-73
Open Access

Acoustic Feature Excitation-and-Aggregation Network Based on Multi-Task Learning for Speech Emotion Recognition
Xin Qi, Qing Song, Guowei Chen, et al.
Electronics (2025) Vol. 14, Iss. 5, pp. 844-844
Open Access

Multimodal speech emotion recognition via dynamic multilevel contrastive loss under local enhancement network
Weiquan Fan, Xiangmin Xu, Fang Liu, et al.
Expert Systems with Applications (2025), pp. 127669-127669
Closed Access

FREE-Net: A dual-modality emotion recognition network for fusing raw and enhanced data
Wei Wei, Bingkun Zhang, Yibing Wang
Neurocomputing (2025), pp. 130361-130361
Closed Access

Modality fusion using auxiliary tasks for dementia detection
Huaizhi Shao, Yilin Pan, Yue Wang, et al.
Computer Speech & Language (2025), pp. 101814-101814
Closed Access

HAFFormer: A Hierarchical Attention-Free Framework for Alzheimer’s Disease Detection From Spontaneous Speech
Zhongren Dong, Zixing Zhang, Weixiang Xu, et al.
ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2024), pp. 11246-11250
Open Access | Times Cited: 2

Speechformer-CTC: Sequential modeling of depression detection with speech temporal classification
Jinhan Wang, Vijay Ravi, Jonathan Flint, et al.
Speech Communication (2024) Vol. 163, pp. 103106-103106
Open Access | Times Cited: 2

MelTrans: Mel-Spectrogram Relationship-Learning for Speech Emotion Recognition via Transformers
Hui Li, Jiawen Li, Hai Liu, et al.
Sensors (2024) Vol. 24, Iss. 17, pp. 5506-5506
Open Access | Times Cited: 2

ESERNet: Learning Spectrogram Structure Relationship for Effective Speech Emotion Recognition with Swin Transformer in Classroom Discourse Analysis
Tingting Liu, Minghong Wang, Bing Yang, et al.
Neurocomputing (2024) Vol. 612, pp. 128711-128711
Closed Access | Times Cited: 2

Avoiding dominance of speaker features in speech-based depression detection
Lishi Zuo, Man‐Wai Mak
Pattern Recognition Letters (2023) Vol. 173, pp. 50-56
Closed Access | Times Cited: 6

Speech based detection of Alzheimer’s disease: a survey of AI techniques, datasets and challenges
Kewen Ding, Madhu Chetty, Azadeh Noori Hoshyar, et al.
Artificial Intelligence Review (2024) Vol. 57, Iss. 12
Open Access | Times Cited: 1

Exploring Sequential Feature Selection in Deep Bi-LSTM Models for Speech Emotion Recognition
Fatma Harby, Mansor Alohali, Adel Thaljaoui, et al.
Computers, materials & continua/Computers, materials & continua (Print) (2024) Vol. 78, Iss. 2, pp. 2689-2719
Open Access | Times Cited: 1

DESTformer: A Transformer Based on Explicit Seasonal–Trend Decomposition for Long-Term Series Forecasting
Yajun Wang, Jianping Zhu, Renke Kang
Applied Sciences (2023) Vol. 13, Iss. 18, pp. 10505-10505
Open Access | Times Cited: 3

Disentangling Prosody Representations With Unsupervised Speech Reconstruction
Leyuan Qu, Taihao Li, Cornelius Weber, et al.
IEEE/ACM Transactions on Audio Speech and Language Processing (2023) Vol. 32, pp. 39-54
Open Access | Times Cited: 2

Comprehensive Speech Emotion Recognition System Employing Multi-Layer Perceptron (MLP) Classifier and libRosa Feature Extraction
Sushmitha Saro R, Jaya Suriya B, R Rajakumari
(2023), pp. 1204-1211
Closed Access | Times Cited: 2

Automatic classification of emotions in speech: methods and data
Vladislav Igorevich Lemaev, Natal'ya Valentinovna Lukashevich
Litera (2024), Iss. 4, pp. 159-173
Open Access

Hierarchical convolutional neural networks with post-attention for speech emotion recognition
Yonghong Fan, Heming Huang, Henry Han
Neurocomputing (2024) Vol. 615, pp. 128879-128879
Closed Access

Page 1 - Next Page

Scroll to top