Ranipa, Kalpeshkumar (2025) Advanced Deep Learning Architectures for Automated Heart Sound Classification. PhD thesis, Concordia University.
Text (application/pdf)
7MBRanipa_PhD_F2025.pdf - Accepted Version Restricted to Repository staff only until 1 September 2027. Available under License Spectrum Terms of Access. |
Abstract
Automatic heart sound classification is a longstanding and challenging problem in the research field of biomedical engineering due to the fact that heart sound is very complex and highly nonstationary.
The most machine learning-based heart sound classification methods achieve limited accuracy and are primarily developed using a segmentation-based approach. Since they depend on single-domain feature information and also tend to focus equally on each part of the signal, rather than employing a selective attention mechanism.
Additionally, the existing studies rely on single-stream architectures, overlooking the advantages of multi-resolution features. Apart from this, these methods rely solely on early, late, or intermediate fusion, which often fail to fully integrate the diverse and multiscale features required for robust classification.
The proposed framework addresses these problems through three core components of deep learning architecture design: the selection of diverse feature representations combined through an effective feature fusion strategy, a learning mechanism that leverages multi-resolution feature fusion, and the utilization of multi-scale representations.
In this thesis, three novel deep learning architectures and methods are developed that do not require a segmentation approach to deal with the existing limitations in the field of heart sound classification.
First, we propose a novel multimodal attention convolutional neural network (MACNN) that incorporates a feature-level fusion strategy. The architecture consists of three parallel branches, each employing low-complexity convolutional neural networks combined with attention mechanisms to process different feature domains. This design enhances feature diversity and enables more effective multi-feature fusion.
Second, we designed a novel attention fusion-based two-stream vision transformer (AFTViT) architecture that captures long-range dependencies and diverse contextual information at multiple scales. A novel attention block is then used to integrate cross-context features at the feature level, enhancing the overall feature representation.
Third, a novel hierarchical multiscale swin (HM-Swin) transformer network with attention fusion is developed. The proposed model benefited from extracting hierarchical multiscale features from swin transformer backbone, which are further fused using a dynamic fusion block. This approach ensures the effective capture of interdependencies.
Comprehensive experiments are conducted on publicly available datasets to demonstrate the efficiency of MACNN, AFTViT and HM-Swin architectures. The results show that these methods consistently outperform existing state-of-the-art approaches in terms of classification accuracy while maintaining a minimal number of training parameters.
| Divisions: | Concordia University > Gina Cody School of Engineering and Computer Science > Electrical and Computer Engineering |
|---|---|
| Item Type: | Thesis (PhD) |
| Authors: | Ranipa, Kalpeshkumar |
| Institution: | Concordia University |
| Degree Name: | Ph. D. |
| Program: | Electrical and Computer Engineering |
| Date: | 10 July 2025 |
| Thesis Supervisor(s): | Zhu, Wei-Ping and Swamy, M.N.S. |
| ID Code: | 996170 |
| Deposited By: | Kalpeshkumar Ranipa |
| Deposited On: | 04 Nov 2025 16:14 |
| Last Modified: | 04 Nov 2025 16:14 |
References:
[1] J. Wang, Y. Chen, Y. Gu, Y. Xiao, and H. Pan, “SensoryGANs: AnEffective Generative Adversarial Framework for Sensor-based Human Activity
Recognition,” in 2018 International Joint Conference on Neural Networks
(IJCNN), pp. 1–8, jul 2018.
[2] World Health Organization (WHO), “Cardiovascular diseases (cvds) key facts,”
2025. Accessed: 2025-01-08.
[3] Centers for Disease Control and Prevention (CDC), “Heart disease facts &
statistics,” 2025. Accessed: 2024-01-02.
[4] F. De Lima, H. Advisor, M. Tavares, C. Co-Advisor:, S. Da, and S. Mattos,
“Heart Sound Segmentation for Digital Stethoscope Integration,” tech. rep.,
2011.
[5] D. S. M. Amin and B.-R. Fethi, “Features for Heartbeat Sound Signal Normal
and Pathological Computer,” may 2008.
[6] D. Roy, J. Sargeant, J. Gray, B. Hoyt, M. Allen, and M. Fleming, “Helping
family physicians improve their cardiac auscultation skills with an interactive
CD-ROM,” Journal of Continuing Education in the Health Professions, vol. 22,
pp. 152–159, may 2002.
[7] E. Etchells, C. Bell, and K. Robb, “Does This Patient Have an Abnormal
Systolic Murmur?,” JAMA, vol. 277, pp. 564–571, may 1997.
[8] M. Z. C. Lam, T. J. Lee, P. Y. Boey, W. F. Ng, H. W. Hey, K. Y. Ho, and
P. Y. Cheong, “Factors influencing cardiac auscultation proficiency in physician
trainees,” p. 4, 2005.
[9] S. M. Debbal and F. Bereksi-Reguig, “Computerized heart sounds analysis,”
Computers in Biology and Medicine, vol. 38, pp. 263–280, may 2008.
[10] F. Renna, J. Oliveira, and M. T. Coimbra, “Deep convolutional neural
networks for heart sound segmentation,” IEEE Journal of Biomedical and
Health Informatics, vol. 23, no. 6, pp. 2435–2445, 2019.
[11] C. Ahlstrom, K. Höglund, P. Hult, J. Häggström, C. Kvart, and P. Ask,
“Distinguishing Innocent Murmurs from Murmurs caused by Aortic Stenosis
by Recurrence Quantification Analysis,” vol. 2, no. 6, p. 6, 2008.
[12] W. Phanphaisarn, A. Roeksabutr, P. Wardkein, J. Koseeyaporn, and
P. P. Yupapin, “Heart detection and diagnosis based on ECG and EPCG
relationships,” Medical Devices (Auckland, N.Z.), vol. 4, pp. 133–144, may 2011.
[13] S. Yuenyong, A. Nishihara, W. Kongprawechnon, and K. Tungpimolrut,
“A framework for automatic heart sound analysis without segmentation,”
BioMedical Engineering OnLine, vol. 10, p. 13, may 2011.
[14] I. Maglogiannis, E. Loukis, E. Zafiropoulos, and A. Stasis, “Support Vectors
Machine-based identification of heart valve diseases using heart sounds,”
Computer Methods and Programs in Biomedicine, vol. 95, pp. 47–61, may 2009.
[15] N. E. Huang, Z. Shen, S. R. Long, M. C. Wu, H. H. Shih, Q. Zheng, N.-C. Yen,
C. C. Tung, and H. H. Liu, “The empirical mode decomposition and the Hilbert
spectrum for nonlinear and non-stationary time series analysis,” Proceedings of
the Royal Society of London. Series A: Mathematical, Physical and Engineering
Sciences, vol. 454, pp. 903–995, may 1998.
[16] Z. Jiang and S. Choi, “A cardiac sound characteristic waveform method for
in-home heart disorder monitoring with electric stethoscope,” Expert Systems
with Applications, vol. 31, pp. 286–298, may 2006.
[17] D. Kumar, P. Carvalho, M. Antunes, J. Henriques, A. Sa e Melo, and
J. Habetha, “Heart murmur recognition and segmentation by complexity
signatures,” in 2008 30th Annual International Conference of the IEEE
Engineering in Medicine and Biology Society, pp. 2128–2132, 2008.
[18] S. Sun, Z. Jiang, H. Wang, and Y. Fang, “Automatic moment segmentation and
peak detection analysis of heart sound pattern via short-time modified Hilbert
transform,” Computer Methods and Programs in Biomedicine, vol. 114, no. 3,
pp. 219–230, 2014.
[19] A. Moukadem, A. Dieterlen, N. Hueber, and C. Brandt, “A robust heart sounds
segmentation module based on S-transform,” Biomedical Signal Processing and
Control, vol. 8, no. 3, pp. 273–281, 2013.
[20] H. Tang, T. Li, T. Qiu, and Y. Park, “Segmentation of heart sounds based on
dynamic clustering,” Biomedical Signal Processing and Control, vol. 7, no. 5,
pp. 509–516, 2012.
[21] D. Gill, N. Gavrieli, and N. Intrator, “Detection and identification of heart
sounds using homomorphic envelogram and self-organizing probabilistic model,”
in Computers in Cardiology, 2005, pp. 957–960, IEEE, 2005.
[22] S. E. Schmidt, C. Holst-Hansen, C. Graff, E. Toft, and J. J. Struijk,
“Segmentation of heart sound recordings by a duration-dependent hidden
Markov model,” Physiological Measurement, vol. 31, no. 4, p. 513, 2010.
[23] C. D. Papadaniil and L. J. Hadjileontiadis, “Efficient heart sound segmentation
and extraction using ensemble empirical mode decomposition and kurtosis
features,” IEEE Journal of Biomedical and Health Informatics, vol. 18, no. 4,
pp. 1138–1152, 2013.
[24] J. Herzig, A. Bickel, A. Eitan, and N. Intrator, “Monitoring Cardiac Stress
Using Features Extracted From S1 Heart Sounds,” IEEE Transactions on
Biomedical Engineering, vol. 62, no. 4, pp. 1169–1178, 2015.
[25] C. Ahlstrom, P. Hult, P. Rask, J.-E. Karlsson, E. Nylander, U. Dahlström, and
P. Ask, “Feature Extraction for Systolic Heart Murmur Classification,” Annals
of Biomedical Engineering, vol. 34, pp. 1666–1677, may 2006.
[26] S. Jabbari and H. Ghassemian, “Modeling of heart systolic murmurs based on
multivariate matching pursuit for diagnosis of valvular disorders,” Computers
in Biology and Medicine, vol. 41, pp. 802–811, may 2011.
[27] A. A. Sepehri, A. Gharehbaghi, T. Dutoit, A. Kocharian, and A. Kiani, “A
novel method for pediatric heart sound segmentation without using the ECG,”
Computer Methods and Programs in Biomedicine, vol. 99, no. 1, pp. 43–48,
2010.
[28] S. W. Deng and J. Q. Han, “Towards heart sound classification without
segmentation via autocorrelation feature and diffusion maps,” Future
Generation Computer Systems, vol. 60, pp. 13–21, 2016.
[29] V. Maknickas and A. Maknickas, “Recognition of normal–abnormal phonocardiographic
signals using deep convolutional neural networks and mel-frequency
spectral coefficients,” Physiological Measurement, vol. 38, pp. 1671–1684, may
2017.
[30] F. Plesinger, I. Viscor, J. Halamek, J. Jurco, and P. Jurak, “Heart sounds
analysis using probability assessment,” Physiological Measurement, vol. 38,
pp. 1685–1700, may 2017.
[31] B. M. Whitaker, P. B. Suresha, C. Liu, G. D. Clifford, and D. V.
Anderson, “Combining sparse coding and time-domain features for heart sound
classification,” Physiological Measurement, vol. 38, pp. 1701–1713, may 2017.
[32] P. Langley and A. Murray, “Heart sound classification from unsegmented
phonocardiograms,” Physiological Measurement, vol. 38, no. 8, pp. 1658–1670,
2017.
[33] W. Zhang, J. Han, and S. Deng, “Heart sound classification based on scaled
spectrogram and partial least squares regression,” Biomedical Signal Processing
and Control, vol. 32, pp. 20–28, may 2017.
[34] Z. Ren, N. Cummins, V. Pandit, J. Han, K. Qian, and B. Schuller, “Learning
Image-based Representations for Heart Sound Classification,” DH ’18, pp. 143–
147, Association for Computing Machinery, may 2018.
[35] S. Amiriparian, M. Schmitt, N. Cummins, K. Qian, F. Dong, and B. Schuller,
“Deep Unsupervised Representation Learning for Abnormal Heart Sound
Classification,” in 2018 40th Annual International Conference of the IEEE
Engineering in Medicine and Biology Society (EMBC), pp. 4776–4779, 2018.
[36] M. E. Karar, S. H. El-Khafif, and M. A. El-Brawany, “Automated Diagnosis
of Heart Sounds Using Rule-Based Classification Tree,” Journal of Medical
Systems, vol. 41, p. 60, may 2017.
[37] S. Dodge and L. Karam, “A study and comparison of human and deep learning
recognition performance under visual distortions,” in 26th International Conf.
on Computer Communication and Networks (ICCCN), pp. 1–7, IEEE, 2017.
[38] X. Jiang, Y. Pang, X. Li, and J. Pan, “Speed up deep neural network
based pedestrian detection by sharing features across multi-scale models,”
Neurocomputing, vol. 185, pp. 163–170, may 2016.
[39] S. Amiriparian, M. Gerczuk, S. Ottl, N. Cummins, S. Pugachevskiy, and
B. Schuller, “Bag-of-Deep-Features: Noise-Robust Deep Feature Representations
for Audio Analysis,” in 2018 International Joint Conference on Neural
Networks (IJCNN), pp. 1–7, 2018.
[40] C. Szegedy, S. Ioffe, V. Vanhoucke, and A. A. Alemi, “Inception-v4, inceptionresnet
and the impact of residual connections on learning,” in Thirty-First
AAAI Conference on Artificial Intelligence, 2017.
[41] T. Young, D. Hazarika, S. Poria, and E. Cambria, “Recent Trends in
Deep Learning Based Natural Language Processing [Review Article],” IEEE
Computational Intelligence Magazine, vol. 13, no. 3, pp. 55–75, 2018.
[42] Y. Bengio, R. Ducharme, P. Vincent, and C. Jauvin, “A Neural Probabilistic
Language Model,” Journal of Machine Learning Research, vol. 3, pp. 1137–1155,
may 2003.
[43] S. Amiriparian, A. Baird, S. Julka, A. Alcorn, S. Ottl, S. Petrović,
E. Ainger, N. Cummins, and B. Schuller, “Recognition of Echolalic Autistic
Child Vocalisations Utilising Convolutional Recurrent Neural Networks,” in
Interspeech 2018, pp. 2334–2338, ISCA, may 2018.
[44] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks
for biomedical image segmentation,” in International Conf. on Medical Image
Computing and Computer-assisted Intervention, pp. 234–241, Springer, 2015.
[45] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet Classification with
Deep Convolutional Neural Networks,” in Advances in Neural Information
Processing Systems 25 (F. Pereira, C. J. C. Burges, L. Bottou, and K. Q.
Weinberger, eds.), pp. 1097–1105, Curran Associates, Inc., may 2012.
[46] E. Shelhamer, J. Long, and T. Darrell, “Fully Convolutional Networks for
Semantic Segmentation,” IEEE Transactions on Pattern Analysis and Machine
Intelligence, vol. 39, no. 4, pp. 640–651, 2017.
[47] K. Jarrett, K. Kavukcuoglu, M. Ranzato, and Y. LeCun, “What is the
best multi-stage architecture for object recognition?,” in 2009 IEEE 12th
International Conference on Computer Vision, pp. 2146–2153, 2009.
[48] Y. LeCun, K. Kavukcuoglu, and C. Farabet, “Convolutional networks and
applications in vision,” in Proceedings of 2010 IEEE International Symposium
on Circuits and Systems, pp. 253–256, 2010.
[49] Z. Li, Y. Chang, and B. W. Schuller, “CNN-based heart sound classification
with an imbalance-compensating weighted loss function,” in 2022 44th Annual
International Conference of the IEEE Engineering in Medicine Biology Society
(EMBC), pp. 4934–4937, 2022.
[50] X. Huai, L. Jiang, C. Wang, P. Chen, and H. Li, “Heart sound classification
based on convolutional neural network with convolutional block attention
module,” Frontiers in Physiology, vol. 16, p. 1596150, 2025.
[51] L. Zacharia, M. Asprou, and E. Kyriakides, “Measurement errors and delays on
wide-area control based on ieee std c37.118.1-2011: Impact and compensation,”
IEEE Systems Journal, vol. 14, no. 1, pp. 422–432, 2020.
[52] V. N. Varghees and K. Ramachandran, “Effective heart sound segmentation
and murmur classification using empirical wavelet transform and instantaneous
phase for electronic stethoscope,” IEEE Sensors Journal, vol. 17, no. 12,
pp. 3861–3872, 2017.
[53] G. Eslamizadeh and R. Barati, “Heart murmur detection based on wavelet
transformation and a synergy between artificial neural network and modified
neighbor annealing methods,” Artificial intelligence in medicine, vol. 78, pp. 23–
40, 2017.
[54] O. Deperlioglu, “Heart sound classification with signal instant energy and
stacked autoencoder network,” Biomedical Signal Processing and Control,
vol. 64, p. 102211, 2021.
[55] E. Soares, P. Angelov, and X. Gu, “Autonomous learning multiple-model zeroorder
classifier for heart sound classification,” Applied Soft Computing, vol. 94,
p. 106449, 2020.
[56] M. V. Shervegar and G. V. Bhat, “Heart sound classification using gaussian
mixture model,” Porto Biomedical Journal, vol. 3, no. 1, p. e4, 2018.
[57] J. Dastagir, F. A. Khan, M. S. Khan, and K. N. Khan, “Computeraided
phonocardiogram classification using multidomain time and frequency
features,” in 2021 International Conference on Artificial Intelligence (ICAI),
pp. 50–55, 2021.
[58] F. Noman, S.-H. Salleh, C.-M. Ting, S. B. Samdin, H. Ombao, and H. Hussain,
“A markov-switching model approach to heart sound segmentation and
classification,” IEEE Journal of Biomedical and Health Informatics, vol. 24,
no. 3, pp. 705–716, 2020.
[59] S. Khaled, M. Fakhry, H. Esmail, A. Ezzat, and E. Hamad, “Analysis of training
optimization algorithms in the narx neural network for classification of heart
sound signals,” International Journal of Scientific and Engineering Research,
vol. 13, no. 2, pp. 382–390, 2022.
[60] V. Arora, R. Leekha, R. Singh, and I. Chana, “Heart sound classification using
machine learning and phonocardiogram,” Modern Physics Letters B, vol. 33,
no. 26, p. 1950321, 2019.
[61] N. M. Khan, M. S. Khan, and G. M. Khan, “Automated heart sound
classification from unsegmented phonocardiogram signals using time frequency
features,” Int. J. Comput. Inf. Eng, vol. 12, no. 8, pp. 598–603, 2018.
[62] M. Sotaquirá, D. Alvear, and M. Mondragon, “Phonocardiogram classification
using deep neural networks and weighted probability comparisons,” Journal of
medical engineering & technology, vol. 42, no. 7, pp. 510–517, 2018.
[63] N. Ibrahim, N. Jamal, M. N. A.-H. Sha’abani, and L. F. Mahadi, “A
comparative study of heart sound signal classification based on temporal,
spectral and geometric features,” in 2020 IEEE-EMBS Conference on
Biomedical Engineering and Sciences (IECBES), pp. 24–29, 2021.
[64] B. Al-Naami, H. Fraihat, N. Y. Gharaibeh, and A.-R. M. Al-Hinnawi, “A
framework classification of heart sound signals in physionet challenge 2016 using
high order statistics and adaptive neuro-fuzzy inference system,” IEEE Access,
vol. 8, pp. 224852–224859, 2020.
[65] Y. Tan, Z. Wang, K. Qian, B. Hu, S. Zhao, B. W. Schuller, and Y. Yamamoto,
“Heart sound classification based on fractional fourier transformation entropy,”
in 2022 IEEE 4th Global Conference on Life Sciences and Technologies
(LifeTech), pp. 588–589, IEEE, 2022.
[66] W. Zhang, J. Han, and S. Deng, “Heart sound classification based on scaled
spectrogram and partial least squares regression,” Biomedical Signal Processing
and Control, vol. 32, pp. 20–28, 2017.
[67] A. Yadav, M. K. Dutta, C. M. Travieso, and J. B. Alonso, “Automatic
classification of normal and abnormal pcg recording heart sound recording using
fourier transform,” in 2018 IEEE International Work Conference on Bioinspired
Intelligence (IWOBI), pp. 1–9, 2018.
[68] B. Ergen, Y. Tatar, and H. O. Gulcur, “Time–frequency analysis of
phonocardiogram signals using wavelet transform: a comparative study,”
Computer Methods in Biomechanics and Biomedical Engineering, vol. 15,
pp. 371–381, may 2012.
[69] S. Debbal and A. M. Tani, “Heart sounds analysis and murmurs,” International
Journal of Medical Engineering and Informatics, vol. 8, pp. 49–62, may 2015.
[70] S. Chauhan, P. Wang, C. S. Lim, and V. Anantharaman, “A computer-aided
MFCC-based HMM system for automatic auscultation,” Computers in biology
and medicine, vol. 38, no. 2, pp. 221–233, 2008.
[71] Y. Soeta and Y. Bito, “Detection of features of prosthetic cardiac valve sound
by spectrogram analysis,” Applied Acoustics, vol. 89, pp. 28–33, may 2015.
[72] J. Singh and R. S. Anand, “Computer aided analysis of phonocardiogram,”
Journal of Medical Engineering & Technology, vol. 31, pp. 319–323, may 2007.
[73] M. Baydoun, L. Safatly, H. Ghaziri, and A. El Hajj, “Analysis of heart sound
anomalies using ensemble learning,” Biomedical Signal Processing and Control,
vol. 62, p. 102019, 2020.
[74] W. Zeng, J. Yuan, C. Yuan, Q. Wang, F. Liu, and Y. Wang, “A new approach
for the detection of abnormal heart sound signals using tqwt, vmd and neural
networks,” Artificial Intelligence Review, vol. 54, no. 3, pp. 1613–1647, 2021.
[75] H. M. Hadi, M. Y. Mashor, M. S. Mohamed, and K. B. Tat, “Classification of
heart sounds using wavelets and neural networks,” in 2008 5th International
Conference on Electrical Engineering, Computing Science and Automatic
Control, pp. 177–180, 2008.
[76] F. Rios-Gutierrez, R. Alba-Flores, and S. Strunic, “Recognition and classification
of cardiac murmurs using ann and segmentation,” in CONIELECOMP
2012, 22nd International Conference on Electrical Communications and
Computers, pp. 219–223, 2012.
[77] T. Ölmez and Z. Dokur, “Classification of heart sounds using an artificial neural
network,” Pattern recognition letters, vol. 24, no. 1-3, pp. 617–629, 2003.
[78] C. N. Gupta, R. Palaniappan, S. Swaminathan, and S. M. Krishnan, “Neural
network classification of homomorphic segmented heart sounds,” Applied soft
computing, vol. 7, no. 1, pp. 286–297, 2007.
[79] R. K. Sinha, Y. Aggarwal, and B. N. Das, “Backpropagation artificial neural
network classifier to detect changes in heart sound due to mitral valve
regurgitation,” Journal of Medical Systems, vol. 31, pp. 205–209, 2007.
[80] J. Martínez-Alajarín, J. López-Candel, and R. Ruiz-Merino, “Classification and
diagnosis of heart sounds and murmurs using artificial neural networks,” in Bioinspired
Modeling of Cognitive Tasks: Second International Work-Conference
on the Interplay Between Natural and Artificial Computation, IWINAC 2007,
La Manga del Mar Menor, Spain, June 18-21, 2007, Proceedings, Part I 2,
pp. 303–312, Springer, 2007.
[81] S. R. Bhatikar, C. DeGroff, and R. L. Mahajan, “A classifier based on the
artificial neural network approach for cardiologic auscultation in pediatrics,”
Artificial intelligence in medicine, vol. 33, no. 3, pp. 251–260, 2005.
[82] T.-c. I. Yang and H. Hsieh, “Classification of acoustic physiological signals based
on deep learning neural networks with augmented features,” in 2016 Computing
in Cardiology Conference (CinC), pp. 569–572, 2016.
[83] A. Raza, A. Mehmood, S. Ullah, M. Ahmad, G. S. Choi, and B.-W. On,
“Heartbeat sound signal classification using deep learning,” Sensors, vol. 19,
no. 21, p. 4819, 2019.
[84] F. A. Khan, A. Abid, and M. S. Khan, “Automatic heart sound
classification from segmented/unsegmented phonocardiogram signals using time
and frequency features,” Physiological measurement, vol. 41, no. 5, p. 055006,
2020.
[85] S. Latif, M. Usman, R. Rana, and J. Qadir, “Phonocardiographic sensing using
deep learning for abnormal heartbeat detection,” IEEE Sensors Journal, vol. 18,
no. 22, pp. 9393–9400, 2018.
[86] N. Baghel, M. K. Dutta, and R. Burget, “Automatic diagnosis of multiple
cardiac diseases from pcg signals using convolutional neural network,” Computer
Methods and Programs in Biomedicine, vol. 197, p. 105750, 2020.
[87] R. Avanzato and F. Beritelli, “Heart sound multiclass analysis based on raw
data and convolutional neural network,” IEEE Sensors Letters, vol. 4, no. 12,
pp. 1–4, 2020.
[88] T. Nilanon, J. Yao, J. Hao, S. Purushotham, and Y. Liu, “Normal / abnormal
heart sound recordings classification using convolutional neural network,” in
Computing in Cardiology, vol. 43, pp. 585–588, IEEE, 2016.
[89] J. P. Dominguez-Morales, A. F. Jimenez-Fernandez, M. J. Dominguez-Morales,
and G. Jimenez-Moreno, “Deep Neural Networks for the Recognition and
Classification of Heart Murmurs Using Neuromorphic Auditory Sensors,” IEEE
Transactions on Biomedical Circuits and Systems, vol. 12, no. 1, pp. 24–34,
2018.
[90] M. Tschannen, T. Kramer, G. Marti, M. Heinzmann, and T. Wiatowski, “Heart
sound classification using deep structured features,” in 2016 Computing in
Cardiology Conference (CinC), pp. 565–568, 2016.
[91] Y. Zhang, S. Ayyar, L.-H. Chen, and E. J. Li, “Segmental Convolutional
Neural Networks for Detection of Cardiac Abnormality With Noisy Heart Sound
Recordings,” dec 2016.
[92] B. Xiao, Y. Xu, X. Bi, W. Li, Z. Ma, J. Zhang, and X. Ma, “Follow the sound of
children’s heart: a deep-learning-based computer-aided pediatric chds diagnosis
system,” IEEE Internet of Things Journal, vol. 7, no. 3, pp. 1994–2004, 2019.
[93] S. L. Oh, V. Jahmunah, C. P. Ooi, R.-S. Tan, E. J. Ciaccio, T. Yamakawa,
M. Tanabe, M. Kobayashi, and U. R. Acharya, “Classification of heart sound
signals using a novel deep wavenet model,” Computer Methods and Programs
in Biomedicine, vol. 196, p. 105604, 2020.
[94] P. Upretee and M. E. Yüksel, “Accurate classification of heart sounds for disease
diagnosis by a single time-varying spectral feature: Preliminary results,” in
2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and
Computer Science (EBBT), pp. 1–4, IEEE, 2019.
[95] Z. Abduh, E. A. Nehary, M. A. Wahed, and Y. M. Kadah, “Classification of
heart sounds using fractional fourier transform based mel-frequency spectral
coefficients and traditional classifiers,” Biomedical Signal Processing and
Control, vol. 57, p. 101788, 2020.
[96] F. Li, H. Tang, S. Shang, K. Mathiak, and F. Cong, “Classification of heart
sounds using convolutional neural network,” Applied Sciences, vol. 10, no. 11,
p. 3956, 2020.
[97] B. Xiao, Y. Xu, X. Bi, J. Zhang, and X. Ma, “Heart sounds classification
using a novel 1-d convolutional neural network with extremely low parameter
consumption,” Neurocomputing, vol. 392, pp. 153–159, 2020.
[98] C. Thomae and A. Dominik, “Using deep gated rnn with a convolutional
front end for end-to-end classification of heart sound,” in 2016 Computing in
Cardiology Conference (CinC), pp. 625–628, IEEE, 2016.
[99] M. Alkhodari and L. Fraiwan, “Convolutional and recurrent neural networks
for the detection of valvular heart diseases in phonocardiogram recordings,”
Computer Methods and Programs in Biomedicine, vol. 200, p. 105940, 2021.
[100] A. Maity, A. Pathak, and G. Saha, “Transfer learning based heart valve disease
classification from phonocardiogram signal,” Biomedical Signal Processing and
Control, vol. 85, p. 104805, 2023.
[101] J. Karhade, S. Dash, S. K. Ghosh, D. K. Dash, and R. K. Tripathy,
“Time–frequency-domain deep learning framework for the automated detection
of heart valve disorders using pcg signals,” IEEE Transactions on Instrumentation
and Measurement, vol. 71, pp. 1–11, 2022.
[102] B. Kaushal, S. Raveendran, M. D. Patil, and G. K. Birajdar, “Deep autoencoder
neural networks for heart sound classification,” in Artificial Intelligence in
Medical Virology, pp. 165–189, Springer, 2023.
[103] D. Ramachandram and G. W. Taylor, “Deep multimodal learning: A survey on
recent advances and trends,” IEEE Signal Processing Magazine, vol. 34, no. 6,
pp. 96–108, 2017.
[104] D. Michelsanti, Z.-H. Tan, S.-X. Zhang, Y. Xu, M. Yu, D. Yu, and J. Jensen,
“An overview of deep-learning-based audio-visual speech enhancement and
separation,” arXiv preprint arXiv:2008.09586, 2020.
[105] G.-Y. Son, S. Kwon, et al., “Classification of heart sound signal using multiple
features,” Applied Sciences, vol. 8, no. 12, p. 2344, 2018.
[106] S. K. Ghosh, R. K. Tripathy, R. Ponnalagu, and R. B. Pachori, “Automated
detection of heart valve disorders from the pcg signal using time-frequency
magnitude and phase features,” IEEE Sensors Letters, vol. 3, no. 12, pp. 1–4,
2019.
[107] S. Aziz, M. U. Khan, M. Alhaisoni, T. Akram, and M. Altaf, “Phonocardiogram
signal processing for automatic diagnosis of congenital heart disorders through
fusion of temporal and cepstral features,” Sensors, vol. 20, no. 13, p. 3790, 2020.
[108] K. Iqtidar, U. Qamar, S. Aziz, and M. U. Khan, “Phonocardiogram signal
analysis for classification of coronary artery diseases using mfcc and 1d adaptive
local ternary patterns,” Computers in Biology and Medicine, vol. 138, p. 104926,
2021.
[109] M. Abdollahpur, A. Ghaffari, S. Ghiasi, and M. J. Mollakazemi, “Detection of
pathological heart sounds,” Physiological measurement, vol. 38, no. 8, p. 1616,
2017.
[110] C. Potes, S. Parvaneh, A. Rahman, and B. Conroy, “Ensemble of feature-based
and deep learning-based classifiers for detection of abnormal heart sounds,” in
2016 computing in cardiology conference (CinC), pp. 621–624, IEEE, 2016.
[111] J. M.-T. Wu, M.-H. Tsai, Y. Z. Huang, S. H. Islam, M. M. Hassan, A. Alelaiwi,
and G. Fortino, “Applying an ensemble convolutional neural network with
savitzky–golay filter to construct a phonocardiogram prediction model,” Applied
Soft Computing, vol. 78, pp. 29–40, 2019.
[112] F. Noman, C.-M. Ting, S.-H. Salleh, and H. Ombao, “Short-segment heart
sound classification using an ensemble of deep convolutional neural networks,”
in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP), pp. 1318–1322, IEEE, 2019.
[113] S. K. Ghosh, R. Ponnalagu, R. Tripathy, and U. R. Acharya, “Automated
detection of heart valve diseases using chirplet transform and multiclass
composite classifier with pcg signals,” Computers in biology and medicine,
vol. 118, p. 103632, 2020.
[114] H. Li, X. Wang, C. Liu, Q. Zeng, Y. Zheng, X. Chu, L. Yao, J. Wang,
Y. Jiao, and C. Karmakar, “A fusion framework based on multi-domain features
and deep learning features of phonocardiogram for coronary artery disease
detection,” Computers in biology and medicine, vol. 120, p. 103733, 2020.
[115] F. Demir, A. Şengür, V. Bajaj, and K. Polat, “Towards the classification of
heart sounds based on convolutional deep neural network,” Health information
science and systems, vol. 7, no. 1, pp. 1–9, 2019.
[116] S. B. Shuvo, S. N. Ali, S. I. Swapnil, M. S. Al-Rakhami, and A. Gumaei,
“Cardioxnet: A novel lightweight deep learning framework for cardiovascular
disease classification using heart sound recordings,” IEEE Access, vol. 9,
pp. 36955–36967, 2021.
[117] K. Ranipa, W.-P. Zhu, and M. Swamy, “Multimodal CNN fusion architecture
with multi-features for heart sound classification,” in 2021 IEEE International
Symposium on Circuits and Systems (ISCAS), pp. 1–5, IEEE, 2021.
[118] S. Davis and P. Mermelstein, “Comparison of parametric representations
for monosyllabic word recognition in continuously spoken sentences,” IEEE
transactions on acoustics, speech, and signal processing, vol. 28, no. 4, pp. 357–
366, 1980.
[119] A. v. d. Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves,
N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, “Wavenet: A generative
model for raw audio,” arXiv preprint arXiv:1609.03499, 2016.
[120] L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution
for semantic image segmentation,” arXiv preprint arXiv:1706.05587,
2017.
[121] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image
recognition,” in Proceedings of the IEEE conference on computer vision and
pattern recognition, pp. 770–778, 2016.
[122] S. Woo, J. Park, J.-Y. Lee, and I. S. Kweon, “Cbam: Convolutional block
attention module,” in Proceedings of the European conference on computer
vision (ECCV), pp. 3–19, 2018.
[123] C. Liu, D. Springer, Q. Li, B. Moody, R. A. Juan, F. J. Chorro, F. Castells, J. M.
Roig, I. Silva, A. E. Johnson, et al., “An open access database for the evaluation
of heart sound algorithms,” Physiological Measurement, vol. 37, no. 12, p. 2181,
2016.
[124] M. Guven and F. Uysal, “A new method for heart disease detection: Long
short-term feature extraction from heart sound data,” Sensors, vol. 23, no. 13,
p. 5835, 2023.
[125] P. T. Krishnan, P. Balasubramanian, and S. Umapathy, “Automated heart
sound classification system from unsegmented phonocardiogram (pcg) using
deep neural network,” Physical and Engineering Sciences in Medicine, pp. 1–
11, 2020.
[126] M. T. Nguyen, W. W. Lin, and J. H. Huang, “Heart sound classification using
deep learning techniques based on log-mel spectrogram,” Circuits, Systems, and
Signal Processing, vol. 42, no. 1, pp. 344–360, 2023.
[127] X. Zhou, X. Guo, Y. Zheng, and Y. Zhao, “Detection of coronary heart disease
based on mfcc characteristics of heart sound,” Applied Acoustics, vol. 212,
p. 109583, 2023.
[128] A. Harimi, Y. Majd, A. A. Gharahbagh, V. Hajihashemi, Z. Esmaileyan, J. J.
Machado, and J. M. R. Tavares, “Classification of heart sounds using chaogram
transform and deep convolutional neural network transfer learning,” Sensors,
vol. 22, no. 24, p. 9569, 2022.
[129] O. R. A. Almanifi, A. F. Ab Nasir, M. A. M. Razman, R. M. Musa, and A. P. A.
Majeed, “Heartbeat murmurs detection in phonocardiogram recordings via
transfer learning,” Alexandria Engineering Journal, vol. 61, no. 12, pp. 10995–
11002, 2022.
[130] M. Xiang, J. Zang, J. Wang, H. Wang, C. Zhou, R. Bi, Z. Zhang, and
C. Xue, “Research of heart sound classification using two-dimensional features,”
Biomedical Signal Processing and Control, vol. 79, p. 104190, 2023.
[131] J. Torre-Cruz, F. Canadas-Quesada, N. Ruiz-Reyes, P. Vera-Candeas, S. Garcia-
Galan, J. Carabias-Orti, and J. Ranilla, “Detection of valvular heart diseases
combining orthogonal non-negative matrix factorization and convolutional
neural networks in pcg signals,” Journal of Biomedical Informatics, vol. 145,
p. 104475, 2023.
[132] E. Bruoth, P. Bugata, D. Gajdoš, D. Hudák, V. Kmečová, M. Staňková,
A. Szabari, G. Vozáriková, et al., “Murmur identification using supervised
contrastive learning,” in 2022 Computing in Cardiology (CinC), vol. 498, pp. 1–
4, IEEE, 2022.
[133] D. S. Panah, A. Hines, and S. McKeever, “Exploring wav2vec 2.0 model for
heart murmur detection,” in 2023 31st European Signal Processing Conference
(EUSIPCO), pp. 1010–1014, IEEE, 2023.
[134] Yaseen, G.-Y. Son, and S. Kwon, “Classification of heart sound signal using
multiple features,” Applied Sciences, vol. 8, no. 12, p. 2344, 2018.
[135] F. B. Azam, M. I. Ansari, I. Mclane, and T. Hasan, “Heart sound
classification considering additive noise and convolutional distortion,” arXiv
preprint arXiv:2106.01865, 2021.
[136] C. Xu, X. Li, X. Zhang, R. Wu, Y. Zhou, Q. Zhao, Y. Zhang, S. Geng,
Y. Gu, and S. Hong, “Cardiac murmur grading and risk analysis of cardiac
diseases based on adaptable heterogeneous-modality multi-task learning,”
Health Information Science and Systems, vol. 12, no. 1, p. 2, 2023.
[137] E. A. Nehary and S. Rajan, “Phonocardiogram classification by learning from
positive and unlabeled examples,” IEEE Transactions on Instrumentation and
Measurement, vol. 73, pp. 1–14, 2024.
[138] J. Zang, C. Lian, B. Xu, Z. Zhang, Y. Su, and C. Xue, “Amtnet: Attentional
multi-scale temporal network for phonocardiogram signal classification,”
Biomedical Signal Processing and Control, vol. 85, p. 104934, 2023.
[139] T. Koike, K. Qian, Q. Kong, M. D. Plumbley, B. W. Schuller, and Y. Yamamoto,
“Audio for audio is better? an investigation on transfer learning models for
heart sound classification,” in 2020 42nd Annual International Conference of the
IEEE Engineering in Medicine & Biology Society (EMBC), pp. 74–77, IEEE,
2020.
[140] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez,
Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural
information processing systems, vol. 30, 2017.
[141] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “Bert: Pre-training of deep
bidirectional transformers for language understanding. arxiv,” arXiv preprint
arXiv:1810.04805, 2019.
[142] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, “Improving
language understanding by generative pre-training,” 2018. Technical report,
OpenAI, San Francisco, CA.
[143] D. Yang, Y. Lin, J. Wei, X. Lin, X. Zhao, Y. Yao, T. Tao, B. Liang, and S.-G.
Lu, “Assisting heart valve diseases diagnosis via transformer-based classification
of heart sound signals,” Electronics, vol. 12, no. 10, p. 2221, 2023.
[144] M. Alkhodari, L. J. Hadjileontiadis, and A. H. Khandoker, “Identification
of congenital valvular murmurs in young patients using deep learning-based
attention transformers and phonocardiograms,” IEEE Journal of Biomedical
and Health Informatics, pp. 1–13, 2024.
[145] T. Jumphoo, K. Phapatanaburi, W. Pathonsuwan, P. Anchuen, M. Uthansakul,
and P. Uthansakul, “Exploiting data-efficient image transformer-based transfer
learning for valvular heart diseases detection,” IEEE Access, 2024.
[146] A. Graves, “Neural turing machines,” arXiv preprint arXiv:1410.5401, 2014.
[147] M. A. Reyna, Y. Kiarashi, A. Elola, J. Oliveira, F. Renna, A. Gu, E. A.
Perez Alday, N. Sadr, A. Sharma, J. Kpodonu, et al., “Heart murmur detection
from phonocardiogram recordings: The george b. moody physionet challenge
2022,” PLOS Digital Health, vol. 2, no. 9, p. e0000324, 2023.
[148] K. Simonyan, “Very deep convolutional networks for large-scale image
recognition,” arXiv preprint arXiv:1409.1556, 2014.
[149] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking
the inception architecture for computer vision,” in Proceedings of the IEEE
conference on computer vision and pattern recognition, pp. 2818–2826, 2016.
[150] F. Chollet, “Xception: Deep learning with depthwise separable convolutions,” in
Proceedings of the IEEE conference on computer vision and pattern recognition,
pp. 1251–1258, 2017.
[151] W. Chen, Z. Zhou, J. Bao, C. Wang, H. Chen, C. Xu, G. Xie, H. Shen, and
H. Wu, “Classifying heart-sound signals based on CNN trained on melspectrum
and log-melspectrum features,” Bioengineering, vol. 10, no. 6, p. 645, 2023.
[152] D. Riccio, N. Brancati, G. Sannino, L. Verde, and M. Frucci, “CNN-based
classification of phonocardiograms using fractal techniques,” Biomedical Signal
Processing and Control, vol. 86, p. 105186, 2023.
[153] C. Xu, Z. Yang, and C. Zhu, “A heart sound classification algorithm based on
bispectral analysis and deep learning,” in Proceedings of the 2022 International
Conference on Intelligent Medicine and Health, pp. 1–6, 2022.
[154] G. Tian, C. Lian, Z. Zeng, B. Xu, Y. Su, J. Zang, Z. Zhang, and C. Xue,
“Imbalanced heart sound signal classification based on two-stage trained
dsanet,” Cognitive Computation, vol. 14, no. 4, pp. 1378–1391, 2022.
[155] C. Yin, Y. Zheng, X. Ding, Y. Shi, J. Qin, and X. Guo, “Detection of coronary
artery disease based on clinical phonocardiogram and multiscale attention
convolutional compression network,” IEEE Journal of Biomedical and Health
Informatics, 2024.
[156] S. Han, W. Jeon, W. Gong, and I.-Y. Kwak, “Mcheart: multi-channel-based
heart signal processing scheme for heart noise detection using deep learning,”
Biology, vol. 12, no. 10, p. 1291, 2023.
[157] S. B. Shuvo, S. S. Alam, S. U. Ayman, A. Chakma, P. D. Barua, and
U. R. Acharya, “Nrc-net: Automated noise robust cardio net for detecting
valvular cardiac diseases using optimum transformation method with heart
sound signals,” Biomedical Signal Processing and Control, vol. 86, p. 105272,
2023.
[158] H. Yadav, P. Shah, N. Gandhi, T. Vyas, A. Nair, S. Desai, L. Gohil, S. Tanwar,
R. Sharma, V. Marina, et al., “CNN and bidirectional gru-based heartbeat
sound classification architecture for elderly people,” Mathematics, vol. 11, no. 6,
p. 1365, 2023.
[159] H. Lu, J. B. Yip, T. Steigleder, S. Grießhammer, M. Heckel, N. V. S. J.
Jami, B. Eskofier, C. Ostgathe, and A. Koelpin, “A lightweight robust
approach for automatic heart murmurs and clinical outcomes classification
from phonocardiogram recordings,” in 2022 Computing in Cardiology (CinC),
vol. 498, pp. 1–4, IEEE, 2022.
[160] Z. Guo, J. Chen, T. He, W. Wang, H. Abbas, and Z. Lv, “DS-CNN: Dual-stream
convolutional neural networks-based heart sound classification for wearable
devices,” IEEE Transactions on Consumer Electronics, vol. 69, no. 4, pp. 1186–
1194, 2023.
[161] K. Ranipa, W.-P. Zhu, and M. Swamy, “A novel feature-level fusion scheme with
multimodal attention CNN for heart sound classification,” Computer Methods
and Programs in Biomedicine, vol. 248, p. 108122, 2024.
[162] H. Zhang, P. Zhang, Z. Wang, L. Chao, Y. Chen, and Q. Li, “Multifeature
decision fusion network for heart sound abnormality detection and
classification,” IEEE Journal of Biomedical and Health Informatics, 2023.
[163] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai,
T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., “An
image is worth 16x16 words: Transformers for image recognition at scale,” arXiv
preprint arXiv:2010.11929, 2020.
[164] J. Kim, G. Park, and B. Suh, “Classification of phonocardiogram recordings
using vision transformer architecture,” in 2022 Computing in Cardiology
(CinC), vol. 498, pp. 1–4, IEEE, 2022.
[165] Z. Liu, W. Liu, Z. Gu, and F. Yang, “Efficient anomaly detection algorithm for
heart sound signal,” IEEE Access, 2024.
[166] Q. Abbas, A. Hussain, and A. R. Baig, “Automatic detection and classification
of cardiovascular disorders using phonocardiogram and convolutional vision
transformers,” diagnostics, vol. 12, no. 12, p. 3109, 2022.
[167] M. Hassanuzzaman, N. A. Hasan, M. A. Al Mamun, M. Alkhodari, K. I.
Ahmed, A. H. Khandoker, and R. Mostafa, “Recognition of pediatric
congenital heart diseases by using phonocardiogram signals and transformerbased
neural networks,” in 2023 45th Annual International Conference of the
IEEE Engineering in Medicine & Biology Society (EMBC), pp. 1–4, IEEE,
2023.
[168] M. Hassanuzzaman, N. A. Hasan, M. A. A. Mamun, K. I. Ahmed, A. H.
Khandoker, and R. Mostafa, “Classification of short segment pediatric heart
sounds based on a transformer-based convolutional neural network,” arXiv
preprint arXiv:2404.00470, 2024.
[169] P. Fan, Y. Shu, and Y. Han, “Transformer embedded with learnable filters for
heart murmur detection,” in 2022 Computing in Cardiology (CinC), vol. 498,
pp. 1–4, IEEE, 2022.
[170] J. L. Costa, P. Couto, and R. Rodrigues, “Multitask and transfer learning
for cardiac abnormality detections in heart sounds,” in 2022 Computing in
Cardiology (CinC), vol. 498, pp. 1–4, IEEE, 2022.
[171] D. Niizumi, D. Takeuchi, Y. Ohishi, N. Harada, and K. Kashino, “Exploring
pre-trained general-purpose audio representations for heart murmur detection,”
in 2024 46th Annual International Conference of the IEEE Engineering in
Medicine and Biology Society (EMBC), pp. 1–4, IEEE, 2024.
[172] Z. Tao, Z. Ren, X. Yang, Y. Liang, and X. Shi, “2d vit and 1d crnn-based
heart sound signals detection model,” in 2023 IEEE International Conference
on Bioinformatics and Biomedicine (BIBM), pp. 2265–2268, IEEE, 2023.
[173] Z. Zhang, T. Pang, J. Han, and B. W. Schuller, “Intelligent cardiac auscultation
for murmur detection via parallel-attentive models with uncertainty estimation,”
in ICASSP 2024-2024 IEEE International Conference on Acoustics,
Speech and Signal Processing (ICASSP), pp. 861–865, IEEE, 2024.
[174] J. Wang, J. Zang, S. Yao, Z. Zhang, and C. Xue, “Multiclassification for heart
sound signals under multiple networks and multi-view feature,” Measurement,
vol. 225, p. 114022, 2024.
[175] R. Wang, Y. Duan, Y. Li, D. Zheng, X. Liu, C. T. Lam, and T. Tan, “Pctmfnet:
heart sound classification with parallel CNNs-transformer and second-order
spectral analysis,” The Visual Computer, vol. 39, no. 8, pp. 3811–3822, 2023.
[176] Y. Liu, E. Sangineto, W. Bi, N. Sebe, B. Lepri, and M. Nadai, “Efficient training
of visual transformers with small datasets,” Advances in Neural Information
Processing Systems, vol. 34, pp. 23818–23830, 2021.
[177] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo,
“Swin transformer: Hierarchical vision transformer using shifted windows,”
in Proceedings of the IEEE/CVF international conference on computer vision,
pp. 10012–10022, 2021.
[178] V. C. Nitesh, “Smote: synthetic minority over-sampling technique,” J Artif
Intell Res, vol. 16, no. 1, p. 321, 2002.
[179] H. M. Nguyen, E. W. Cooper, and K. Kamei, “Borderline over-sampling
for imbalanced data classification,” International Journal of Knowledge
Engineering and Soft Data Paradigms, vol. 3, no. 1, pp. 4–21, 2011.
[180] E. Messner, M. Zöhrer, and F. Pernkopf, “Heart Sound Segmentation—An
Event Detection Approach Using Deep Recurrent Neural Networks,” IEEE
Transactions on Biomedical Engineering, vol. 65, no. 9, pp. 1964–1974, 2018.
[181] M. Frid-Adar, E. Klang, M. Amitai, J. Goldberger, and H. Greenspan, “Synthetic
data augmentation using GAN for improved liver lesion classification,” in
2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018),
pp. 289–293, apr 2018.
[182] A. Madani, M. Moradi, A. Karargyris, and T. Syeda-Mahmood, “Chest
x-ray generation and data augmentation for cardiovascular abnormality
classification,” in Medical Imaging 2018: Image Processing (E. D. Angelini and
B. A. Landman, eds.), vol. 10574, pp. 415–420, International Society for Optics
and Photonics, SPIE, 2018.
[183] H. Wu, B. Xiao, N. Codella, M. Liu, X. Dai, L. Yuan, and L. Zhang,
“Cvt: Introducing convolutions to vision transformers,” in Proceedings of the
IEEE/CVF international conference on computer vision, pp. 22–31, 2021.
[184] L. Yuan, Y. Chen, T. Wang, W. Yu, Y. Shi, Z.-H. Jiang, F. E. Tay, J. Feng,
and S. Yan, “Tokens-to-token vit: Training vision transformers from scratch
on imagenet,” in Proceedings of the IEEE/CVF international conference on
computer vision, pp. 558–567, 2021.
[185] W. Wang, E. Xie, X. Li, D.-P. Fan, K. Song, D. Liang, T. Lu, P. Luo,
and L. Shao, “Pyramid vision transformer: A versatile backbone for
dense prediction without convolutions,” in Proceedings of the IEEE/CVF
international conference on computer vision, pp. 568–578, 2021.
[186] T. Ye, L. Dong, Y. Xia, Y. Sun, Y. Zhu, G. Huang, and F. Wei, “Differential
transformer,” arXiv preprint arXiv:2410.05258, 2024.
Repository Staff Only: item control page


Download Statistics
Download Statistics