Europe PMC

This website requires cookies, and the limited processing of your personal data in order to function. By using the site you are agreeing to this as outlined in our privacy notice and cookie policy.

Abstract 


Increased life expectancy in most countries is a result of continuous improvements at all levels, starting from medicine and public health services, environmental and personal hygiene to the use of the most advanced technologies by healthcare providers. Despite these significant improvements, especially at the technological level in the last few decades, the overall access to healthcare services and medical facilities worldwide is not equally distributed. Indeed, the end beneficiary of these most advanced healthcare services and technologies on a daily basis are mostly residents of big cities, whereas the residents of rural areas, even in developed countries, have major difficulties accessing even basic medical services. This may lead to huge deficiencies in timely medical advice and assistance and may even cause death in some cases. Remote healthcare is considered a serious candidate for facilitating access to health services for all; thus, by using the most advanced technologies, providing at the same time high quality diagnosis and ease of implementation and use. ECG analysis and related cardiac diagnosis techniques are the basic healthcare methods providing rapid insights in potential health issues through simple visualization and interpretation by clinicians or by automatic detection of potential cardiac anomalies. In this paper, we propose a novel machine learning (ML) architecture for the ECG classification regarding five heart diseases based on temporal convolution networks (TCN). The proposed design, which implements a dilated causal one-dimensional convolution on the input heartbeat signals, seems to be outperforming all existing ML methods with an accuracy of 96.12% and an F1 score of 84.13%, using a reduced number of parameters (10.2 K). Such results make the proposed TCN architecture a good candidate for low power consumption hardware platforms, and thus its potential use in low cost embedded devices for remote health monitoring.

Free full text 


Logo of sensorsLink to Publisher's site
Sensors (Basel). 2023 Feb; 23(3): 1697.
Published online 2023 Feb 3. https://doi.org/10.3390/s23031697
PMCID: PMC9920651
PMID: 36772737

ECG Classification Using an Optimal Temporal Convolutional Network for Remote Health Monitoring

Annie Lanzolla, Academic Editor

Abstract

Increased life expectancy in most countries is a result of continuous improvements at all levels, starting from medicine and public health services, environmental and personal hygiene to the use of the most advanced technologies by healthcare providers. Despite these significant improvements, especially at the technological level in the last few decades, the overall access to healthcare services and medical facilities worldwide is not equally distributed. Indeed, the end beneficiary of these most advanced healthcare services and technologies on a daily basis are mostly residents of big cities, whereas the residents of rural areas, even in developed countries, have major difficulties accessing even basic medical services. This may lead to huge deficiencies in timely medical advice and assistance and may even cause death in some cases. Remote healthcare is considered a serious candidate for facilitating access to health services for all; thus, by using the most advanced technologies, providing at the same time high quality diagnosis and ease of implementation and use. ECG analysis and related cardiac diagnosis techniques are the basic healthcare methods providing rapid insights in potential health issues through simple visualization and interpretation by clinicians or by automatic detection of potential cardiac anomalies. In this paper, we propose a novel machine learning (ML) architecture for the ECG classification regarding five heart diseases based on temporal convolution networks (TCN). The proposed design, which implements a dilated causal one-dimensional convolution on the input heartbeat signals, seems to be outperforming all existing ML methods with an accuracy of 96.12% and an F1 score of 84.13%, using a reduced number of parameters (10.2 K). Such results make the proposed TCN architecture a good candidate for low power consumption hardware platforms, and thus its potential use in low cost embedded devices for remote health monitoring.

Keywords: ECG, temporal convolution, TCN, healthcare

1. Introduction

Electrocardiogram (ECG) is a rapid bedside inspection that measures the electrical activity generated by the heart as it contracts. It is commonly used to recognize diverse heart diseases such as arrhythmia, cardiomyopathy, coronary heart disease, cardiovascular disease, and many others. The inspection process has always been carried out by physicians and clinicians, which is a time-consuming procedure requiring significant medical and human resources to process the large amount of ECG data [1]. On the other hand, due to the diversity of ECG signals, many issues could arise, making this process of ECG inspection even more challenging. For example, the ECG of two healthy people may not be completely similar. Moreover, two patients suffering from the same heart disease could show different signs in their ECGs. Another issue could be that two different diseases have very close signals at the ECG level. It seems that there are no definite standards to be used in the diagnosis process [2]. For that reason, the use of artificial intelligence (AI) methods are needed, as these methods are learnable through accumulated experiences such that they could find hidden patterns that humans cannot find.

The computerized analysis of ECG signals was mainly meant to improve the diagnosis process, save time, and target rural and remote regions where medical specialists are not always affordable [3]. To this end, millions of ECG schemes are recorded worldwide every year, where most of them are automatically analyzed and decided afterwards. However, a false analysis is very likely, especially in the case of inexperienced clinicians who might endorse any automated results without further analysis. Such clinical mismanagement mostly ends up with a useless or even dangerous treatment. Thus, it becomes necessary that the ECG results are read and approved by well-experienced physicians. On the other hand, doctors highly recommend the modernization of existing computerized ECG analysis methods as well as the improvement of their robustness for more reliable medication.

Machine learning (ML) has proven to be eminently successful in different classification problems. This opened the door for its use in ECG analysis problems, and various ML-based methods have been recorded in this domain. In [4], ECG signals of normal people are collected and compared to ECG signals under tests using cross-correlation techniques. This allows for the detection of ECG signals of patients with myocardial infarction with an accuracy of 91.5% and a F1 score of 90.8%. In [5], a hybrid model of a decision tree with the C4.5 algorithm is applied on ECG features after they have been extracted using the genetic algorithm. The model was tested on the UCI arrhythmia dataset in two modes: 2-class and 16-class, ending up with highest evaluation metrics that have been recorded for the UCI arrhythmia dataset. In [6], two life-threatening arrhythmias, AFIB and AFL, are considered. A residual deep neural network architecture is proposed to detect the presence of such arrhythmias based on the RR interval of the ECG signals. The inputs are first extracted, denoised, and then normalized before being introduced to the network. A 10-fold cross-validation is carried out in training, leading to massive results in terms of accuracy and other metrics. In [7], the authors introduced a model based on a feed-forward multilayer neural network with error back propagation learning algorithm for the diagnosis of ischemic heart disease. The resulting high-order statistics facilitate the discrimination between the nonlinear dynamics of normal and diseased cases. Another feed-forward network for the same task was proposed in [8]. The network is deeper in terms of used layers, which makes it capable of classifying 6 ECG abnormalities that are representative of both rhythmic and morphological ECG abnormalities. The authors in this paper emphasize the need for expert review of borderline and complex cases after any automated classification. On the other hand, different prevalent deep learning architectures such as GoogLeNet, ResNet, and LSTM have shown great performance in the ECG domain [9,10,11]. This comes at the cost of storage and computation, as these models include millions of learnable parameters. Therefore, the adoption of such networks in hardware applications is often avoided. Other studies showing different cardiopathologies can also be considered in the future [12]. The main problem of machine learning methods that are data-driven remains the availability of datasets and different conditions of acquisition between different available datasets.

Convolution neural networks (CNN) are mainly intended for visual imagery analysis and computer vision tasks. However, with the great success these networks have shown in classification tasks, they started to be involved in automated ECG analysis as well. Nevertheless, the performance of convolutional networks could degrade due to the impurity of data as well as the imbalance in the number of examples between classes. For that, it is often required to utilize some effective data augmentation techniques with the raw data before the recruitment of robust-to-data convolutional network models. In this paper, we present an optimal architecture for sequential data processing based on 1D Temporal Convolutional Networks (TCN). A database of five classes named ECG5000 [13], originally established from the BIDMC Congestive Heart Failure Database [14], is considered in our study. The ECG database signals used are noise-free clean, which makes them ready for use without any preprocessing. Each sample includes a single heartbeat. The ECG5000 database is enhanced by three data augmentation techniques for better performance of the network. The network involves various diluted causal one-dimensional convolutions with padding. As a result, the output signal is the same length as the input heartbeat. The convolution layers are followed by a softmax unit that matches the heartbeats with their classes. Accordingly, the network is evaluated. Due to the unique internal design of TCNs, these networks are lighter in weight, faster, and more stable than conventional convolutional networks. This allows the implementation of efficient embedded systems suited to remote health monitoring systems. ECG analysis devices are thus realized using low complexity and power consumption hardware.

The rest of the paper is organized as follows: in Section 2, we provide a background on the machine learning methods used in ECG analysis. In Section 3, we present in detail our proposed temporal convolution network architecture; next we demonstrate the data augmentation techniques applied to the ECG data and designate the values used in the training process; and lastly we explain the standard used in the evaluation of the network. In Section 4, we display the results of the multiple trials that we carried out and compare our optimized model to some existing machine learning networks implemented on the same dataset. The whole paper is summarized and concluded in Section 5.

2. Background and Related Work

A large part of the world’s population resides in a spread-out remote or rural area. In these rural and remote areas, besides other basic needs of life, the overall access to medical facilities ranges from difficult to deficient, and the availability of doctors is scarce. The deficiency of timely medical advice and assistance to the patients, due to distance and lack of adequate infrastructure, is the source of critical situations and may lead to death in some cases. Remote health care is considered a serious candidate for facilitating access to health services for all. Sensing and actuating technologies along with big data analysis provide basic building blocks for remote health monitoring (RHM). The concept of RHM is not new, but newer and efficient systems are still being designed to overcome the weaknesses of existing systems, especially for rural areas. Indeed, in rural areas, the main challenges are related to communication latency and bandwidth availability, autonomy and energy consumption, and low cost devices.

In the context of SAFE-RH (Sensing, ArtiFicial intelligence, and Edge networking towards Rural Health monitoring), a framework is proposed to cope with the above-mentioned problems by sending (and thus recording permanently) only the relevant data. Indeed, these relevant data to send are related to generation of alarms identified mostly by AI or machine learning (ML) driven methods, and thus significantly limit the bandwidth cost and communication overhead. Figure 1 shows the overall architecture of the proposed RHM system where ECG related flow is depicted in red.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g001.jpg

Overall framework of remote health monitoring in rural areas. In red is the flow related to ECG data. The classification is done by embedding intelligence near sensors and sending only alarms. If necessary, raw data can be sent to fog or cloud for further analysis or storage.

The study of ECG signals has become an essential tool in the clinical diagnosis of various heart diseases. This study is mainly based on the detailed characteristics of the ECG signal. In detail, an ECG signal is composed of numerous heartbeats connected together. Each heartbeat consists of different parts, namely: P wave, QRS complex, and T wave (see Figure 2). A normal heartbeat is characterized by given amplitude values for its peaks (P, Q, R, S, T, and U), as well as given duration values for its intervals (PR, RR, QRS, ST, and QT) and segments (PR and ST). The variation of any of these values indicates a certain abnormality at the diagnosis level. More details of an ECG beat can be found in [15]. These peaks, intervals, and segments are called the ECG features, on which the ECG classification is mainly based. The ECG classification problem is often a multi-class classification problem. It includes several classes not limited to: normal (N), right bundle branch block (RBBB), and left bundle branch block (LBBB). An ECG classification process involves multiple steps starting from signal preprocessing, feature extraction, then normalization, and ending with classification. In the first phase, signals are filtered to remove any kind of possible noise that could affect the extraction of the features. This includes powerline interference [16], EMG noise [17], baseline wander [18], and electrode motion artifacts [19]. Various techniques are proposed for noise removal such as low and high pass linear phase filters. For baseline adjustment, techniques such as linear phase high pass filter, median filter, and mean median filter are usually employed. In the second phase, the main features are collected to be used as inputs to a classification model. The commonly used techniques for this purpose are: Continuous Wavelet Transform (CWT) [20], Discrete Wavelet Transform (DWT) [21], Discrete Fourier Transform (DFT) [22], Discrete Cosine Transform (DCT) [23], S-Transform (ST) [24], Principal Component Analysis (PCA) [25], Pan–Tompkins Algorithm [26], Daubechies Wavelet (Db4) [27], and Independent Component Analysis (ICA) [28]. For the normalization of the features, two main approaches are commonly used: Z-score [29] and Unity Standard Deviation (SD) [30]. Finally, in the classification stage, different models are utilized such as: Multilayer Perceptron Neural Network (MLPNN) [31], Quantum Neural Network (QNN) [32], Radial Basis Function Neural Network (RBFNN) [33], Fuzzy C-Means Clustering (FCM) [34], ID3 Decision Tree [35], Support Vector Machine (SVM) [36], Type2 Fuzzy Clustering Neural Network (T2FCNN) [37], and Probabilistic Neural Network (PNN) [38].

Different ECG classifications have been recorded. In [39], the authors established a four-class ECG classification problem using the RR intervals as inputs. The data were collected from the MIT-BIH arrhythmia database. The raw signals were first subjected to baseline adjustment. After that, the RR intervals were extracted using DWT and then normalized using Z-score. The classification was done using FCM with an accuracy of 99.05%. No other metrics were calculated. The same authors considered in [40] another four-class ECG classification problem using the same database but with different classes. The features used were the RR intervals and the R point location. The feature extraction was done using DWT with Daubechies wavelet of order 3. The outputs were classified at two stages: preclassification using FCM and final classification using a three-layer MLPNN. The final accuracy was up to 99.99%. A two-class problem was demonstrated in [41] using the MIT-BIH arrhythmia database. The RR interval and R location were extracted using Db4 discrete wavelet transform. An FFNN, trained with back propagation algorithm, was used as a classifier. The final results showed an accuracy of 95%, a sensitivity of 90%, and a specificity of 90%. Another two-class ECG classification problem was investigated in [42]. Data were collected from the two databases: MIT-BIH arrhythmia and normal sinus rhythm. Noise was removed by band pass filter. The features, RR interval and R peak, were first extracted using DWT and then normalized by zero mean. An FFNN, with error back propagation algorithm, was used for classification. The performance of the model was estimated by the calculation of the classification accuracy (96.77%) and Youden index (0.9415). In [43], the authors constructed an ECG classification model for six classes. Data were collected from the MIT-BIH arrhythmia database. The utilized features were QRSh (QRS complex height), QRS width, R peak, RRt interval (current RR interval at time t), and RRt+1 interval (next RR interval at time t+1). The Pan–Tompkins algorithm was used in feature extraction. A low pass linear phase filter was built for noise removal, whereas a median filter was built for baseline correction. Outputs were classified using a particle swarm optimization (PSO) RBFNN. The sensitivity and specificity of the model were 96.251% and 99.104%, respectively. In [44], an ECG classifier was built based on the MIT-BIH arrhythmia database for the classification of eight heart diseases. The involved features, R peak, QRS segment, and RR interval were normalized before being fed into the classification model consisting of a PNN (radial basis layer and competitive layer) and a three-layer FFNN with back propagation algorithm, using zero mean and unity standard deviation. The model evaluation showed a sensitivity of 98.508%, a specificity of 99.906%, and an overall accuracy of 98.710%. Another ECG classifier was investigated in [45] using the MIT-BIH arrhythmia database. R peak and RR interval were extracted by the use of DWT, whereas the classification process was done by an MLPNN. The model was shown to be reliable with a mean square error of 0.00621.

Convolution neural networks (CNN) have also been involved in ECG analysis. In [46], a 2D CNN approach for ECG classification is investigated. The sequential vectors representing the heartbeats are transformed into binary images via one-hot encoding [47] before being introduced to the network. The morphology of the heartbeats as well as the temporal relationship between every two adjacent heartbeats is captured in such images. The learning process is accelerated using ADADELTA [48], a per-dimension learning rate method for gradient descent. The network also involves a biased dropout [49] to mitigate the overfitting of the network. The network, when tested on the MIT-BIH arrhythmia database, has shown to be highly effective in the detection of various cardiovascular diseases. Another work is investigated in [50]. A 1D CNN approach for arrhythmia detection is proposed. The sequential data are extracted using two leads and then injected directly into the network without any preprocessing. Although the network achieves high accuracy when tested on the MIT-BIH database, some classes are hardly recognized. This can be the result of the impurity of the data as well as the imbalance between classes.

3. Method

3.1. Proposed Architecture

Until recently, sequential data was mostly analyzed and modelled using recurrence-dependent networks such as recurrent neural networks and LSTM architectures. However, the most problematic issue that arises in training such networks is the vanishing or exploding of gradients. In other words, the network is often incapable of learning its weights from long-past values. For that, convolutional neural networks, widely used for computer vision and visual imagery tasks, are currently used for signal processing as well, under the name 1D Temporal Convolutional Network (1D TCN). The convolution in a TCN is uni-dimensional, causal, and dilated. A causal convolution means that the computation at a given unit of the network only depends on present and past values; this suits sequential data where each point of a sequence depends on previous ones. On the other side, dilation is set to increase the sparsity of a kernel so that the receptive field of the convolution layer can be enlarged without using additional parameters. Note that a receptive field is the region in the input that produces a feature at the output. The receptive field (R) of a dilated convolution layer with factor d is R=d(k1).

To build up a TCN, multiple convolution layers are stacked above each other as shown in Figure 3. The dilation factor of layer i1 is defined as di=2i1; this leads to an exponential growth of the receptive field size. Finally, with the scheme shown above, the receptive field of a TCN of l layers and convolution kernels of size k is defined as:

R=2l(k1).
(1)

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g003.jpg

Dilated causal convolution in a TCN of four layers.

The convolution in a TCN layer is defined as follows:

F(x(t))=(xdf)(t)=j=0kf(j)x(td·j)
(2)

where x is the input sequence, d is the dilation factor, and f is a convolution filter of size k applied at time t. It should be noted that an input sequence of length n is introduced to the above vanilla 1D convolution layer in order to generate an output sequence of length nk+1. A zero padding of length k1 is often applied at the beginning of the sequence so that the length of the sequence is preserved.

The full TCN model that we propose in this work is demonstrated in Figure 4. It is made up of multiple residual blocks, followed by a fully connected layer (FC), a softmax function, and a classifier. Each block contains a group of layers and a skip connection that links its input to its output. A 1×1 convolution is set on the skip connection of the first block in case the input and output mismatch in size. The skip connection is mainly used to revive gradients so that they can flow from one block to another without passing through non-linear activation functions. This, along with the dilation property of the utilized convolution, helps in solving the gradients’ exploding/fading issues. Within a residual block, the layers can be described as two sets connected in series where each set is composed of a dilated causal convolution layer, followed by layer normalization and spatial dropout layers. After the second normalization layer, a rectified linear unit (ReLU) is added. Note that the two convolution layers are identical (same filters and dilation factor) in one block, whereas they differ from one block to another. A normalization layer is added for faster and better performance of the TCN. It helps improve the stability of the network in case the weight’s initialization or the used regularization techniques are not helpful. This occurs by normalizing each of the inputs in the batch independently across all channels (features) using the mean and variance values [51]. This is different from batch normalization, mainly used in conventional convolutional networks, where normalization takes place at the level of batches in each channel separately. Layer normalization is well suited for sequence data where, unlike batch normalization, the batch size has no role. This allows the processing of large input sequences using any batch size for data division. Moreover, with layer normalization, parallelization is easily implemented with no need to have communication and synchronization between the different computing engines, as each one is computing separate data. On the other hand, in the dropout layer, some of the neurons are randomly deactivated during the network training. This aims to provide different forms of the network while training to avoid overfitting afterwards. Finally, the rectified linear unit (ReLU) performs a threshold operation on its input, where any element of a value less than zero is set to zero. This is done to achieve a non-linear transformation of the data so that they can be linearly separable during classification. The used values of the model hyperparameters can all be found in Section 4.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g004.jpg

TCN architecture for ECG heartbeat classification. The input signal is made of 140 points (0.56 s). Number of filters is FT. Filter length is KT. Dilation factors of blocks are 1, 2, 4, and 8, respectively.

Temporal convolution networks have been shown to outperform recurrent neural networks [52]. In terms of memory, the sparse kernels in TCNs allow the prediction of a time series from their long-past values using a very low number of parameters. In recurrent networks, this is done by the use of cycles and condensed recurrent connections, resulting in a large number of parameters. Moreover, the internal structure of a TCN is independent of the input signal. This allows the processing of excessively long sequences using a small TCN structures. Additionally, the receptive field size in such networks is easily tuned by modifying the number of layers, the filter size, and the dilation factors; this facilitates control of the model’s memory for various requirements. In terms of performance, temporal convolution networks are much faster than recurrent networks. This is due the fact that these networks compute their outputs in parallel. On the other side, the structure of a TCN leads to more stable gradients where these gradients vary in the direction of the layers not in the temporal direction (also thanks to the residual connections). Temporal convolution networks are not without flaws. Indeed, these networks do not function well in the case of domain transfer especially from a domain that requires a short history to another that requires a long one.

3.2. Data Augmentation

In terms of classification problems, the lack of sufficient training samples of certain classes is often fixed by using data augmentation techniques. This occurs by adding new copies of the existing samples of the deficient classes after applying certain minor alterations or using machine learning models to generate new examples in the latent space of the original data. Various techniques can be found in the literature for the augmentation of image data such as rotation, flipping, cropping, and color transformation [53]. Such mechanisms preserve the main features of an image while providing a bigger space for training. This seems more complicated in the case of one-dimensional data. In this work, we propose three simple types of data augmentation that can be applied to the ECG signals: amplitude shifting, time shifting, and amplification. Amplitude shifting, as shown in Figure 5a, means moving the signal a certain number of steps either upward or downward. Time shifting is presented in Figure 5b; it is about moving the signal a given number of steps either to the right or to the left. Amplification, as shown in Figure 5c, is done by vertically extending the signal by a certain ratio. Like conventional 2D data augmentation techniques, these proposed techniques, along with many others, provide additional data for use in training while preserving the main features of a signal.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g005.jpg

Data augmentation techniques applied to the classes PVC, SP, and UB of the ECG5000 dataset.

3.3. Training Process

The general TCN model, proposed in Section 3.1 and shown in Figure 4, is trained using Adam optimizer on batches of size 20 at a learning rate of 0.0025. Different trials are carried out, as will be shown in Section 4, by changing the number of blocks, the number of filters, the filters’ size, and the applied data augmentation techniques. The training process lasts for 250 epochs. The dropout ratio of all dropout layers is 0.3. The raw dataset (before augmentation) is divided into two sets: training dataset and testing dataset, with a ratio of 90:10. The training dataset is shuffled once before use, and no cross-validation is applied. On the other hand, the evaluation of the network is done on the testing dataset at the end of every epoch. The final accuracy is chosen at the epoch where the parameters of the network produce the minimum loss.

3.4. Network Evaluation

Classification model performance is mostly evaluated based on the “confusion matrix” [54]. This is a quite common measure that can be applied to both binary and multiclass classification problems where the counts of predicted and actual outcomes are all represented. For each class, four quantities can be identified: TP, FP, TN, and FN. The term “TP” denotes True Positive, which represents the number of positive examples that are correctly classified by the model. Similarly, the term “TN” stands for True Negative, which represents the number of negative examples that are correctly classified. The term “FP” stands for False Positive, which is the number of negative examples classified by the model as positive; whereas the term “FN” denotes False Negative, i.e., the number of positive examples classified as negative. The most commonly used criterion in the evaluation of a classification model is accuracy, which presents the fraction of true examples over all examples:

Accuracy=AllcorrectAll=TP+TNTP+TN+FP+FN.
(3)

However, the accuracy defined in Equation (3) can be misleading when the datasets used for training and test purposes are imbalanced. For that, there are other metrics that could be involved in the evaluation process for better analysis: precision and sensitivity. Precision is the proportion of correctly classified positive cases, i.e., the fraction of positive examples over the total predicted positive examples. On other hand, sensitivity is the proportion of correctly recognized actual positive cases, i.e., the fraction of positive instances over the total actual positive instances. The formulas are defined below:

Precision=TruePositivesPredictedPositives=TPTP+FP;Sensitivity=TruePositivesAllActualPositives=TPTP+FN.

It should be noted that the two quantities above are computed for each class separately, and therefore the overall quantities are deduced by averaging. For the sake of brevity, precision and sensitivity can be combined into one term, the F-score, as follows:

Fβ=(1+β2)Sensitivity×Precision(β2·Sensitivity)+Precision.
(4)

The balanced F-score is the harmonic mean of precision and recall; that is, the F1-score (β=1):

F1=2·Sensitivity×PrecisionSensitivity+Precision.
(5)

It is always preferable to achieve a good performance of the TCN model using the most concise network structure. For that, the number of parameters that comprise the TCN is another criterion to be considered in the evaluation process. A TCN, as shown in Figure 4, is made up of multiple layers, where each distinct layer involves a different number of parameters. The only layers that have no learnable parameters are the input layer, the ReLU layer, and the dropout layer. The input layer only provides the shape of the input signal and has nothing to do with the training process. The ReLU performs a threshold operation on its input x according to the fixed equation f(x)=max(0,x). Similarly, this layer is not changed during the training phase. The dropout layer is set to reduce the number of activated neurons in the training phase where elimination takes place in a merely random way. The layers that include learnable parameters are the convolution layer and the normalization layer. Each convolution layer has two types of parameters: weights and biases. The weights are learned during the training process. They are matrices that affect the prediction ability of the model, which is altered during the back-propagation process based on the used optimization strategy. The biases are set to delay or accelerate the activation of nodes. The total number of parameters in a convolution layer is the sum of all present weights and biases. Knowing that the size of a filter of a given convolution layer is denoted by KT, the number of filters of this layer is denoted by FT, and the number of filters of the previous layer is denoted by FTp, then the total number of parameters of this convolution layer (Pconv) is defined as follows:

Pconv=W+B=(KT×FTp×FT)+FT
(6)

where W and B are the number of weights and the number of biases of the convolution layer, respectively. Note that the dilation, stride, and padding are hyperparameters that do not interfere in the learning process [55,56]. Similarly, a normalization layer also has two learnable parameters of its own: offset (also called beta) and scale (gamma). Each channel of a normalization layer has one parameter of each kind. Attached to the convolution layer, the normalization layer thus has FT betas and FT gammas. Hence, the total number of parameters in a normalization layer is:

Pnorm=2×FT.
(7)

4. Results and Discussion

The TCN model that we proposed in Section 3.1 is tested on the ECG5000 dataset [13] that has been collected from the BIDMC Congestive Heart Failure Database [14]. The raw record is composed of 17,998,834 data points including 92,584 heartbeats. The heartbeats are first extracted and then interpolated so that they all become the same length. After that, the heartbeats are annotated according to five classes holding the following labels: Normal (N), R-on-T Premature Ventricular Contraction (Ron-T PVC), Premature Ventricular Contraction (PVC), Supraventricular Premature or Ectopic Beat (SP or EB), and Unclassified Beat (UB). About 128,570 annotated heartbeats are present from which a dataset of 5000 randomly selected heartbeats is created. The new dataset is then divided into two sets: one for training that contains 4500 samples and another for testing that contains 500 samples.

On the other hand, different forms of the TCN model are built and experimentally tested. All details are summarized in Table 1, where 12 different experiments are demonstrated. The number of blocks varies between 3 and 5. The size of the filters of the dilated convolution layers is constant in one block whereas it grows from one block to another. However, the number of filters is always the same. Consequently, the number of parameters of the TCN model is calculated (see Section 3.4 for more details). In all experiments, the mini-batch has a size of 20, whereas the training process lasts for 250 epochs. The dilation factors (di) of the five blocks (if they exist) are 1, 2, 4, 8, and 16, respectively. For data augmentation, different factors are tried out on the three classes: PVC, SP, and UB. In all experiments, the classes PVC and UB (classes 3 and 5, respectively) are augmented 24 times. On the other hand, class SP is augmented 12 times in Experiments 1–10, 24 times in Experiment 11, and 6 times in Experiment 12. Based on the available training data and the size of the training batch, the size of one epoch is determined. The best results are obtained for Experiments 1 and 5. In Experiment 1, the TCN model involves about 10,200 parameters and achieves an F1 score of 84.13% and an accuracy of 96.12%. The TCN model in Experiment 5 involves about 39,900 parameters and achieves an F1 score of 85.43% and an accuracy of 96.62%.

Table 1

Different forms of the TCN model proposed in Section 3.1 with different data augmentation cases. #Epochs=250 and Mini-BatchSize=20.

Exp#Blocks#FiltersFilter Size#ParmsData Augmentation FactorEpoch SizeF1 ScoreAccuracy
#1 416 {2,4,6,8} 10.2 Kclass3:24, class4:12, class5:24460 84.13% 96.12%
#2 316 {2,4,6} 6 Kclass3:24, class4:12, class5:24460 68.54% 94.80%
#3 516 {2,4,6} 15.4 Kclass3:24, class4:12, class5:24460 83.76% 96.26%
#4 48 {2,4,6,8} 2.7 Kclass3:24, class4:12, class5:24460 73.40% 95.20%
#5 432 {2,4,6,8} 39.9 Kclass3:24, class4:12, class5:24460 85.43% 96.62%
#6 416 {4,6,8,12} 14.9 Kclass3:24, class4:12, class5:24460 81.04% 95.87%
#7 432 {4,6,8,12} 58.4 Kclass3:24, class4:12, class5:24460 75.60% 96.07%
#8 316 {4,6,8,12} 8.6 Kclass3:24, class4:12, class5:24460 78.32% 95.57%
#9 316 {6,8,12,14} 12.2 Kclass3:24, class4:12, class5:24460 73.57% 95.70%
#10 332 {2,4,6} 23.3 Kclass3:24, class4:12, class5:24460 73.80% 96.00%
#11 416 {2,4,6,8} 10.2 Kclass3:24, class4:24, class5:24566 80.85% 96.35%
#12 416 {2,4,6,8} 10.2 Kclass3:24, class4:06, class5:24407 79.70% 96.02%

The TCN model in Experiment 1 is the most optimal among all models present in Table 1. In detail, it has convolution layers that involve 16 filters (FT=16). The size of the filters augments from one block to another as KT{2,4,6,8}. The training data of classes PVC and UB are augmented 24 times: 4 times by amplitude shifting, 12 times by time shifting, and 8 times by amplification; whereas the training data of class SP are augmented only 12 times: 2 times by amplitude shifting, 6 times by time shifting, and 4 times by amplification. More details are found in Table 2 and Table 3. After data augmentation, the number of samples of the training data set grows from 4500 to 9204. The mini-batch has a size of 20, which leads to the distribution of the samples on an epoch of 460 iterations. With 250 epochs, the training process ends after 115,000 iterations. The evolution of the training and testing accuracy is shown in Figure 6. As shown, the model gains its optimal parameters after nearly 75 epochs with a testing accuracy of 96.12%. A more detailed evaluation of Experiment 1 can be done based on the confusion matrix shown in Figure 7. The sensitivity percentages of the classes are presented to the right of the 5×5 confusion matrix, whereas the precision percentages are placed below it (all in blue). The classification results for Classes 1–3 are very competitive where the sensitivity and precision percentages are high. By contrast, the classification is misleading for Class 4 with a sensitivity of 42.1%, and for Class 5 with a precision of 50%. In the latter two classes, the applied data augmentation techniques improve the classification results to a limited range. We should mention here that without data augmentation, the sensitivity of Class 4 is 31.5%, whereas the precision of Class 5 is 25%. Apart from data augmentation, multiple techniques could be employed to handle imbalanced data such as K-fold cross-validation, the use of specialized models like XGBoost, and the aggregation of more raw data. The receiver operating characteristic (ROC) curves of the five classes are shown in Figure 8, which demonstrates an excellent classification of the proposed model for four classes (Class 1, Class 2, Class 3, Class 5) and satisfactory classification results for Class 4.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g006.jpg

Training and validation process of TCN model in Experiment 1 of Table 4.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g007.jpg

Confusion matrix of the TCN model proposed in Experiment 1 of Table 1, applied to the ECG5000 dataset.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g008.jpg

ROC curves of the trained model with marked operating points (dots) obtained for the five-class use case (Experiment #1, see Table 1).

Table 2

Data augmentation applied on ECG5000 dataset in Experiments 1–10 of Table 1.

ClassData Augmentation Type
Amplitude ShiftTime ShiftAmplification
PVC [0.4,0.4]0.2 [6,6]1 [1.05,1.4]0.05
SP [0.4,0.4]0.4 [5,5]1 [1.05,1.2]0.05
UB [0.4,0.4]0.2 [6,6]1 [1.05,1.4]0.05

Table 3

Distribution of samples of ECG5000 dataset in Experiments 1–10 of Table 1, before and after data augmentation.

ClassCount
Testing SetBefore AugmentationAfter Augmentation
Training SetTotalTraining SetTotal
N2922627291926272919
Ron-T PVC1771590176715901767
PVC10869621502160
SP1917519422872306
UB20224550552
Total5004500500092049704

The ECG5000 dataset has been adopted in various classification works for heart disease detection. The comparison between our optimal model and some state-of-the-art models is demonstrated in Table 4. We show in this table the accuracy, the F1 score (combination between sensitivity and precision), and the number of parameters of the different models. As shown, our TCN model with the fewest number of parameters among all other existing models (10.2 K) attains the highest accuracy among all of these models (96.12%). On the other hand, our TCN model affords an F1 score of 84.13%, which is lower only than that of the TCN model proposed in [57]. The effect of the applied data augmentation is also investigated. The accuracy is only 93.4% and the F1 score is down to 70.18% for the same TCN model when applied to the raw data.

Table 4

Evaluation of different methods on the ECG5000 dataset.

ArchitectureAccuracy (%)F1 Score (%)#Parameters
TCN [57]94.289.014.88 K
LSTM-FCN [58]94.172.5404.74 K
CCN [58]93.481.5266.37 K
LSTM [58]93.168.9138.37 K
1-NN (L2 dist.) [59]92.554.970 K
Our TCN96.1284.1310.2 K
Our TCN (First 3 classes)98.5494.5110.2 K
Our TCN (Without Data Augmentation)93.470.1810.2 K

As noticed in the confusion matrix of Figure 7, the slight deficiency in the F1 score is mainly due to classes SP and UB (see orange boxes). For that, we retrain the network while excluding these classes. As shown in Table 4, the accuracy now is up to 98.54%, whereas the F1 score reaches 94.51%. The detailed results are present in the confusion matrix of Figure 9. The receiver operating characteristic (ROC) curves of the three classes are shown in Figure 10 with an AUC1 demonstrating excellent classification results for this use case.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g009.jpg

Confusion matrix of the TCN model proposed in Experiment 1 of Table 1, applied to the ECG5000 dataset excluding classes SP and UB.

An external file that holds a picture, illustration, etc.
Object name is sensors-23-01697-g010.jpg

ROC curves of the trained model corresponding to Experiment 1 of Table 1, where the network is trained only on the first three classes.

5. Conclusions

In this paper, we proposed a 1D Temporal Convolutional Network (TCN) based architecture for ECG classification of five heart diseases. The main goal was to provide a low complexity architecture aimed at being used in low-cost embedded devices for remote health monitoring. The proposed architecture is characterized by its simplicity and the lowest number of used parameters compared to the state-of-the-art approaches (10.2 K parameters < 15 K in the literature). Moreover, the proposed architecture is outperforming all existing state-of-the-art ML methods in terms of overall accuracy and F1 score, reaching up to 96.12% and 84.13%, respectively. The ROC curves of the proposed model show excellent classification performances with an average AUC1. As future work and perspectives, an extensive study to interpret and explain the obtained results will be conducted, along with the study of quantization and pruning of the network’s parameters and their influence on the overall accuracy of the proposed architecture as well as the hardware resources used and power consumption needed. Both microcontroller-based and circuit-specific implementations will be targeted with a comparison in terms of power consumption, ease of implementation and use, and overall cost.

Funding Statement

This research was funded by the European Erasmus+ capacity building for higher education program under grant agreement no. 619483.

Author Contributions

Conceptualization and methodology A.R.I., S.J. and H.R.; software, A.R.I.; validation, A.R.I., S.J. and H.R.; investigation, A.R.I., S.J. and H.R.; resources, H.R. and N.R.; writing—original draft preparation, A.R.I., S.J. and H.R.; writing—review and editing, S.J., A.R.I., N.R. and H.R.; visualization, H.R., S.J., N.R. and A.R.I.; supervision, H.R. and S.J.; project administration, H.R.; funding acquisition, H.R. and N.R. All authors have read and agreed to the published version of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

1. Wu M., Lu Y., Yang W., Wong S.Y. A Study on Arrhythmia via ECG Signal Classification Using the Convolutional Neural Network. Front. Comput. Neurosci. 2021;14:564015. 10.3389/fncom.2020.564015. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
2. Cook D., Oh S.Y., Pusic M. Accuracy of Physicians’ Electrocardiogram Interpretations: A Systematic Review and Meta-analysis. JAMA Intern. Med. 2020;180:1461–1471. 10.1001/jamainternmed.2020.3989. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
3. Schläpfer J., Wellens H.J. Computer-Interpreted Electrocardiograms: Benefits and Limitations. J. Am. Coll. Cardiol. 2017;70:1183–1192. 10.1016/j.jacc.2017.07.723. [Abstract] [CrossRef] [Google Scholar]
4. Ghafoor M.J., Ahmed S., Riaz K. Exploiting Cross-Correlation Between ECG signals to Detect Myocardial Infarction; Proceedings of the 2020 17th International Bhurban Conference on Applied Sciences and Technology (IBCAST); Islamabad, Pakistan. 14–18 January 2020; pp. 321–325. [CrossRef] [Google Scholar]
5. Ayar M., Sabamoniri S. An ECG-based feature selection and heartbeat classification model using a hybrid heuristic algorithm. Inform. Med. Unlocked. 2018;13:167–175. 10.1016/j.imu.2018.06.002. [CrossRef] [Google Scholar]
6. Faust O., Kareem M., Ali A., Ciaccio E.J., Acharya U.R. Automated Arrhythmia Detection Based on RR Intervals. Diagnostics. 2021;11:1446. 10.3390/diagnostics11081446. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
7. El-Khafif S.H., El-Brawany M.A. Artificial Neural Network-Based Automated ECG Signal Classifier. Int. Sch. Res. Not. 2013;2013:1–6. 10.1155/2013/261917. [CrossRef] [Google Scholar]
8. Ribeiro A.H., Ribeiro M.H., Paixão G.M.M., Oliveira D.M., Gomes P.R., Canazart J.A., Ferreira M.P.S., Andersson C.R., Macfarlane P.W., Meira W., et al. Automatic diagnosis of the 12-lead ECG using a deep neural network. Nat. Commun. 2020;11:1760. 10.1038/s41467-020-15432-4. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
9. Kim J.H., Seo S.Y., Song C.G., Kim K.S. Assessment of Electrocardiogram Rhythms by GoogLeNet Deep Neural Network Architecture. J. Healthc. Eng. 2019;2019:2826901. 10.1155/2019/2826901. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
10. Jing E., Zhang H., Li Z., Liu Y., Ji Z., Ganchev I. ECG Heartbeat Classification Based on an Improved ResNet-18 Model. Comput. Math. Methods Med. 2021;2021:6649970. 10.1155/2021/6649970. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
11. Saadatnejad S., Oveisi M., Hashemi M. LSTM-Based ECG Classification for Continuous Monitoring on Personal Wearable Devices. IEEE J. Biomed. Health Inform. 2019;24:515–523. 10.1109/JBHI.2019.2911367. [Abstract] [CrossRef] [Google Scholar]
12. Hussain I., Park S.J. Big-ECG: Cardiographic Predictive Cyber-Physical System for Stroke Management. IEEE Access. 2021;9:123146. 10.1109/ACCESS.2021.3109806. [CrossRef] [Google Scholar]
13. Chen Y., Keogh E. Time Series Classification. [(accessed on 8 April 2022)]. Available online: http://www.timeseriesclassification.com/description.php?Dataset=ECG5000.
14. Baim D.S., Colucci W.S., Monrad E.S., Smith H.S., Wright R.F., Lanoue A., Gauthier D.F., Ransil B.J., Grossman W., Braunwald E. Survival of patients with severe congestive heart failure treated with oral milrinone. J. Am. Coll. Cardiol. 1986;7:661–670. 10.1016/S0735-1097(86)80478-8. [Abstract] [CrossRef] [Google Scholar]
15. Wikipedia Contributors Electrocardiography—Wikipedia, The Free Encyclopedia. 2022. [(accessed on 9 November 2022)]. Available online: https://en.wikipedia.org/w/index.php?title=Electrocardiography&oldid=1120710988.
16. Chen B., Li Y., Cao X., Sun W., He W. Removal of Power Line Interference From ECG Signals Using Adaptive Notch Filters of Sharp Resolution. IEEE Access. 2019;7:150667. 10.1109/ACCESS.2019.2944027. [CrossRef] [Google Scholar]
17. Bhaskar P.C., Uplane M.D. High Frequency Electromyogram Noise Removal from Electrocardiogram Using FIR Low Pass Filter Based on FPGA. Procedia Technol. 2016;25:497–504. 10.1016/j.protcy.2016.08.137. [CrossRef] [Google Scholar]
18. Romero F.P., Romaguera L.V., V’azquez-Seisdedos C.R., Filho C.F.F.C., Costa M.G.F., Neto J.E. Baseline wander removal methods for ECG signals: A comparative study. arXiv. 20181807.11359 [Google Scholar]
19. Subramaniam S.R., Ling B.W.K., Georgakis A. Motion artifact suppression in the ECG signal by successive modifications in frequency and time; Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); Osaka, Japan. 3–7 July 2013; pp. 425–428. [Abstract] [CrossRef] [Google Scholar]
20. Shi Y., Ruan Q. Continuous wavelet transforms; Proceedings of the 7th International Conference on Signal Processing; Beijing, China. 31 August–4 September 2004; pp. 207–210. [CrossRef] [Google Scholar]
21. Alessio S. Digital Signal Processing and Spectral Analysis for Scientists. Springer; Cham, Switzerland: 2016. Discrete Wavelet Transform (DWT) pp. 645–714. [CrossRef] [Google Scholar]
22. Thyagarajan K. Introduction to Digital Signal Processing Using MATLAB with Application to Digital Communications. Springer; Cham, Switzerland: 2019. Discrete Fourier Transform; pp. 151–188. [CrossRef] [Google Scholar]
23. Burger W., Burge M. Digital Image Processing: An Algorithmic Introduction Using Java. Springer; Berlin/Heidelberg, Germany: 2016. The Discrete Cosine Transform (DCT) pp. 503–511. [CrossRef] [Google Scholar]
24. Wang Y.H. The Tutorial: S Transform. National Taiwan University; Taipei, Taiwan: 2006. [Google Scholar]
25. Mishra S., Sarkar U., Taraphder S., Datta S., Swain D., Saikhom R., Panda S., Laishram M. Principal Component Analysis. Int. J. Livest. Res. 2017;7:60–78. 10.5455/ijlr.20170415115235. [CrossRef] [Google Scholar]
26. Wikipedia Contributors Pan–Tompkins Algorithm—Wikipedia, The Free Encyclopedia. 2022. [(accessed on 21 November 2022)]. Available online: https://encyclopedia.pub/entry/30809.
27. Wikipedia Contributors Daubechies Wavelet—Wikipedia, The Free Encyclopedia. 2022. [(accessed on 21 November 2022)]. Available online: https://en.wikipedia.org/wiki/Daubechies_wavelet.
28. Tharwat A. Independent Component Analysis: An Introduction. Appl. Comput. Inform. 2018;17:222–249. 10.1016/j.aci.2018.08.006. [CrossRef] [Google Scholar]
29. Z-Score. 2010. [(accessed on 21 November 2022)]. Available online: https://onlinelibrary.wiley.com/doi/10.1002/9780470479216.corpsy1047.
30. Lee D., In J., Lee S. Standard deviation and standard error of the mean. Korean J. Anesthesiol. 2015;68:220–223. 10.4097/kjae.2015.68.3.220. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
31. Popescu M.C., Balas V., Perescu-Popescu L., Mastorakis N. Multilayer perceptron and neural networks. WSEAS Trans. Circuits Syst. 2009;8:579–588. [Google Scholar]
32. Kwak Y., Yun W., Jung S., Kim J. Quantum Neural Networks: Concepts, Applications, and Challenges; Proceedings of the 2021 Twelfth International Conference on Ubiquitous and Future Networks (ICUFN); Jeju Island, Republic of Korea. 17–20 August 2021; pp. 413–416. [CrossRef] [Google Scholar]
33. Dash C., Behera A.K., Dehuri S., Cho S.B. Radial basis function neural networks: A topical state-of-the-art survey. Open Comput. Sci. 2016;6:33–63. 10.1515/comp-2016-0005. [CrossRef] [Google Scholar]
34. Hung M.C., Yang D.L. An efficient Fuzzy C-Means clustering algorithm; Proceedings of the 2001 IEEE International Conference on Data Mining; San Jose, CA, USA. 29 November–2 December 2001; pp. 225–232. [CrossRef] [Google Scholar]
35. Ogheneovo E., Nlerum P. Iterative Dichotomizer 3 (ID3) Decision Tree: A Machine Learning Algorithm for Data Classification and Predictive Analysis. Int. J. Adv. Eng. Res. Sci. 2020;7:514–521. 10.22161/ijaers.74.60. [CrossRef] [Google Scholar]
36. Pradhan A. Support vector machine—A survey. Int. J. Emerg. Technol. Adv. Eng. 2012;2:82–85. [Google Scholar]
37. Aliev R., Guirimov B. Type-2 Fuzzy Neural Networks and Their Applications. Springer; Berlin/Heidelberg, Germany: 2014. Type-2 Fuzzy Clustering; pp. 153–166. [CrossRef] [Google Scholar]
38. Zeinali Y., Story B. Competitive probabilistic neural network. Integr. Comput.-Aided Eng. 2017;24:105–118. 10.3233/ICA-170540. [CrossRef] [Google Scholar]
39. Dallali A., Kachouri A., Samet M. A Classification of Cardiac Arrhythmia Using WT, HRV, and Fuzzy C-Means Clustering. Signal Process. Int. J. (SPJI) 2011;5:101–108. [Google Scholar]
40. Dallali A., Kachouri A., Samet M. Fuzzy C-means clustering, neural network, WT, and HRV for classification of cardiac arrhythmia. J. Eng. Appl. Sci. 2011;6:112–118. [Google Scholar]
41. Khazaee A. Heart Beat Classification Using Particle Swarm Optimization. Int. J. Intell. Syst. Appl. 2013;5:25–33. 10.5815/ijisa.2013.06.03. [CrossRef] [Google Scholar]
42. Vishwa A., Lal M., Dixit S., Varadwaj P. Clasification Of Arrhythmic ECG Data Using Machine Learning Techniques. Int. J. Interact. Multimed. Artif. Intell. 2011;1:67–70. 10.9781/ijimai.2011.1411. [CrossRef] [Google Scholar]
43. Korürek M., Doğan B. ECG beat classification using particle swarm optimization and radial basis function neural network. Expert Syst. Appl. 2010;37:7563–7569. 10.1016/j.eswa.2010.04.087. [CrossRef] [Google Scholar]
44. Yu S.N., Chou K.T. Integration of independent component analysis and neural networks for ECG beat classification. Expert Syst. Appl. 2008;34:2841–2846. 10.1016/j.eswa.2007.05.006. [CrossRef] [Google Scholar]
45. Ayub S., Saini J. ECG classification and abnormality detection using cascade forward neural network. Int. J. Eng. Sci. Technol. 2011;3:41–46. 10.4314/ijest.v3i3.68420. [CrossRef] [Google Scholar]
46. Li J., Si Y., Xu T., Saibiao J. Deep Convolutional Neural Network Based ECG Classification System Using Information Fusion and One-Hot Encoding Techniques. Math. Probl. Eng. 2018;2018:1–10. 10.1155/2018/7354081. [CrossRef] [Google Scholar]
47. Wikipedia Contributors One-Hot—Wikipedia, The Free Encyclopedia. 2022. [(accessed on 14 November 2022)]. Available online: https://en.wikipedia.org/wiki/One-hot.
48. Zeiler M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv. 2012 10.48550/ARXIV.1212.5701.1212.5701 [CrossRef] [Google Scholar]
49. Russakovsky O., Deng J., Su H., Krause J., Satheesh S., Ma S., Huang Z., Karpathy A., Khosla A., Bernstein M., et al. ImageNet Large Scale Visual Recognition Challenge. arXiv. 2014 10.1007/s11263-015-0816-y.1409.0575 [CrossRef] [Google Scholar]
50. Ferretti J., Randazzo V., Cirrincione G., Pasero E. Progresses in Artificial Intelligence and Neural Systems. Springer; Berlin/Heidelberg, Germany: 2021. 1-D Convolutional Neural Network for ECG Arrhythmia Classification; pp. 269–279. [CrossRef] [Google Scholar]
51. Ba J., Kiros J., Hinton G. Layer Normalization. arXiv. 20161607.06450 [Google Scholar]
52. Bai S., Kolter J., Koltun V. An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling. arXiv. 20181803.01271 [Google Scholar]
53. Shorten C., Khoshgoftaar T. A survey on Image Data Augmentation for Deep Learning. J. Big Data. 2019;6:1–48. 10.1186/s40537-019-0197-0. [Europe PMC free article] [Abstract] [CrossRef] [Google Scholar]
54. Shultz T.R., Fahlman S.E., Craw S., Andritsos P., Tsaparas P., Silva R., Drummond C., Ling C.X., Sheng V.S., Drummond C., et al. Encyclopedia of Machine Learning. Springer; New York, NY, USA: 2011. Confusion Matrix; p. 209. [CrossRef] [Google Scholar]
55. Roy S., Rodrigues N., Taguchi Y.h. Incremental Dilations Using CNN for Brain Tumor Classification. Appl. Sci. 2020;10:4915. 10.3390/app10144915. [CrossRef] [Google Scholar]
56. Lee W.Y., Park S., Sim K.B. Optimal hyperparameter tuning of convolutional neural networks based on the parameter-setting-free harmony search algorithm. Optik. 2018;172:359–367. 10.1016/j.ijleo.2018.07.044. [CrossRef] [Google Scholar]
57. Ingolfsson T.M., Wang X., Hersche M., Burrello A., Cavigelli L., Benini L. ECG-TCN: Wearable Cardiac Arrhythmia Detection with a Temporal Convolutional Network; Proceedings of the 2021 IEEE 3rd International Conference on Artificial Intelligence Circuits and Systems (AICAS); Washington, DC, USA. 3–7 May 2021; pp. 1–4. [CrossRef] [Google Scholar]
58. Karim F., Majumdar S., Darabi H., Chen S. LSTM Fully Convolutional Networks for Time Series Classification. IEEE Access. 2018;6:1662–1669. 10.1109/ACCESS.2017.2779939. [CrossRef] [Google Scholar]
59. Chen Y., Keogh E., Hu B., Begum N., Bagnall A., Mueen A., Batista G. The UCR Time Series Classification Archive. 2015. [(accessed on 8 April 2002)]. Available online: https://www.cs.ucr.edu/~eamonn/time_series_data/

Articles from Sensors (Basel, Switzerland) are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

Citations & impact 


Impact metrics

Jump to Citations

Citations of article over time

Alternative metrics

Altmetric item for https://www.altmetric.com/details/143752356
Altmetric
Discover the attention surrounding your research
https://www.altmetric.com/details/143752356

Article citations

Funding 


Funders who supported this work.

European Erasmus+ capacity building for higher educa- 342 tion programme (1)

  • Grant ID: 619483

European Erasmus+ capacity building for higher education program (1)

  • Grant ID: 619483