NEURAL PROCESSING LETTERS

An Optimal Stacked ResNet-BiLSTM-Based Accurate Detection and Classification of Genetic Disorders
Nandhini K and Tamilpavai G
Gene is located inside the nuclease and the genetic data is contained in deoxyribonucleic acid (DNA). A person's gene count ranges from 20,000 to 30,000. Even a minor alteration to the DNA sequence can be harmful if it affects the cell's fundamental functions. As a result, the gene begins to act abnormally. The sorts of genetic abnormalities brought on by mutation include chromosomal disorders, complex disorders, and single-gene disorders. Therefore, a detailed diagnosis method is required. Thus, we proposed an Elephant Herd Optimization-Whale Optimization Algorithm (EHO-WOA) optimized Stacked ResNet-Bidirectional Long Term Short Memory (ResNet-BiLSTM) model for detecting genetic disorders. Here, a hybrid EHO-WOA algorithm is presented to assess the Stacked ResNet-BiLSTM architecture's fitness. The ResNet-BiLSTM design uses the genotype and gene expression phenotype as input data. Furthermore, the proposed method identifies rare genetic disorders such as Angelman Syndrome, Rett Syndrome, and Prader-Willi Syndrome. It demonstrates the effectiveness of the developed model with greater accuracy, recall, specificity, precision, and f1-score. Thus, a wide range of DNA deficiencies including Prader-Willi syndrome, Marfan syndrome, Early Onset Morbid Obesity, Rett syndrome, and Angelman syndrome are predicted accurately.
A Rumor Detection Model Incorporating Propagation Path Contextual Semantics and User Information
Bai L, Han X and Jia C
Currently, social media is full of rumors. To stop rumors from spreading further, rumor detection has received increasing attention. Recent rumor detection methods treat all propagation paths and all nodes on the paths as equally important, resulting in models that fail to extract the key features. In addition, most methods ignore user features, leading to limitations in the performance improvement of rumor detection. To address these problems, we propose a Dual-Attention Network model on propagation Tree structures named DAN-Tree, where a node-and-path dual-attention mechanism is designed to organically fuse deep structure and semantic information on the propagation structures of rumors, and path oversampling and structural embedding are employed to enhance the learning of deep structures. Finally, we deeply integrate user profiles into the propagation trees in DAN-Tree, thus proposing the DAN-Tree++ model to further improve performance. Empirical studies on four rumor datasets have shown that DAN-Tree outperforms the state-of-the-art rumor detection models learning on propagation structures, and the results on two datasets with user information validate the superior performance of DAN-Tree++ over other models using both user profiles and propagation structures. What's more, DAN-Tree, especially DAN-Tree++, has achieved the best performance on early detection tasks.
Differential Neural Networks Prediction Using Slow and Fast Hybrid Learning: Application to Prognosis of Infectionsand Deaths of COVID-19 Dynamics
Poznyak A, Chairez I and Anyutin A
This essay discusses a potential method for predicting the behavior of various physical processes and uses the COVID-19 outbreak to demonstrate its applicability. This study assumes that the current data set reflects the output of a dynamic system that is governed by a nonlinear ordinary differential equation. This dynamic system may be described by a Differential Neural Network (DNN) with time-varying weights matrix parameters. A new hybrid learning scheme based on the decomposition of the signal to be predicted. The decomposition considers the slow and fast components of the signal which is more natural to signals such as the ones corresponding to the number of infected and deceased patients who suffered of COVID 2019 sickness. The paper results demonstrate the recommended method offers competitive performance (70 days of COVID prediction) in comparison to similar studies.
Knowledge Fusion Distillation: Improving Distillation with Multi-scale Attention Mechanisms
Li L, Su W, Liu F, He M and Liang X
The success of deep learning has brought breakthroughs in many fields. However, the increased performance of deep learning models is often accompanied by an increase in their depth and width, which conflicts with the storage, energy consumption, and computational power of edge devices. Knowledge distillation, as an effective model compression method, can transfer knowledge from complex teacher models to student models. Self-distillation is a special type of knowledge distillation, which does not to require a pre-trained teacher model. However, existing self-distillation methods rarely consider how to effectively use the early features of the model. Furthermore, most self-distillation methods use features from the deepest layers of the network to guide the training of the branches of the network, which we find is not the optimal choice. In this paper, we found that the feature maps obtained by early feature fusion do not serve as a good teacher to guide their own training. Based on this, we propose a selective feature fusion module and further obtain a new self-distillation method, knowledge fusion distillation. Extensive experiments on three datasets have demonstrated that our method has comparable performance to state-of-the-art distillation methods. In addition, the performance of the network can be further enhanced when fused features are integrated into the network.
A Radial Basis Function Neural Network for Stochastic Frontier Analyses of General Multivariate Production and Cost Functions
Pendharkar PC
Production function techniques often impose functional form and other restrictions that limit their applicability. One common limitation in popular production function techniques is the requirement that all inputs and outputs must be positive numbers. There is a need to develop a production function analysis technique that is less restrictive in the assumptions it makes, and inputs it can process. This paper proposes such a general technique by linking fields of neural networks and econometrics. Specifically, two radial basis function (RBF) neural networks are proposed for stochastic production and cost frontier analyses. The functional forms of production and cost functions are considered unknown except that they are multivariate. Using simulated and real-world datasets, experiments are performed, and results are provided. The results illustrate that the proposed technique has broad applicability and performs equal to or better than the traditional stochastic frontier analysis technique.
Co-Membership-based Generic Anomalous Communities Detection
Lapid S, Kagan D and Fire M
Nowadays, detecting anomalous communities in networks is an essential task in research, as it helps discover insights into community-structured networks. Most of the existing methods leverage either information regarding attributes of vertices or the topological structure of communities. In this study, we introduce the Co-Membership-based Generic Anomalous Communities Detection Algorithm (referred as to ), a novel and generic method that utilizes the information of vertices co-membership in multiple communities. is domain-free and almost unaffected by communities' sizes and densities. Specifically, we train a classifier to predict the probability of each vertex in a community being a member of the community. We then rank the communities by the aggregated membership probabilities of each community's vertices. The lowest-ranked communities are considered to be anomalous. Furthermore, we present an algorithm for generating a community-structured random network enabling the infusion of anomalous communities to facilitate research in the field. We utilized it to generate two datasets, composed of thousands of labeled anomaly-infused networks, and published them. We experimented extensively on thousands of simulated, and real-world networks, infused with artificial anomalies. outperformed other existing methods in a range of settings. Additionally, we demonstrated that can identify abnormal communities in real-world unlabeled networks in different domains, such as Reddit and Wikipedia.
Improving the Polarity of Text through word2vec Embedding for Primary Classical Arabic Sentiment Analysis
Aoumeur NE, Li Z and Alshari EM
Over the past decade, Sentiment analysis has attracted significant researcher attention. Despite a huge number of studies in this field, Sentiment analysis of authors' books (classical Arabic) with extracting the embedding features has not yet been done. The recent feature extraction of Arabic text depends on the frequency of the words within the corpus without extracting the relation between these words. This paper aims to create a new classical Arabic dataset CASAD from many art books by collecting sentences from several stories with human-expert labeling. Additionally, the feature extraction of those datasets is created by word embedding techniques equivalent to Word2vec that are able to extract the deep relation which means features of the formal Arabic language. These features are evaluated by several types of machine learning for classical Arabic, for example, support vector machines (SVM), Logistic Regression (LR), Naive Bayes (NB) K-Nearest Neighbors (KNN), Latent Dirichlet Allocation (LDA) and Classification And Regression Trees (CART). Moreover, statistical methods such as validation and reliability are applied to evaluate this dataset's label. Finally, our experiments evaluated the classification rate of the feature-extraction matrices in two and three classes using six machine-learning algorithms for tenfold cross-validation that showed that the Logistic Regression with Word2Vec approach is the most accurate in predicting topic-polarity occurrence.
Edge Weight Updating Neural Network for Named Entity Normalization
Jeon SH and Cho S
Discriminating the matched named entity pairs or identifying the entities' canonical forms are critical in text mining tasks. More precise named entity normalization in text mining will benefit other subsequent text analytic applications. We built the named entity normalization model with a novel edge weight updating neural network. We, next, verify our model's performance on NCBI disease, BC5CDR disease, and BC5CDR chemical databases, which are widely used named entity normalization datasets in the bioinformatics field. We also tested our model with our own financial named entity normalization dataset to validate the efficacy for more general applications. Using the constructed dataset, we differentiate named entity pairs. Our model achieved the highest named entity normalization performances in terms of various evaluation metrics. Our proposed model when tested on four different datasets achieved state-of-the-art results.
AI Assisted Attention Mechanism for Hybrid Neural Model to Assess Online Attitudes About COVID-19
Kour H and Gupta MK
COVID-19 is a novel virus that presents challenges due to a lack of consistent and in-depth research. The news of the COVID-19 spreads across the globe, resulting in a flood of posts on social media sites. Apart from health, social, and economic disturbances brought by the COVID-19 pandemic, another important consequence involves public mental health crises which is of greater concern. Data related to COVID-19 is a valuable asset for researchers in understanding people's feelings related to the pandemic. It is thus important to extract the early information evolving public sentiments on social platforms during the outbreak of COVID-19. The objective of this study is to look at people's perceptions of the COVID-19 pandemic who interact with each other and share tweets on the Twitter platform. COVIDSenti, a large-scale benchmark dataset comprising 90,000 COVID-19 tweets collected from February to March 2020, during the initial phases of the outbreak served as the foundation for our experiments. A pre-trained bidirectional encoder representations from transformers (BERT) model is fine-tuned and embeddings generated are combined with two long short-term memory networks to propose the residual encoder transformation network model. The proposed model is used for multiclass text classification on a large dataset labeled as positive, negative, and neutral. The experimental outcomes validate that: (1) the proposed model is the best performing model, with 98% accuracy and 96% F1-score; (2) It also outperforms conventional machine learning algorithms and different variants of BERT, and (3) the approach achieves better results as compared to state-of-the-art on different benchmark datasets.
Prediction of SO Concentration Based on AR-LSTM Neural Network
Ju J, Liu K and Liu F
Sulphur dioxide is one of the most common air pollutants, forming acid rain and other harmful substances in the atmosphere, which can further damage our ecosystem and cause respiratory diseases in humans. Therefore, it is essential to monitor the concentration of sulphur dioxide produced in industrial processes in real-time to predict the concentration of sulphur dioxide emissions in the next few hours or days and to control them in advance. To address this problem, we propose an AR-LSTM analytical forecasting model based on ARIMA and LSTM. Based on the sensor's time series data set, we preprocess the data set and then carry out the modeling and analysis work. We analyze and predict the proposed analysis and prediction model in two data sets and conduct comparative experiments with other comparison models based on the three evaluation indicators of R, RMSE and MAE. The results demonstrated the effectiveness of the AR-LSTM analytical prediction model; Finally, a forecasting exercise was carried out for emissions in the coming weeks using our proposed AR-LSTM analytical forecasting model.
Attention Based Convolutional Neural Network with Multi-frequency Resolution Feature for Environment Sound Classification
Li M, Huang W and Zhang T
The environmental sound classification has great research significance in the fields of intelligent audio monitoring and other fields. A novel multi-frequency resolution (MFR) feature is proposed in this paper to solve the problem that the existing single frequency resolution time-frequency features of sound cannot effectively express the characteristics of multiple types of sound. The MFR feature is composed of three features with different frequency resolutions, which are compressed in varying degrees at the time dimension. This method not only has the effect of data augmentation but also can obtain more context information during the feature extraction. And the MFR features of Log-Mel Spectrogram, Cochleagram, and Constant Q-Transform are combined to form a multi-channel MFR feature. Also, a network named SacNet is built, which can effectively solve the problem that the time-frequency feature map of sound contains more invalid information. The basic structural unit of the SacNet consists of two parallel branches, one using depthwise separable convolution as the main feature extractor, and the other using spatial attention module to extract more effective information. Experiment results have demonstrated that the proposed method achieves the state-of-the-art accuracy of 97.5%, 93.1%, and 95.3% on three benchmark datasets of ESC10, ESC50, and UrbanSound8K respectively, which are increased by 3.3%, 0.5%, and 2.3% respectively compared with the previous advanced methods.
COVID-19 Detection from Chest X-rays Using Trained Output Based Transfer Learning Approach
Kumar S and Mallik A
The recent Coronavirus disease (COVID-19), which started in 2019, has spread across the globe and become a global pandemic. The efficient and effective COVID-19 detection using chest X-rays helps in early detection and curtailing the spread of the disease. In this paper, we propose a novel Trained Output-based Transfer Learning (TOTL) approach for COVID-19 detection from chest X-rays. We start by preprocessing the Chest X-rays of the patients with techniques like denoising, contrasting, segmentation. These processed images are then fed to several pre-trained transfer learning models like InceptionV3, InceptionResNetV2, Xception, MobileNet, ResNet50, ResNet50V2, VGG16, and VGG19. We fine-tune these models on the processed chest X-rays. Then we further train the outputs of these models using a deep neural network architecture to achieve enhanced performance and aggregate the capabilities of each of them. The proposed model has been tested on four recent COVID-19 chest X-rays datasets by computing several popular evaluation metrics. The performance of our model has also been compared with various deep transfer learning models and several contemporary COVID-19 detection methods. The obtained results demonstrate the efficiency and efficacy of our proposed model.
A New Time Series Forecasting Model Based on Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Temporal Convolutional Network
Guo C, Kang X, Xiong J and Wu J
In this paper, a new hybrid time series forecasting model based on the complete ensemble empirical mode decomposition with adaptive noise (CEEMDAN) and a temporal convolutional network (TCN) (CEEMDAN-TCN) is proposed. The CEEMDAN is used to decompose the time series data and the TCN is used to obtain a good prediction accuracy. The effectiveness of the model is verified in univariate and multivariate time series forecasting tasks. The experimental results indicate that compared with the long short-term memory model and other hybrid models, the proposed CEEMDAN-TCN model shows a better performance in both univariate and multivariate prediction tasks.
Application of Meta-Heuristic Algorithms for Training Neural Networks and Deep Learning Architectures: A Comprehensive Review
Kaveh M and Mesgari MS
The learning process and hyper-parameter optimization of artificial neural networks (ANNs) and deep learning (DL) architectures is considered one of the most challenging machine learning problems. Several past studies have used gradient-based back propagation methods to train DL architectures. However, gradient-based methods have major drawbacks such as stucking at local minimums in multi-objective cost functions, expensive execution time due to calculating gradient information with thousands of iterations and needing the cost functions to be continuous. Since training the ANNs and DLs is an NP-hard optimization problem, their structure and parameters optimization using the meta-heuristic (MH) algorithms has been considerably raised. MH algorithms can accurately formulate the optimal estimation of DL components (such as hyper-parameter, weights, number of layers, number of neurons, learning rate, etc.). This paper provides a comprehensive review of the optimization of ANNs and DLs using MH algorithms. In this paper, we have reviewed the latest developments in the use of MH algorithms in the DL and ANN methods, presented their disadvantages and advantages, and pointed out some research directions to fill the gaps between MHs and DL methods. Moreover, it has been explained that the evolutionary hybrid architecture still has limited applicability in the literature. Also, this paper classifies the latest MH algorithms in the literature to demonstrate their effectiveness in DL and ANN training for various applications. Most researchers tend to extend novel hybrid algorithms by combining MHs to optimize the hyper-parameters of DLs and ANNs. The development of hybrid MHs helps improving algorithms performance and capable of solving complex optimization problems. In general, the optimal performance of the MHs should be able to achieve a suitable trade-off between exploration and exploitation features. Hence, this paper tries to summarize various MH algorithms in terms of the convergence trend, exploration, exploitation, and the ability to avoid local minima. The integration of MH with DLs is expected to accelerate the training process in the coming few years. However, relevant publications in this way are still rare.
Application of Deep Learning Techniques in Diagnosis of Covid-19 (Coronavirus): A Systematic Review
Bhosale YH and Patnaik KS
Covid-19 is now one of the most incredibly intense and severe illnesses of the twentieth century. Covid-19 has already endangered the lives of millions of people worldwide due to its acute pulmonary effects. Image-based diagnostic techniques like X-ray, CT, and ultrasound are commonly employed to get a quick and reliable clinical condition. Covid-19 identification out of such clinical scans is exceedingly time-consuming, labor-intensive, and susceptible to silly intervention. As a result, radiography imaging approaches using Deep Learning (DL) are consistently employed to achieve great results. Various artificial intelligence-based systems have been developed for the early prediction of coronavirus using radiography pictures. Specific DL methods such as CNN and RNN noticeably extract extremely critical characteristics, primarily in diagnostic imaging. Recent coronavirus studies have used these techniques to utilize radiography image scans significantly. The disease, as well as the present pandemic, was studied using public and private data. A total of 64 pre-trained and custom DL models concerning imaging modality as taxonomies are selected from the studied articles. The constraints relevant to DL-based techniques are the sample selection, network architecture, training with minimal annotated database, and security issues. This includes evaluating causal agents, pathophysiology, immunological reactions, and epidemiological illness. DL-based Covid-19 detection systems are the key focus of this review article. Covid-19 work is intended to be accelerated as a result of this study.
Intelligent Identification of Jute Pests Based on Transfer Learning and Deep Convolutional Neural Networks
Sourav MSU and Wang H
Pest attacks pose a substantial threat to jute production and other significant crop plants. Jute farmers in Bangladesh generally distinguish between different pests that appear to be the same using their eyes and expertise, which isn't always accurate. We developed an intelligent model for jute pests identification based on transfer learning (TL) and deep convolutional neural networks (DCNN) to solve this practical problem. The proposed DCNN model can realize fast and accurate automatic identification of jute pests based on photographs. Specifically, the VGG19 CNN model was trained by TL on the ImageNet database. A well-structured image dataset of four dominant jute pests is also established. Our model shows a final accuracy of 95.86% on the four most vital jute pest classes. The model's performance is further demonstrated by the precision, recall, F1-score, and confusion matrix results. The proposed model is integrated into Android and IOS applications for practical uses.
Effectiveness of Federated Learning and CNN Ensemble Architectures for Identifying Brain Tumors Using MRI Images
Islam M, Reza MT, Kaosar M and Parvez MZ
Medical institutions often revoke data access due to the privacy concern of patients. Federated Learning (FL) is a collaborative learning paradigm that can generate an unbiased global model based on collecting updates from local models trained by client's data while keeping the local data private. This study aims to address the centralized data collection issue through the application of FL on brain tumor identification from MRI images. At first, several CNN models were trained using the MRI data and the best three performing CNN models were selected to form different variants of ensemble classifiers. Afterward, the FL model was constructed using the ensemble architecture. It was trained using model weights from the local model without sharing the client's data (MRI images) using the FL approach. Experimental results show only a slight decline in the performance of the FL approach as it achieved 91.05% accuracy compared to the 96.68% accuracy of the base ensemble model. Additionally, same approach was taken for another slightly larger dataset to prove the scalability of the method. This study shows that the FL approach can achieve privacy-protected tumor classification from MRI images without compromising much accuracy compared to the traditional deep learning approach.
A Novel Distant Domain Transfer Learning Framework for Thyroid Image Classification
Tang F, Ding J, Wang L and Ning C
Medical ultrasound imaging technology is currently the preferred method for early diagnosis of thyroid nodules. Radiologists' analysis of ultrasound images is highly dependent on their clinical experience and is susceptible to intra- and inter-observer variability. Although end-to-end deep learning technique can address these limitations, the difficulty of acquiring annotated medical image makes it very challenging. Transfer learning can alleviate the problems, but the large gap between source and target domain will lead to negative transfer. In this paper, a novel transfer learning method with distant domain high-level feature fusion (DHFF) model is proposed. It reduces the distribution distance between the source domain and the target domain while maintaining the characteristics of respective domains, which can avoid excessive feature fusion while enabling the model to learn more valuable transfer knowledge. The DHFF is validated by multiple public source and private target datasets in experiments. The results show that the classification accuracy of DHFF is up to 88.92% with thyroid ultrasound auxiliary source domains, which is up to 8% higher than existing transfer and distant transfer algorithms.
COVID-19 Variants and Transfer Learning for the Emerging Stringency Indices
Sohail A, Yu Z and Nutini A
The pandemics in the history of world health organization have always left memorable hallmarks, on the health care systems and on the economy of highly effected areas. The ongoing pandemic is one of the most harmful pandemics and is threatening due to its transformation to more contiguous variants. Here in this manuscript, we will first outline the variants and then their impact on the associated health issues. The deep learning algorithms are useful in developing models, from a higher dimensional problem/ dataset, but these algorithms fail to provide insight during the training process and do not generalize the conditions. Transfer learning, a new subfield of machine learning has acquired fame due to its ability to exploit the information/learning gained from a previous process to improve generalization for the next. In short, transfer learning is the optimization of the stored knowledge. With the aid of transfer learning, we will show that the stringency index and cardiovascular death rates were the most important and appropriate predictors to develop the model for the forecasting of the COVID-19 death rates.
Multi-Source Selection Transfer Learning with Privacy-Preserving
Wu W
Transfer learning has ability to create learning task of weakly labeled or unlabeled target domain by using knowledge of source domain to help, which can effectively improve the performance of target learning task. At present, the increased awareness of privacy protection restricts access to data sources and poses new challenges to the development of transfer learning. However, the research on privacy protection in transfer learning is very rare. The existing work mainly uses differential privacy technology and does not consider the distribution difference between data sources, or does not consider the conditional probability distribution of data, which causes negative transfer to harm the effect of algorithm. Therefore, this paper proposes multi-source selection transfer learning algorithm with privacy-preserving MultiSTLP, which is used in scenarios where target domain contains unlabeled data sets with only a small amount of group probability information and multiple source domains with a large number of labeled data sets. Group probability means that the class label of each sample in target data set is unknown, but the probability of each class in a given data group is available, and multiple source domains indicate that there are more than two source domains. The number of data set contains more than two data sets of source domain and one data set of target domain. The algorithm adapts to the marginal probability distribution and conditional probability distribution differences between domains, and can protect the privacy of target data and improve classification accuracy by fusing the idea of multi-source transfer learning and group probability into support vector machine. At the same time, it can select the representative dataset in source domains to improve efficiency relied on speeding up the training process of algorithm. Experimental results on several real datasets show the effectiveness of MultiSTLP, and it also has some advantages compared with the state-of-the-art transfer learning algorithm.
A Novel CNN-TLSTM Approach for Dengue Disease Identification and Prevention using IoT-Fog Cloud Architecture
Manoharan SN, Kumar KMVM and Vadivelan N
One of the mosquito-borne pandemic viral infections is Dengue which is mostly transmitted to humans by the Aedes agypti or female Aedes albopictis mosquitoes. The dengue disease expansion is mainly due to the different factors such as climate change, socioeconomic factors, viral evolution, globalization, etc. The unavailability of certain antiviral therapy and specific vaccine increases the risk of the dengue disease spreading even further. This arises the need for a novel technique that overcomes the complexities associated with dengue disease prediction such as low reporting level, misclassification, and incompatible disease monitoring framework. This paper mainly overcomes the above-mentioned problems by integrating the Internet of Things (IoT), fog-cloud, and deep learning techniques for efficient dengue monitoring. A compatible disease monitoring framework is formed via the IoT devices and the reports are effectively created and transferred to the healthcare facilities via the fog-cloud model. The misdiagnosis error is overcome in this paper using the novel Hybrid Convolutional Neural Network (CNN) with Tan Long and Short Term Memory (TLSTM) based Adaptive Teaching Learning Based Optimization (ATLBO) algorithm. The ATLBO optimized CNN-TLSTM architecture mainly analyzes the dengue-related parameters such as Soft Bleeding, Muscle Pain, Joint Pain, Skin rash, Fever, Water Site, Carbon Dioxide, Water Site Humidity, Water Site Temperature, etc. for an efficient clinical decision making and timely disease diagnosis. The experimental results are conducted using a real-time dataset and its performance is validated using various performance metrics. When compared in terms of different statistical parameters such as accuracy, f-measure, mean square error, and reliability, the proposed method offers superior results in the case of dengue disease detection than other existing methods. The ATLBO optimized hybrid CNN-TLSTM shows an accuracy of 96.9%, a precision of 95.7%, recall of 96.8%, and F-measure of 96.2% which is relatively high when compared to the existing techniques. The proposed model is capable of identifying the patients in a certain geographical region and preventing the disease emergency via immediate disease diagnosis and alerting the healthcare officials to offer the stipulated services.