Search results
21 – 30 of over 7000Hongfang Zhou, Xiqian Wang and Yao Zhang
Feature selection is an essential step in data mining. The core of it is to analyze and quantize the relevancy and redundancy between the features and the classes. In CFR feature…
Abstract
Feature selection is an essential step in data mining. The core of it is to analyze and quantize the relevancy and redundancy between the features and the classes. In CFR feature selection method, they rarely consider which feature to choose if two or more features have the same value using evaluation criterion. In order to address this problem, the standard deviation is employed to adjust the importance between relevancy and redundancy. Based on this idea, a novel feature selection method named as Feature Selection Based on Weighted Conditional Mutual Information (WCFR) is introduced. Experimental results on ten datasets show that our proposed method has higher classification accuracy.
Details
Keywords
Sandeep Kumar Hegde and Monica R. Mundada
Chronic diseases are considered as one of the serious concerns and threats to public health across the globe. Diseases such as chronic diabetes mellitus (CDM), cardio…
Abstract
Purpose
Chronic diseases are considered as one of the serious concerns and threats to public health across the globe. Diseases such as chronic diabetes mellitus (CDM), cardio vasculardisease (CVD) and chronic kidney disease (CKD) are major chronic diseases responsible for millions of death. Each of these diseases is considered as a risk factor for the other two diseases. Therefore, noteworthy attention is being paid to reduce the risk of these diseases. A gigantic amount of medical data is generated in digital form from smart healthcare appliances in the current era. Although numerous machine learning (ML) algorithms are proposed for the early prediction of chronic diseases, these algorithmic models are neither generalized nor adaptive when the model is imposed on new disease datasets. Hence, these algorithms have to process a huge amount of disease data iteratively until the model converges. This limitation may make it difficult for ML models to fit and produce imprecise results. A single algorithm may not yield accurate results. Nonetheless, an ensemble of classifiers built from multiple models, that works based on a voting principle has been successfully applied to solve many classification tasks. The purpose of this paper is to make early prediction of chronic diseases using hybrid generative regression based deep intelligence network (HGRDIN) model.
Design/methodology/approach
In the proposed paper generative regression (GR) model is used in combination with deep neural network (DNN) for the early prediction of chronic disease. The GR model will obtain prior knowledge about the labelled data by analyzing the correlation between features and class labels. Hence, the weight assignment process of DNN is influenced by the relationship between attributes rather than random assignment. The knowledge obtained through these processes is passed as input to the DNN network for further prediction. Since the inference about the input data instances is drawn at the DNN through the GR model, the model is named as hybrid generative regression-based deep intelligence network (HGRDIN).
Findings
The credibility of the implemented approach is rigorously validated using various parameters such as accuracy, precision, recall, F score and area under the curve (AUC) score. During the training phase, the proposed algorithm is constantly regularized using the elastic net regularization technique and also hyper-tuned using the various parameters such as momentum and learning rate to minimize the misprediction rate. The experimental results illustrate that the proposed approach predicted the chronic disease with a minimal error by avoiding the possible overfitting and local minima problems. The result obtained with the proposed approach is also compared with the various traditional approaches.
Research limitations/implications
Usually, the diagnostic data are multi-dimension in nature where the performance of the ML algorithm will degrade due to the data overfitting, curse of dimensionality issues. The result obtained through the experiment has achieved an average accuracy of 95%. Hence, analysis can be made further to improve predictive accuracy by overcoming the curse of dimensionality issues.
Practical implications
The proposed ML model can mimic the behavior of the doctor's brain. These algorithms have the capability to replace clinical tasks. The accurate result obtained through the innovative algorithms can free the physician from the mundane care and practices so that the physician can focus more on the complex issues.
Social implications
Utilizing the proposed predictive model at the decision-making level for the early prediction of the disease is considered as a promising change towards the healthcare sector. The global burden of chronic disease can be reduced at an exceptional level through these approaches.
Originality/value
In the proposed HGRDIN model, the concept of transfer learning approach is used where the knowledge acquired through the GR process is applied on DNN that identified the possible relationship between the dependent and independent feature variables by mapping the chronic data instances to its corresponding target class before it is being passed as input to the DNN network. Hence, the result of the experiments illustrated that the proposed approach obtained superior performance in terms of various validation parameters than the existing conventional techniques.
Details
Keywords
Nitha Thomas, Joshin John Mathew and Alex James
The real-time generation of feature descriptors for object recognition is a challenging problem. In this research, the purpose of this paper is to provide a hardware friendly…
Abstract
Purpose
The real-time generation of feature descriptors for object recognition is a challenging problem. In this research, the purpose of this paper is to provide a hardware friendly framework to generate sparse features that can be useful for key feature point selection, feature extraction, and descriptor construction. The inspiration is drawn from feature formation processes of the human brain, taking into account the sparse, modular, and hierarchical processing of visual information.
Design/methodology/approach
A sparse set of neurons referred as active neurons determines the feature points necessary for high-level vision applications such as object recognition. A psycho-physical mechanism of human low-level vision relates edge detection to noticeable local spatial stimuli, representing this set of active neurons. A cognitive memory cell array-based implementation of low-level vision is proposed. Applications of memory cell in edge detection are used for realizing human vision inspired feature selection and leading to feature vector construction for high-level vision applications.
Findings
True parallel architecture and faster response of cognitive circuits avoid time costly and redundant feature extraction steps. Validation of proposed feature vector toward high-level computer vision applications is demonstrated using standard object recognition databases. The comparison against existing state-of-the-art object recognition features and methods shows an accuracy of 97, 95, 69 percent for Columbia Object Image Library-100, ALOI, and PASCAL VOC 2007 databases indicating an increase from benchmark methods by 5, 3 and 10 percent, respectively.
Originality/value
A hardware friendly low-level sparse edge feature processing system is proposed for recognizing objects. The edge features are developed based on threshold logic of neurons, and the sparse selection of the features applies a modular and hierarchical processing inspired from the human neural system.
Details
Keywords
Jia Yan, Shukai Duan, Tingwen Huang and Lidan Wang
The purpose of this paper is to improve the performance of E-nose in the detection of wound infection. Feature extraction and selection methods have a strong impact on the…
Abstract
Purpose
The purpose of this paper is to improve the performance of E-nose in the detection of wound infection. Feature extraction and selection methods have a strong impact on the performance of pattern classification of electronic nose (E-nose). A new hybrid feature matrix construction method and multi-objective binary quantum-behaved particle swarm optimization (BQPSO) have been proposed for feature extraction and selection of sensor array.
Design/methodology/approach
A hybrid feature matrix constructed by maximum value and wavelet coefficients is proposed to realize feature extraction. Multi-objective BQPSO whose fitness function contains classification accuracy and a number of selected sensors is used for feature selection. Quantum-behaved particle swarm optimization (QPSO) is used for synchronization optimization of selected features and parameter of classifier. Radical basis function (RBF) network is used for classification.
Findings
E-nose obtains the highest classification accuracy when the maximum value and db 5 wavelet coefficients are extracted as the hybrid features and only six sensors are selected for classification. All results make it clear that the proposed method is an ideal feature extraction and selection method of E-nose in the detection of wound infection.
Originality/value
The innovative concept improves the performance of E-nose in wound monitoring, and is beneficial for realizing the clinical application of E-nose.
Details
Keywords
Fuzan Chen, Harris Wu, Runliang Dou and Minqiang Li
The purpose of this paper is to build a compact and accurate classifier for high-dimensional classification.
Abstract
Purpose
The purpose of this paper is to build a compact and accurate classifier for high-dimensional classification.
Design/methodology/approach
A classification approach based on class-dependent feature subspace (CFS) is proposed. CFS is a class-dependent integration of a support vector machine (SVM) classifier and associated discriminative features. For each class, our genetic algorithm (GA)-based approach evolves the best subset of discriminative features and SVM classifier simultaneously. To guarantee convergence and efficiency, the authors customize the GA in terms of encoding strategy, fitness evaluation, and genetic operators.
Findings
Experimental studies demonstrated that the proposed CFS-based approach is superior to other state-of-the-art classification algorithms on UCI data sets in terms of both concise interpretation and predictive power for high-dimensional data.
Research limitations/implications
UCI data sets rather than real industrial data are used to evaluate the proposed approach. In addition, only single-label classification is addressed in the study.
Practical implications
The proposed method not only constructs an accurate classification model but also obtains a compact combination of discriminative features. It is helpful for business makers to get a concise understanding of the high-dimensional data.
Originality/value
The authors propose a compact and effective classification approach for high-dimensional data. Instead of the same feature subset for all the classes, the proposed CFS-based approach obtains the optimal subset of discriminative feature and SVM classifier for each class. The proposed approach enhances both interpretability and predictive power for high-dimensional data.
Details
Keywords
Arslan Akram, Saba Ramzan, Akhtar Rasool, Arfan Jaffar, Usama Furqan and Wahab Javed
This paper aims to propose a novel splicing detection method using a discriminative robust local binary pattern (DRLBP) with a support vector machine (SVM). Reliable detection of…
Abstract
Purpose
This paper aims to propose a novel splicing detection method using a discriminative robust local binary pattern (DRLBP) with a support vector machine (SVM). Reliable detection of image splicing is of growing interest due to the extensive utilization of digital images as a communication medium and the availability of powerful image processing tools. Image splicing is a commonly used forgery technique in which a region of an image is copied and pasted to a different image to hide the original contents of the image.
Design/methodology/approach
The structural changes caused due to splicing are robustly described by DRLBP. The changes caused by image forgery are localized, so as a first step, localized description is divided into overlapping blocks by providing an image as input. DRLBP descriptor is calculated for each block, and the feature vector is created by concatenation. Finally, features are passed to the SVM classifier to predict whether the image is genuine or forged.
Findings
The performance and robustness of the method are evaluated on public domain benchmark data sets and achieved 98.95% prediction accuracy. The results are compared with state-of-the-art image splicing finding approaches, and it shows that the performance of the proposed method is improved using the given technique.
Originality/value
The proposed method is using DRLBP, an efficient texture descriptor, which combines both corner and inside design detail in a single representation. It produces discriminative and compact features in such a way that there is no need for the feature selection process to drop the redundant and insignificant features.
Details
Keywords
Oladosu Oyebisi Oladimeji, Abimbola Oladimeji and Olayanju Oladimeji
Diabetes is one of the life-threatening chronic diseases, which is already affecting 422m people globally based on (World Health Organization) WHO report as at 2018. This costs…
Abstract
Purpose
Diabetes is one of the life-threatening chronic diseases, which is already affecting 422m people globally based on (World Health Organization) WHO report as at 2018. This costs individuals, government and groups a whole lot; right from its diagnosis stage to the treatment stage. The reason for this cost, among others, is that it is a long-term treatment disease. This disease is likely to continue to affect more people because of its long asymptotic phase, which makes its early detection not feasible.
Design/methodology/approach
In this study, the authors have presented machine learning models with feature selection, which can detect diabetes disease at its early stage. Also, the models presented are not costly and available to everyone, including those in the remote areas.
Findings
The study result shows that feature selection helps in getting better model, as it prevents overfitting and removes redundant data. Hence, the study result when compared with previous research shows the better result has been achieved, after it was evaluated based on metrics such as F-measure, Precision-Recall curve and Receiver Operating Characteristic Area Under Curve. This discovery has the potential to impact on clinical practice, when health workers aim at diagnosing diabetes disease at its early stage.
Originality/value
This study has not been published anywhere else.
Details
Keywords
Ali Leylavi Shoushtari, Stefano Mazzoleni and Paolo Dario
This paper aims to propose an innovative kinematic control algorithm for redundant robotic manipulators. The algorithm takes advantage of a bio-inspired approach.
Abstract
Purpose
This paper aims to propose an innovative kinematic control algorithm for redundant robotic manipulators. The algorithm takes advantage of a bio-inspired approach.
Design/methodology/approach
A simplified two-degree-of-freedom model is presented to handle kinematic redundancy in the x-y plane; an extension to three-dimensional tracking tasks is presented as well. A set of sample trajectories was used to evaluate the performances of the proposed algorithm.
Findings
The results from the simulations confirm the continuity and accuracy of generated joint profiles for given end-effector trajectories as well as algorithm robustness, singularity and self-collision avoidance.
Originality/value
This paper shows how to control a redundant robotic arm by applying human upper arm-inspired concept of inter-joint dependency.
Details
Keywords
Zainab Akhtar, Jong Weon Lee, Muhammad Attique Khan, Muhammad Sharif, Sajid Ali Khan and Naveed Riaz
In artificial intelligence, the optical character recognition (OCR) is an active research area based on famous applications such as automation and transformation of printed…
Abstract
Purpose
In artificial intelligence, the optical character recognition (OCR) is an active research area based on famous applications such as automation and transformation of printed documents into machine-readable text document. The major purpose of OCR in academia and banks is to achieve a significant performance to save storage space.
Design/methodology/approach
A novel technique is proposed for automated OCR based on multi-properties features fusion and selection. The features are fused using serially formulation and output passed to partial least square (PLS) based selection method. The selection is done based on the entropy fitness function. The final features are classified by an ensemble classifier.
Findings
The presented method was extensively tested on two datasets such as the authors proposed and Chars74k benchmark and achieved an accuracy of 91.2 and 99.9%. Comparing the results with existing techniques, it is found that the proposed method gives improved performance.
Originality/value
The technique presented in this work will help for license plate recognition and text conversion from a printed document to machine-readable.
Details
Keywords
Guan Yuan, Zhaohui Wang, Fanrong Meng, Qiuyan Yan and Shixiong Xia
Currently, ubiquitous smartphones embedded with various sensors provide a convenient way to collect raw sequence data. These data bridges the gap between human activity and…
Abstract
Purpose
Currently, ubiquitous smartphones embedded with various sensors provide a convenient way to collect raw sequence data. These data bridges the gap between human activity and multiple sensors. Human activity recognition has been widely used in quite a lot of aspects in our daily life, such as medical security, personal safety, living assistance and so on.
Design/methodology/approach
To provide an overview, the authors survey and summarize some important technologies and involved key issues of human activity recognition, including activity categorization, feature engineering as well as typical algorithms presented in recent years. In this paper, the authors first introduce the character of embedded sensors and dsiscuss their features, as well as survey some data labeling strategies to get ground truth label. Then, following the process of human activity recognition, the authors discuss the methods and techniques of raw data preprocessing and feature extraction, and summarize some popular algorithms used in model training and activity recognizing. Third, they introduce some interesting application scenarios of human activity recognition and provide some available data sets as ground truth data to validate proposed algorithms.
Findings
The authors summarize their viewpoints on human activity recognition, discuss the main challenges and point out some potential research directions.
Originality/value
It is hoped that this work will serve as the steppingstone for those interested in advancing human activity recognition.
Details