Search results

1 – 10 of over 2000
Open Access
Article
Publication date: 17 July 2020

Sheryl Brahnam, Loris Nanni, Shannon McMurtrey, Alessandra Lumini, Rick Brattin, Melinda Slack and Tonya Barrier

Diagnosing pain in neonates is difficult but critical. Although approximately thirty manual pain instruments have been developed for neonatal pain diagnosis, most are complex…

2284

Abstract

Diagnosing pain in neonates is difficult but critical. Although approximately thirty manual pain instruments have been developed for neonatal pain diagnosis, most are complex, multifactorial, and geared toward research. The goals of this work are twofold: 1) to develop a new video dataset for automatic neonatal pain detection called iCOPEvid (infant Classification Of Pain Expressions videos), and 2) to present a classification system that sets a challenging comparison performance on this dataset. The iCOPEvid dataset contains 234 videos of 49 neonates experiencing a set of noxious stimuli, a period of rest, and an acute pain stimulus. From these videos 20 s segments are extracted and grouped into two classes: pain (49) and nopain (185), with the nopain video segments handpicked to produce a highly challenging dataset. An ensemble of twelve global and local descriptors with a Bag-of-Features approach is utilized to improve the performance of some new descriptors based on Gaussian of Local Descriptors (GOLD). The basic classifier used in the ensembles is the Support Vector Machine, and decisions are combined by sum rule. These results are compared with standard methods, some deep learning approaches, and 185 human assessments. Our best machine learning methods are shown to outperform the human judges.

Details

Applied Computing and Informatics, vol. 19 no. 1/2
Type: Research Article
ISSN: 2634-1964

Keywords

Article
Publication date: 17 September 2019

Chérif Taouche and Hacene Belhadef

Palmprint recognition is a very interesting and promising area of research. Much work has already been done in this area, but much more needs to be done to make the systems more…

73

Abstract

Purpose

Palmprint recognition is a very interesting and promising area of research. Much work has already been done in this area, but much more needs to be done to make the systems more efficient. In this paper, a multimodal biometrics system based on fusion of left and right palmprints of a person is proposed to overcome limitations of unimodal systems.

Design/methodology/approach

Features are extracted using some proposed multi-block local descriptors in addition to MBLBP. Fusion of extracted features is done at feature level by a simple concatenation of feature vectors. Then, feature selection is performed on the resulting global feature vector using evolutionary algorithms such as genetic algorithms and backtracking search algorithm for a comparison purpose. The benefits of such step selecting the relevant features are known in the literature, such as increasing the recognition accuracy and reducing the feature set size, which results in runtime saving. In matching step, Chi-square similarity measure is used.

Findings

The resulting feature vector length representing a person is compact and the runtime is reduced.

Originality/value

Intensive experiments were done on the publicly available IITD database. Experimental results show a recognition accuracy of 99.17 which prove the effectiveness and robustness of the proposed multimodal biometrics system than other unimodal and multimodal biometrics systems.

Details

Information Discovery and Delivery, vol. 48 no. 1
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 29 July 2014

Chengyong Zheng, Hong Li and Guokuan Li

This paper presents a novel printed circuit board (PCB) film image alignment method based on distance context of image components, which can be directly used for PCB film…

Abstract

Purpose

This paper presents a novel printed circuit board (PCB) film image alignment method based on distance context of image components, which can be directly used for PCB film inspection. PCB film inspection plays a very important role in PCB production.

Design/methodology/approach

First, image components of reference film image and inspected film image are extracted. Then, local distance context (LDC) and global distance context (GDC) are computed for each image component. Using LDC and GDC, the similarity of each pair of components between the reference film image and the inspected film image are computed, the component correspondences can be established accordingly and the parameters for aligning these two images can be eventually estimated.

Findings

LDC and GDC act as the local spatial distribution descriptor and the global relative position descriptor of the current component, and they are invariant to translation, rotating and scale. Experimental results on aligning real PCB film images against various rotations and scaling transformation show that the proposed algorithm is fast and accurate and is very suitable for PCB film inspection.

Research limitations/implications

The proposed algorithm is suitable for aligning those images that have some isolated connected components, such as the PCB film images. It is not suitable for general image alignment.

Originality/value

We put forward to use LDC and GDC as the local descriptor and global descriptor of an image component, and designed a PCB film image alignment algorithm that can overcome the shortcomings of that image alignment algorithm that was based on local feature descriptors such as Fourier descriptor.

Details

Circuit World, vol. 40 no. 3
Type: Research Article
ISSN: 0305-6120

Keywords

Article
Publication date: 23 August 2019

Shenlong Wang, Kaixin Han and Jiafeng Jin

In the past few decades, the content-based image retrieval (CBIR), which focuses on the exploration of image feature extraction methods, has been widely investigated. The term of…

Abstract

Purpose

In the past few decades, the content-based image retrieval (CBIR), which focuses on the exploration of image feature extraction methods, has been widely investigated. The term of feature extraction is used in two cases: application-based feature expression and mathematical approaches for dimensionality reduction. Feature expression is a technique of describing the image color, texture and shape information with feature descriptors; thus, obtaining effective image features expression is the key to extracting high-level semantic information. However, most of the previous studies regarding image feature extraction and expression methods in the CBIR have not performed systematic research. This paper aims to introduce the basic image low-level feature expression techniques for color, texture and shape features that have been developed in recent years.

Design/methodology/approach

First, this review outlines the development process and expounds the principle of various image feature extraction methods, such as color, texture and shape feature expression. Second, some of the most commonly used image low-level expression algorithms are implemented, and the benefits and drawbacks are summarized. Third, the effectiveness of the global and local features in image retrieval, including some classical models and their illustrations provided by part of our experiment, are analyzed. Fourth, the sparse representation and similarity measurement methods are introduced, and the retrieval performance of statistical methods is evaluated and compared.

Findings

The core of this survey is to review the state of the image low-level expression methods and study the pros and cons of each method, their applicable occasions and certain implementation measures. This review notes that image peculiarities of single-feature descriptions may lead to unsatisfactory image retrieval capabilities, which have significant singularity and considerable limitations and challenges in the CBIR.

Originality/value

A comprehensive review of the latest developments in image retrieval using low-level feature expression techniques is provided in this paper. This review not only introduces the major approaches for image low-level feature expression but also supplies a pertinent reference for those engaging in research regarding image feature extraction.

Article
Publication date: 19 December 2023

Jinchao Huang

Single-shot multi-category clothing recognition and retrieval play a crucial role in online searching and offline settlement scenarios. Existing clothing recognition methods based…

Abstract

Purpose

Single-shot multi-category clothing recognition and retrieval play a crucial role in online searching and offline settlement scenarios. Existing clothing recognition methods based on RGBD clothing images often suffer from high-dimensional feature representations, leading to compromised performance and efficiency.

Design/methodology/approach

To address this issue, this paper proposes a novel method called Manifold Embedded Discriminative Feature Selection (MEDFS) to select global and local features, thereby reducing the dimensionality of the feature representation and improving performance. Specifically, by combining three global features and three local features, a low-dimensional embedding is constructed to capture the correlations between features and categories. The MEDFS method designs an optimization framework utilizing manifold mapping and sparse regularization to achieve feature selection. The optimization objective is solved using an alternating iterative strategy, ensuring convergence.

Findings

Empirical studies conducted on a publicly available RGBD clothing image dataset demonstrate that the proposed MEDFS method achieves highly competitive clothing classification performance while maintaining efficiency in clothing recognition and retrieval.

Originality/value

This paper introduces a novel approach for multi-category clothing recognition and retrieval, incorporating the selection of global and local features. The proposed method holds potential for practical applications in real-world clothing scenarios.

Details

International Journal of Intelligent Computing and Cybernetics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 17 March 2014

Giulio Reina, Mauro Bellone, Luigi Spedicato and Nicola Ivan Giannoccaro

This research aims to address the issue of safe navigation for autonomous vehicles in highly challenging outdoor environments. Indeed, robust navigation of autonomous mobile…

Abstract

Purpose

This research aims to address the issue of safe navigation for autonomous vehicles in highly challenging outdoor environments. Indeed, robust navigation of autonomous mobile robots over long distances requires advanced perception means for terrain traversability assessment.

Design/methodology/approach

The use of visual systems may represent an efficient solution. This paper discusses recent findings in terrain traversability analysis from RGB-D images. In this context, the concept of point as described only by its Cartesian coordinates is reinterpreted in terms of local description. As a result, a novel descriptor for inferring the traversability of a terrain through its 3D representation, referred to as the unevenness point descriptor (UPD), is conceived. This descriptor features robustness and simplicity.

Findings

The UPD-based algorithm shows robust terrain perception capabilities in both indoor and outdoor environment. The algorithm is able to detect obstacles and terrain irregularities. The system performance is validated in field experiments in both indoor and outdoor environments.

Research limitations/implications

The UPD enhances the interpretation of 3D scene to improve the ambient awareness of unmanned vehicles. The larger implications of this method reside in its applicability for path planning purposes.

Originality/value

This paper describes a visual algorithm for traversability assessment based on normal vectors analysis. The algorithm is simple and efficient providing fast real-time implementation, since the UPD does not require any data processing or previously generated digital elevation map to classify the scene. Moreover, it defines a local descriptor, which can be of general value for segmentation purposes of 3D point clouds and allows the underlining geometric pattern associated with each single 3D point to be fully captured and difficult scenarios to be correctly handled.

Details

Sensor Review, vol. 34 no. 2
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 23 August 2019

Yiqun Kuang, Hong Cheng, Yali Zheng, Fang Cui and Rui Huang

This paper aims to present a one-shot gesture recognition approach which can be a high-efficient communication channel in human–robot collaboration systems.

Abstract

Purpose

This paper aims to present a one-shot gesture recognition approach which can be a high-efficient communication channel in human–robot collaboration systems.

Design/methodology/approach

This paper applies dynamic time warping (DTW) to align two gesture sequences in temporal domain with a novel frame-wise distance measure which matches local features in spatial domain. Furthermore, a novel and robust bidirectional attention region extraction method is proposed to retain information in both movement and hold phase of a gesture.

Findings

The proposed approach is capable of providing efficient one-shot gesture recognition without elaborately designed features. The experiments on a social robot (JiaJia) demonstrate that the proposed approach can be used in a human–robot collaboration system flexibly.

Originality/value

According to previous literature, there are no similar solutions that can achieve an efficient gesture recognition with simple local feature descriptor and combine the advantages of local features with DTW.

Details

Assembly Automation, vol. 40 no. 1
Type: Research Article
ISSN: 0144-5154

Keywords

Article
Publication date: 14 May 2020

Minghua Wei

In order to solve the problem that the performance of the existing local feature descriptors in uncontrolled environment is greatly affected by illumination, background, occlusion…

135

Abstract

Purpose

In order to solve the problem that the performance of the existing local feature descriptors in uncontrolled environment is greatly affected by illumination, background, occlusion and other factors, we propose a novel face recognition algorithm in uncontrolled environment which combines the block central symmetry local binary pattern (CS-LBP) and deep residual network (DRN) model.

Design/methodology/approach

The algorithm first extracts the block CSP-LBP features of the face image, then incorporates the extracted features into the DRN model, and gives the face recognition results by using a well-trained DRN model. The features obtained by the proposed algorithm have the characteristics of both local texture features and deep features that robust to illumination.

Findings

Compared with the direct usage of the original image, the usage of local texture features of the image as the input of DRN model significantly improves the computation efficiency. Experimental results on the face datasets of FERET, YALE-B and CMU-PIE have shown that the recognition rate of the proposed algorithm is significantly higher than that of other compared algorithms.

Originality/value

The proposed algorithm fundamentally solves the problem of face identity recognition in uncontrolled environment, and it is particularly robust to the change of illumination, which proves its superiority.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 13 no. 2
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 24 September 2019

Kun Wei, Yong Dai and Bingyin Ren

This paper aims to propose an identification method based on monocular vision for cylindrical parts in cluttered scene, which solves the issue that iterative closest point (ICP…

Abstract

Purpose

This paper aims to propose an identification method based on monocular vision for cylindrical parts in cluttered scene, which solves the issue that iterative closest point (ICP) algorithm fails to obtain global optimal solution, as the deviation from scene point cloud to target CAD model is huge in nature.

Design/methodology/approach

The images of the parts are captured at three locations by a camera amounted on a robotic end effector to reconstruct initial scene point cloud. Color signatures of histogram of orientations (C-SHOT) local feature descriptors are extracted from the model and scene point cloud. Random sample consensus (RANSAC) algorithm is used to perform the first initial matching of point sets. Then, the second initial matching is conducted by proposed remote closest point (RCP) algorithm to make the model get close to the scene point cloud. Levenberg Marquardt (LM)-ICP is used to complete fine registration to obtain accurate pose estimation.

Findings

The experimental results in bolt-cluttered scene demonstrate that the accuracy of pose estimation obtained by the proposed method is higher than that obtained by two other methods. The position error is less than 0.92 mm and the orientation error is less than 0.86°. The average recognition rate is 96.67 per cent and the identification time of the single bolt does not exceed 3.5 s.

Practical implications

The presented approach can be applied or integrated into automatic sorting production lines in the factories.

Originality/value

The proposed method improves the efficiency and accuracy of the identification and classification of cylindrical parts using a robotic arm.

Article
Publication date: 26 January 2022

Ziming Zeng, Shouqiang Sun, Tingting Li, Jie Yin and Yueyan Shen

The purpose of this paper is to build a mobile visual search service system for the protection of Dunhuang cultural heritage in the smart library. A novel mobile visual search…

Abstract

Purpose

The purpose of this paper is to build a mobile visual search service system for the protection of Dunhuang cultural heritage in the smart library. A novel mobile visual search model for Dunhuang murals is proposed to help users acquire rich knowledge and services conveniently.

Design/methodology/approach

First, local and global features of images are extracted, and the visual dictionary is generated by the k-means clustering. Second, the mobile visual search model based on the bag-of-words (BOW) and multiple semantic associations is constructed. Third, the mobile visual search service system of the smart library is designed in the cloud environment. Furthermore, Dunhuang mural images are collected to verify this model.

Findings

The findings reveal that the BOW_SIFT_HSV_MSA model has better search performance for Dunhuang mural images when the scale-invariant feature transform (SIFT) and the hue, saturation and value (HSV) are used to extract local and global features of the images. Compared with different methods, this model is the most effective way to search images with the semantic association in the topic, time and space dimensions.

Research limitations/implications

Dunhuang mural image set is a part of the vast resources stored in the smart library, and the fine-grained semantic labels could be applied to meet diverse search needs.

Originality/value

The mobile visual search service system is constructed to provide users with Dunhuang cultural services in the smart library. A novel mobile visual search model based on BOW and multiple semantic associations is proposed. This study can also provide references for the protection and utilization of other cultural heritages.

Details

Library Hi Tech, vol. 40 no. 6
Type: Research Article
ISSN: 0737-8831

Keywords

1 – 10 of over 2000