Search results

1 – 10 of 10
Article
Publication date: 25 October 2021

Erdem Tunca, Hasan Saribas, Hasim Kafali and Sinem Kahvecioglu

The purpose of this paper is to monitor the backup indicators in case of indicator failure and to minimize the situations when the pilot may be unable to monitor the indicator…

Abstract

Purpose

The purpose of this paper is to monitor the backup indicators in case of indicator failure and to minimize the situations when the pilot may be unable to monitor the indicator effectively in emergency situations.

Design/methodology/approach

In this study, the pointer positions of different indicators were determined with a deep learning-based algorithm. Within the scope of the study, the pointer on the analog indicators obtained from aircraft cockpits was detected with the YOLOv4 object detector. Then, segmentation was made with the GrabCut algorithm to detect the pointer in the detected region more precisely. Finally, a line including the segmented pointer was found using the least-squares method, and the exact direction of the pointer was determined and the angle value of the pointer was obtained by using the inverse tangent function. In addition, to detect the pointer of the YOLOv4 object detection method and to test the designed method, a data set consisting of videos taken from aircraft cockpits was created and labeled.

Findings

The analog indicator pointers were detected with great accuracy by the YOLOv4 and YOLOv4-Tiny detectors. The experimental results show that the proposed method estimated the angle of the pointer with a high degree of accuracy. The developed method can reduce the workloads of both pilots and flight engineers. Similarly, the performance of pilots can be evaluated with this method.

Originality/value

The authors propose a novel real-time method which consists of detection, segmentation and line regression modules for mapping the angle of the pointers on analog indicators. A data set that includes analog indicators taken from aircraft cockpits was collected and labeled to train and test the proposed method.

Details

Aircraft Engineering and Aerospace Technology, vol. 94 no. 3
Type: Research Article
ISSN: 1748-8842

Keywords

Article
Publication date: 31 December 2021

Ajanthaa Lakkshmanan, C. Anbu Ananth and S. Tiroumalmouroughane S. Tiroumalmouroughane

The advancements of deep learning (DL) models demonstrate significant performance on accurate pancreatic tumor segmentation and classification.

93

Abstract

Purpose

The advancements of deep learning (DL) models demonstrate significant performance on accurate pancreatic tumor segmentation and classification.

Design/methodology/approach

The presented model involves different stages of operations, namely preprocessing, image segmentation, feature extraction and image classification. Primarily, bilateral filtering (BF) technique is applied for image preprocessing to eradicate the noise present in the CT pancreatic image. Besides, noninteractive GrabCut (NIGC) algorithm is applied for the image segmentation process. Subsequently, residual network 152 (ResNet152) model is utilized as a feature extractor to originate a valuable set of feature vectors. At last, the red deer optimization algorithm (RDA) tuned backpropagation neural network (BPNN), called RDA-BPNN model, is employed as a classification model to determine the existence of pancreatic tumor.

Findings

The experimental results are validated in terms of different performance measures and a detailed comparative results analysis ensured the betterment of the RDA-BPNN model with the sensitivity of 98.54%, specificity of 98.46%, accuracy of 98.51% and F-score of 98.23%.

Originality/value

The study also identifies several novel automated deep learning based approaches used by researchers to assess the performance of the RDA-BPNN model on benchmark dataset and analyze the results in terms of several measures.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 15 no. 3
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 23 November 2021

Srinivas Talasila, Kirti Rawal and Gaurav Sethi

Extraction of leaf region from the plant leaf images is a prerequisite process for species recognition, disease detection and classification and so on, which are required for crop…

Abstract

Purpose

Extraction of leaf region from the plant leaf images is a prerequisite process for species recognition, disease detection and classification and so on, which are required for crop management. Several approaches were developed to implement the process of leaf region segmentation from the background. However, most of the methods were applied to the images taken under laboratory setups or plain background, but the application of leaf segmentation methods is vital to be used on real-time cultivation field images that contain complex backgrounds. So far, the efficient method that automatically segments leaf region from the complex background exclusively for black gram plant leaf images has not been developed.

Design/methodology/approach

Extracting leaf regions from the complex background is cumbersome, and the proposed PLRSNet (Plant Leaf Region Segmentation Net) is one of the solutions to this problem. In this paper, a customized deep network is designed and applied to extract leaf regions from the images taken from cultivation fields.

Findings

The proposed PLRSNet compared with the state-of-the-art methods and the experimental results evident that proposed PLRSNet yields 96.9% of Similarity Index/Dice, 94.2% of Jaccard/IoU, 98.55% of Correct Detection Ratio, Total Segmentation Error of 0.059 and Average Surface Distance of 3.037, representing a significant improvement over existing methods particularly taking into account of cultivation field images.

Originality/value

In this work, a customized deep learning network is designed for segmenting plant leaf region under complex background and named it as a PLRSNet.

Details

International Journal of Intelligent Unmanned Systems, vol. 11 no. 1
Type: Research Article
ISSN: 2049-6427

Keywords

Article
Publication date: 27 March 2009

Yuan Tian, Tao Guan, Cheng Wang, Lijun Li and Wei Liu

The purpose of this paper is to present an efficient, interactive foreground/background image segmentation method using mean shift (MS) and graph cuts, in order to improve the…

Abstract

Purpose

The purpose of this paper is to present an efficient, interactive foreground/background image segmentation method using mean shift (MS) and graph cuts, in order to improve the segmentation performance with little user interaction.

Design/methodology/approach

By incorporating the advantages of the mean shift method and the graph cut algorithm, the proposed approach ensures the accuracy of segmentation results. First, the user marks certain pixels as foreground or background. Then the graph is constructed and the cost function composed of the boundary properties and the region properties is defined. To obtain the hidden information of user interaction, the foreground and background marks are clustered separately by the mean shift method. The region properties are determined by the minimum distances from the unmarked pixels to the foreground and background clusters. The boundary properties are determined by the relationship between the unmarked pixels and its neighbor pixels. Finally, using the graph cuts method solves the energy minimization problem to get the foreground which is of interest.

Findings

The paper presents experimental results and compares the results to other methods. It can be seen from the comparison that this method can obtain a better segmentation performance in many cases.

Originality/value

The paper incorporates the advantages of the mean shift method and the graph cut algorithm to obtain better segmentation results, even though the scene is complex.

Details

Sensor Review, vol. 29 no. 2
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 14 July 2022

Alessandro Gambetti and Qiwei Han

The purpose of this paper is to explore and examine discrepancies of food aesthetics portrayed on social media across different types of restaurants using a large-scale data set…

1386

Abstract

Purpose

The purpose of this paper is to explore and examine discrepancies of food aesthetics portrayed on social media across different types of restaurants using a large-scale data set of food images.

Design/methodology/approach

A neural food aesthetic assessment model using computer vision and deep learning techniques is proposed, applied and evaluated on the food images data set. In addition, a set of photographic attributes drawn from food services and cognitive science research, including color, composition and figure–ground relationship attributes is implemented and compared with aesthetic scores for each food image.

Findings

This study finds that restaurants with different rating levels, cuisine types and chain status have different aesthetic scores. Moreover, the authors study the difference in the aesthetic scores between two groups of image posters: customers and restaurant owners, showing that the latter group tends to post more aesthetically appealing food images about the restaurant on social media than the former.

Practical implications

Restaurant owners may consider performing more proactive social media marketing strategies by posting high-quality food images. Likewise, social media platforms should incentivize their users to share high-quality food images.

Originality/value

The main contribution of this paper is to provide a novel methodological framework to assess the aesthetics of food images. Instead of relying on a multitude of standard attributes stemming from food photography, this method yields a unique one-take-all score, which is more straightforward to understand and more accessible to correlate with other target variables.

Details

International Journal of Contemporary Hospitality Management, vol. 34 no. 9
Type: Research Article
ISSN: 0959-6119

Keywords

Article
Publication date: 26 February 2021

Juncheng Zou

The purpose of this paper is to propose a new video prediction-based methodology to solve the manufactural occlusion problem, which causes the loss of input images and uncertain…

Abstract

Purpose

The purpose of this paper is to propose a new video prediction-based methodology to solve the manufactural occlusion problem, which causes the loss of input images and uncertain controller parameters for the robot visual servo control.

Design/methodology/approach

This paper has put forward a method that can simultaneously generate images and controller parameter increments. Then, this paper also introduced target segmentation and designed a new comprehensive loss. Finally, this paper combines offline training to generate images and online training to generate controller parameter increments.

Findings

The data set experiments to prove that this method is better than the other four methods, and it can better restore the occluded situation of the human body in six manufactural scenarios. The simulation experiment proves that it can simultaneously generate image and controller parameter variations to improve the position accuracy of tracking under occlusions in manufacture.

Originality/value

The proposed method can effectively solve the occlusion problem in visual servo control.

Details

Assembly Automation, vol. 41 no. 2
Type: Research Article
ISSN: 0144-5154

Keywords

Abstract

Details

AI in Fashion Industry
Type: Book
ISBN: 978-1-80262-633-9

Abstract

Details

AI in Fashion Industry
Type: Book
ISBN: 978-1-80262-633-9

Article
Publication date: 21 March 2018

Muditha Senanayake, Amar Raheja and Yuhan Zhang

The purpose of this paper is to develop an automated human body measurement extraction system using simple inexpensive equipment with minimum requirement of human assistance. This…

Abstract

Purpose

The purpose of this paper is to develop an automated human body measurement extraction system using simple inexpensive equipment with minimum requirement of human assistance. This research further leads to the comparison of extracted measurements to established methods to analyze the error. The extracted measurements can be used to assist the production of custom-fit apparel. This is an effort to reduce the cost of expensive 3-D body scanners and to make the system available to the user at home.

Design/methodology/approach

A single camera body measurement system is proposed, implemented, and pilot tested. This system involves a personal computer and a webcam operating within a space of controlled lighting. The system will take two images of the user, extract body silhouettes, and perform measurement extraction. The camera is automatically calibrated using the software each time of scanning considering the scanning space. The user will select a front view and a side view among the images captured, and specify the height. In this pilot study, 31 subjects were recruited and the accuracy of 8 human body measurements were compared with the manual measurements and measurements extracted from a commercial 3-D body scanner.

Findings

The system achieved reasonable measurement performance within 10 percent accuracy for seven out of the eight measurements, while four out of eight parameters obtained a performance similar to the commercial scanner. It is proved that human body measurement extraction can be done using inexpensive equipment to obtain reasonable results.

Originality/value

This study is aimed at developing a proof-of-concept for inexpensive body scanning system, with an effort to benchmark measurement accuracy, available to an average user providing the ability to acquire self-body measurements to be used to purchase custom-fit apparel. This system can potentially boost the customization of apparel and revolutionize online shopping of custom-fit apparel.

Details

International Journal of Clothing Science and Technology, vol. 30 no. 2
Type: Research Article
ISSN: 0955-6222

Keywords

Article
Publication date: 19 January 2021

BinBin Zhang, Fumin Zhang and Xinghua Qu

Laser-based measurement techniques offer various advantages over conventional measurement techniques, such as no-destructive, no-contact, fast and long measuring distance. In…

Abstract

Purpose

Laser-based measurement techniques offer various advantages over conventional measurement techniques, such as no-destructive, no-contact, fast and long measuring distance. In cooperative laser ranging systems, it’s crucial to extract center coordinates of retroreflectors to accomplish automatic measurement. To solve this problem, this paper aims to propose a novel method.

Design/methodology/approach

We propose a method using Mask RCNN (Region Convolutional Neural Network), with ResNet101 (Residual Network 101) and FPN (Feature Pyramid Network) as the backbone, to localize retroreflectors, realizing automatic recognition in different backgrounds. Compared with two other deep learning algorithms, experiments show that the recognition rate of Mask RCNN is better especially for small-scale targets. Based on this, an ellipse detection algorithm is introduced to obtain the ellipses of retroreflectors from recognized target areas. The center coordinates of retroreflectors in the camera coordinate system are obtained by using a mathematics method.

Findings

To verify the accuracy of this method, an experiment was carried out: the distance between two retroreflectors with a known distance of 1,000.109 mm was measured, with 2.596 mm root-mean-squar error, meeting the requirements of the coarse location of retroreflectors.

Research limitations/implications

The research limitations/implications are as follows: (i) As the data set only has 200 pictures, although we have used some data augmentation methods such as rotating, mirroring and cropping, there is still room for improvement in the generalization ability of detection. (ii) The ellipse detection algorithm needs to work in relatively dark conditions, as the retroreflector is made of stainless steel, which easily reflects light.

Originality/value

The originality/value of the article lies in being able to obtain center coordinates of multiple retroreflectors automatically even in a cluttered background; being able to recognize retroreflectors with different sizes, especially for small targets; meeting the recognition requirement of multiple targets in a large field of view and obtaining 3 D centers of targets by monocular model-based vision.

Details

Industrial Robot: the international journal of robotics research and application, vol. 48 no. 3
Type: Research Article
ISSN: 0143-991X

Keywords

1 – 10 of 10