Search results

1 – 10 of 98
Article
Publication date: 8 September 2022

Johnny Kwok Wai Wong, Mojtaba Maghrebi, Alireza Ahmadian Fard Fini, Mohammad Amin Alizadeh Golestani, Mahdi Ahmadnia and Michael Er

Images taken from construction site interiors often suffer from low illumination and poor natural colors, which restrict their application for high-level site management purposes…

Abstract

Purpose

Images taken from construction site interiors often suffer from low illumination and poor natural colors, which restrict their application for high-level site management purposes. The state-of-the-art low-light image enhancement method provides promising image enhancement results. However, they generally require a longer execution time to complete the enhancement. This study aims to develop a refined image enhancement approach to improve execution efficiency and performance accuracy.

Design/methodology/approach

To develop the refined illumination enhancement algorithm named enhanced illumination quality (EIQ), a quadratic expression was first added to the initial illumination map. Subsequently, an adjusted weight matrix was added to improve the smoothness of the illumination map. A coordinated descent optimization algorithm was then applied to minimize the processing time. Gamma correction was also applied to further enhance the illumination map. Finally, a frame comparing and averaging method was used to identify interior site progress.

Findings

The proposed refined approach took around 4.36–4.52 s to achieve the expected results while outperforming the current low-light image enhancement method. EIQ demonstrated a lower lightness-order error and provided higher object resolution in enhanced images. EIQ also has a higher structural similarity index and peak-signal-to-noise ratio, which indicated better image reconstruction performance.

Originality/value

The proposed approach provides an alternative to shorten the execution time, improve equalization of the illumination map and provide a better image reconstruction. The approach could be applied to low-light video enhancement tasks and other dark or poor jobsite images for object detection processes.

Details

Construction Innovation , vol. 24 no. 2
Type: Research Article
ISSN: 1471-4175

Keywords

Article
Publication date: 1 November 2023

Juan Yang, Zhenkun Li and Xu Du

Although numerous signal modalities are available for emotion recognition, audio and visual modalities are the most common and predominant forms for human beings to express their…

Abstract

Purpose

Although numerous signal modalities are available for emotion recognition, audio and visual modalities are the most common and predominant forms for human beings to express their emotional states in daily communication. Therefore, how to achieve automatic and accurate audiovisual emotion recognition is significantly important for developing engaging and empathetic human–computer interaction environment. However, two major challenges exist in the field of audiovisual emotion recognition: (1) how to effectively capture representations of each single modality and eliminate redundant features and (2) how to efficiently integrate information from these two modalities to generate discriminative representations.

Design/methodology/approach

A novel key-frame extraction-based attention fusion network (KE-AFN) is proposed for audiovisual emotion recognition. KE-AFN attempts to integrate key-frame extraction with multimodal interaction and fusion to enhance audiovisual representations and reduce redundant computation, filling the research gaps of existing approaches. Specifically, the local maximum–based content analysis is designed to extract key-frames from videos for the purpose of eliminating data redundancy. Two modules, including “Multi-head Attention-based Intra-modality Interaction Module” and “Multi-head Attention-based Cross-modality Interaction Module”, are proposed to mine and capture intra- and cross-modality interactions for further reducing data redundancy and producing more powerful multimodal representations.

Findings

Extensive experiments on two benchmark datasets (i.e. RAVDESS and CMU-MOSEI) demonstrate the effectiveness and rationality of KE-AFN. Specifically, (1) KE-AFN is superior to state-of-the-art baselines for audiovisual emotion recognition. (2) Exploring the supplementary and complementary information of different modalities can provide more emotional clues for better emotion recognition. (3) The proposed key-frame extraction strategy can enhance the performance by more than 2.79 per cent on accuracy. (4) Both exploring intra- and cross-modality interactions and employing attention-based audiovisual fusion can lead to better prediction performance.

Originality/value

The proposed KE-AFN can support the development of engaging and empathetic human–computer interaction environment.

Open Access
Article
Publication date: 22 September 2023

Nengsheng Bao, Yuchen Fan, Chaoping Li and Alessandro Simeone

Lubricating oil leakage is a common issue in thermal power plant operation sites, requiring prompt equipment maintenance. The real-time detection of leakage occurrences could…

Abstract

Purpose

Lubricating oil leakage is a common issue in thermal power plant operation sites, requiring prompt equipment maintenance. The real-time detection of leakage occurrences could avoid disruptive consequences caused by the lack of timely maintenance. Currently, inspection operations are mostly carried out manually, resulting in time-consuming processes prone to health and safety hazards. To overcome such issues, this paper proposes a machine vision-based inspection system aimed at automating the oil leakage detection for improving the maintenance procedures.

Design/methodology/approach

The approach aims at developing a novel modular-structured automatic inspection system. The image acquisition module collects digital images along a predefined inspection path using a dual-light (i.e. ultraviolet and blue light) illumination system, deploying the fluorescence of the lubricating oil while suppressing unwanted background noise. The image processing module is designed to detect the oil leakage within the digital images minimizing detection errors. A case study is reported to validate the industrial suitability of the proposed inspection system.

Findings

On-site experimental results demonstrate the capabilities to complete the automatic inspection procedures of the tested industrial equipment by achieving an oil leakage detection accuracy up to 99.13%.

Practical implications

The proposed inspection system can be adopted in industrial context to detect lubricant leakage ensuring the equipment and the operators safety.

Originality/value

The proposed inspection system adopts a computer vision approach, which deploys the combination of two separate sources of light, to boost the detection capabilities, enabling the application for a variety of particularly hard-to-inspect industrial contexts.

Details

Journal of Quality in Maintenance Engineering, vol. 29 no. 5
Type: Research Article
ISSN: 1355-2511

Keywords

Article
Publication date: 24 July 2024

Arthur de Carvalho Cruzeiro, Leonardo Santana, Danay Manzo Jaime, Sílvia Ramoa, Jorge Lino Alves and Guilherme Mariz de Oliveira Barra

This study aims to evaluate in situ oxidative polymerization of aniline (Ani) as a post-processing method to promote extrusion-based 3D printed parts, made from insulating…

Abstract

Purpose

This study aims to evaluate in situ oxidative polymerization of aniline (Ani) as a post-processing method to promote extrusion-based 3D printed parts, made from insulating polymers, to components with functional properties, including electrical conductivity and chemical sensitivity.

Design/methodology/approach

Extrusion-based 3D printed parts of polyethylene terephthalate modified with glycol (PETG) and polypropylene (PP) were coated in an aqueous acid solution via in situ oxidative polymerization of Ani. First, the feedstocks were characterized. Densely printed samples were then used to assess the adhesion of polyaniline (PAni) and electrical conductivity on printed parts. The best feedstock candidate for PAni coating was selected for further analysis. Last, a Taguchi methodology was used to evaluate the influence of printing parameters on the coating of porous samples. Analysis of variance and Tukey post hoc test were used to identify the best levels for each parameter.

Findings

Colorimetry measurements showed significant color shifts in PP samples and no shifts in PETG samples upon pullout testing. The incorporation of PAni content and electrical conductivity were, respectively, 41% and 571% higher for PETG in comparison to PP. Upon coating, the surface energy of both materials decreased. Additionally, the dynamic mechanical analysis test showed minimal influence of PAni over the dynamic mechanical properties of PETG. The parametric study indicated that only layer thickness and infill pattern had a significant influence on PAni incorporation and electrical conductivity of coated porous samples.

Originality/value

Current literature reports difficulties in incorporating PAni without affecting dimensional precision and feedstock stability. In situ, oxidative polymerization of Ani could overcome these limitations. However, its use as a functional post-processing of extrusion-based printed parts is a novelty.

Details

Rapid Prototyping Journal, vol. 30 no. 8
Type: Research Article
ISSN: 1355-2546

Keywords

Article
Publication date: 17 June 2021

Ambica Ghai, Pradeep Kumar and Samrat Gupta

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered…

1396

Abstract

Purpose

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered with to influence public opinion. Since the consumers of online information (misinformation) tend to trust the content when the image(s) supplement the text, image manipulation software is increasingly being used to forge the images. To address the crucial problem of image manipulation, this study focusses on developing a deep-learning-based image forgery detection framework.

Design/methodology/approach

The proposed deep-learning-based framework aims to detect images forged using copy-move and splicing techniques. The image transformation technique aids the identification of relevant features for the network to train effectively. After that, the pre-trained customized convolutional neural network is used to train on the public benchmark datasets, and the performance is evaluated on the test dataset using various parameters.

Findings

The comparative analysis of image transformation techniques and experiments conducted on benchmark datasets from a variety of socio-cultural domains establishes the effectiveness and viability of the proposed framework. These findings affirm the potential applicability of proposed framework in real-time image forgery detection.

Research limitations/implications

This study bears implications for several important aspects of research on image forgery detection. First this research adds to recent discussion on feature extraction and learning for image forgery detection. While prior research on image forgery detection, hand-crafted the features, the proposed solution contributes to stream of literature that automatically learns the features and classify the images. Second, this research contributes to ongoing effort in curtailing the spread of misinformation using images. The extant literature on spread of misinformation has prominently focussed on textual data shared over social media platforms. The study addresses the call for greater emphasis on the development of robust image transformation techniques.

Practical implications

This study carries important practical implications for various domains such as forensic sciences, media and journalism where image data is increasingly being used to make inferences. The integration of image forgery detection tools can be helpful in determining the credibility of the article or post before it is shared over the Internet. The content shared over the Internet by the users has become an important component of news reporting. The framework proposed in this paper can be further extended and trained on more annotated real-world data so as to function as a tool for fact-checkers.

Social implications

In the current scenario wherein most of the image forgery detection studies attempt to assess whether the image is real or forged in an offline mode, it is crucial to identify any trending or potential forged image as early as possible. By learning from historical data, the proposed framework can aid in early prediction of forged images to detect the newly emerging forged images even before they occur. In summary, the proposed framework has a potential to mitigate physical spreading and psychological impact of forged images on social media.

Originality/value

This study focusses on copy-move and splicing techniques while integrating transfer learning concepts to classify forged images with high accuracy. The synergistic use of hitherto little explored image transformation techniques and customized convolutional neural network helps design a robust image forgery detection framework. Experiments and findings establish that the proposed framework accurately classifies forged images, thus mitigating the negative socio-cultural spread of misinformation.

Details

Information Technology & People, vol. 37 no. 2
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 29 September 2023

Wen-Qian Lou, Bin Wu and Bo-Wen Zhu

This study aims to clarify influencing factors of overcapacity of new energy enterprises in China and accurately predict whether these enterprises have overcapacity.

122

Abstract

Purpose

This study aims to clarify influencing factors of overcapacity of new energy enterprises in China and accurately predict whether these enterprises have overcapacity.

Design/methodology/approach

Based on relevant data including the experience and evidence from the capital market in China, the research establishes a generic univariate selection-comparative machine learning model to study relevant factors that affect overcapacity of new energy enterprises from five dimensions. These include the governmental intervention, market demand, corporate finance, corporate governance and corporate decision. Moreover, the bridging approach is used to strengthen findings from quantitative studies via the results from qualitative studies.

Findings

The authors' results show that the overcapacity of new energy enterprises in China is brought out by the combined effect of governmental intervention corporate governance and corporate decision. Governmental interventions increase the overcapacity risk of new energy enterprises mainly by distorting investment behaviors of enterprises. Corporate decision and corporate governance factors affect the overcapacity mainly by regulating the degree of overconfidence of the management team and the agency cost. Among the eight comparable integrated models, generic univariate selection-bagging exhibits the optimal comprehensive generalization performance and its area under the receiver operating characteristic curve Area under curve (AUC) accuracy precision and recall are 0.719, 0.960, 0.975 and 0.983, respectively.

Originality/value

The proposed integrated model analyzes causes and predicts presence of overcapacity of new energy enterprises to help governments to formulate appropriate strategies to deal with overcapacity and new energy enterprises to optimize resource allocation. Ten main features which affect the overcapacity of new energy enterprises in China are identified through generic univariate selection model. Through the bridging approach, the impact of the main features on the overcapacity of new energy enterprises and the mechanism of the influence are analyzed.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 8 April 2024

Hu Luo, Haobin Ruan and Dawei Tu

The purpose of this paper is to propose a whole set of methods for underwater target detection, because most underwater objects have small samples, low quality underwater images…

Abstract

Purpose

The purpose of this paper is to propose a whole set of methods for underwater target detection, because most underwater objects have small samples, low quality underwater images problems such as detail loss, low contrast and color distortion, and verify the feasibility of the proposed methods through experiments.

Design/methodology/approach

The improved RGHS algorithm to enhance the original underwater target image is proposed, and then the YOLOv4 deep learning network for underwater small sample targets detection is improved based on the combination of traditional data expansion method and Mosaic algorithm, expanding the feature extraction capability with SPP (Spatial Pyramid Pooling) module after each feature extraction layer to extract richer feature information.

Findings

The experimental results, using the official dataset, reveal a 3.5% increase in average detection accuracy for three types of underwater biological targets compared to the traditional YOLOv4 algorithm. In underwater robot application testing, the proposed method achieves an impressive 94.73% average detection accuracy for the three types of underwater biological targets.

Originality/value

Underwater target detection is an important task for underwater robot application. However, most underwater targets have the characteristics of small samples, and the detection of small sample targets is a comprehensive problem because it is affected by the quality of underwater images. This paper provides a whole set of methods to solve the problems, which is of great significance to the application of underwater robot.

Details

Robotic Intelligence and Automation, vol. 44 no. 2
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 27 August 2024

Vali Dalouji and Nasim Rahimi

The purpose of this paper is to study the correlation between the thicknesses of the C–Ni films that have been prepared by RF-magnetron sputtering on quartz substrates and their…

Abstract

Purpose

The purpose of this paper is to study the correlation between the thicknesses of the C–Ni films that have been prepared by RF-magnetron sputtering on quartz substrates and their three-dimensional (3D) micro morphology. In this work by AFM images, this paper studied stereo metric analysis of these films.

Design/methodology/approach

The C–Ni films have been prepared by RF-magnetron sputtering on quartz substrates using a mosaic target consisting of pure graphite and strips of pure nickel approximately 2 cm2 attached to the graphite race track. The field emission scanning electronic microscopy (FESEM) images were used for the morphological characterization.

Findings

The histogram peaks are zero for all samples and the histograms are almost symmetric around zero. Temperature did not have much effect on the degree of isolation, so all four diagrams have similar results. The qualitative observations through statistical parameters of the 3D surface texture revealed that the smoothest surface has been obtained for C-Ni films annealed at 500 °C (Sa, Sq, Sz and Sv have the lower values), while the most irregular topography has been found for C-Ni films annealed at 300 °C (the fractal dimension D = 2.01 ± 0.131).

Originality/value

As shown in FESEM images, the size of the particles was increased for films deposited from 300 ºC to 800ºC; however, at 1000ºC, it decreased significantly. The histogram peaks are zero for all samples and the histograms were almost symmetric around zero. Also, the largest and lowest root mean heights (Sq) belong to films at 300 °C and 500 °C. Furthermore, the more irregular surface was found at 300 °C, and the more regular surface was found at 500 °C. As the temperature was increased to 800 °C, the values of the IAPSD function increased systematically, and then the values of the IAPSD function was decreased in the fourth sample. The surface skewness of samples annealed at 1000 °C was positive which confirms the lack of dominance of cavities on their surface with the highest amount of C-Ni films at 800 °C.

Details

Soldering & Surface Mount Technology, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0954-0911

Keywords

Article
Publication date: 14 March 2024

Qiang Wen, Lele Chen, Jingwen Jin, Jianhao Huang and HeLin Wan

Fixed mode noise and random mode noise always exist in the image sensor, which affects the imaging quality of the image sensor. The charge diffusion and color mixing between…

Abstract

Purpose

Fixed mode noise and random mode noise always exist in the image sensor, which affects the imaging quality of the image sensor. The charge diffusion and color mixing between pixels in the photoelectric conversion process belong to fixed mode noise. This study aims to improve the image sensor imaging quality by processing the fixed mode noise.

Design/methodology/approach

Through an iterative training of an ergoable long- and short-term memory recurrent neural network model, the authors obtain a neural network model able to compensate for image noise crosstalk. To overcome the lack of differences in the same color pixels on each template of the image sensor under flat-field light, the data before and after compensation were used as a new data set to further train the neural network iteratively.

Findings

The comparison of the images compensated by the two sets of neural network models shows that the gray value distribution is more concentrated and uniform. The middle and high frequency components in the spatial spectrum are all increased, indicating that the compensated image edges change faster and are more detailed (Hinton and Salakhutdinov, 2006; LeCun et al., 1998; Mohanty et al., 2016; Zang et al., 2023).

Originality/value

In this paper, the authors use the iterative learning color image pixel crosstalk compensation method to effectively alleviate the incomplete color mixing problem caused by the insufficient filter rate and the electric crosstalk problem caused by the lateral diffusion of the optical charge caused by the adjacent pixel potential trap.

Details

Sensor Review, vol. 44 no. 2
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 19 December 2023

Jinchao Huang

Single-shot multi-category clothing recognition and retrieval play a crucial role in online searching and offline settlement scenarios. Existing clothing recognition methods based…

Abstract

Purpose

Single-shot multi-category clothing recognition and retrieval play a crucial role in online searching and offline settlement scenarios. Existing clothing recognition methods based on RGBD clothing images often suffer from high-dimensional feature representations, leading to compromised performance and efficiency.

Design/methodology/approach

To address this issue, this paper proposes a novel method called Manifold Embedded Discriminative Feature Selection (MEDFS) to select global and local features, thereby reducing the dimensionality of the feature representation and improving performance. Specifically, by combining three global features and three local features, a low-dimensional embedding is constructed to capture the correlations between features and categories. The MEDFS method designs an optimization framework utilizing manifold mapping and sparse regularization to achieve feature selection. The optimization objective is solved using an alternating iterative strategy, ensuring convergence.

Findings

Empirical studies conducted on a publicly available RGBD clothing image dataset demonstrate that the proposed MEDFS method achieves highly competitive clothing classification performance while maintaining efficiency in clothing recognition and retrieval.

Originality/value

This paper introduces a novel approach for multi-category clothing recognition and retrieval, incorporating the selection of global and local features. The proposed method holds potential for practical applications in real-world clothing scenarios.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 17 no. 2
Type: Research Article
ISSN: 1756-378X

Keywords

1 – 10 of 98