Search results

1 – 10 of 169
Open Access
Article
Publication date: 29 July 2020

Mahmood Al-khassaweneh and Omar AlShorman

In the big data era, image compression is of significant importance in today’s world. Importantly, compression of large sized images is required for everyday tasks; including…

Abstract

In the big data era, image compression is of significant importance in today’s world. Importantly, compression of large sized images is required for everyday tasks; including electronic data communications and internet transactions. However, two important measures should be considered for any compression algorithm: the compression factor and the quality of the decompressed image. In this paper, we use Frei-Chen bases technique and the Modified Run Length Encoding (RLE) to compress images. The Frei-Chen bases technique is applied at the first stage in which the average subspace is applied to each 3 × 3 block. Those blocks with the highest energy are replaced by a single value that represents the average value of the pixels in the corresponding block. Even though Frei-Chen bases technique provides lossy compression, it maintains the main characteristics of the image. Additionally, the Frei-Chen bases technique enhances the compression factor, making it advantageous to use. In the second stage, RLE is applied to further increase the compression factor. The goal of using RLE is to enhance the compression factor without adding any distortion to the resultant decompressed image. Integrating RLE with Frei-Chen bases technique, as described in the proposed algorithm, ensures high quality decompressed images and high compression rate. The results of the proposed algorithms are shown to be comparable in quality and performance with other existing methods.

Details

Applied Computing and Informatics, vol. 20 no. 1/2
Type: Research Article
ISSN: 2634-1964

Keywords

Article
Publication date: 21 February 2024

Faguo Liu, Qian Zhang, Tao Yan, Bin Wang, Ying Gao, Jiaqi Hou and Feiniu Yuan

Light field images (LFIs) have gained popularity as a technology to increase the field of view (FoV) of plenoptic cameras since they can capture information about light rays with…

Abstract

Purpose

Light field images (LFIs) have gained popularity as a technology to increase the field of view (FoV) of plenoptic cameras since they can capture information about light rays with a large FoV. Wide FoV causes light field (LF) data to increase rapidly, which restricts the use of LF imaging in image processing, visual analysis and user interface. Effective LFI coding methods become of paramount importance. This paper aims to eliminate more redundancy by exploring sparsity and correlation in the angular domain of LFIs, as well as mitigate the loss of perceptual quality of LFIs caused by encoding.

Design/methodology/approach

This work proposes a new efficient LF coding framework. On the coding side, a new sampling scheme and a hierarchical prediction structure are used to eliminate redundancy in the LFI's angular and spatial domains. At the decoding side, high-quality dense LF is reconstructed using a view synthesis method based on the residual channel attention network (RCAN).

Findings

In three different LF datasets, our proposed coding framework not only reduces the transmitted bit rate but also maintains a higher view quality than the current more advanced methods.

Originality/value

(1) A new sampling scheme is designed to synthesize high-quality LFIs while better ensuring LF angular domain sparsity. (2) To further eliminate redundancy in the spatial domain, new ranking schemes and hierarchical prediction structures are designed. (3) A synthetic network based on RCAN and a novel loss function is designed to mitigate the perceptual quality loss due to the coding process.

Details

Data Technologies and Applications, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 8 July 2022

Uzair Khan, Hikmat Ullah Khan, Saqib Iqbal and Hamza Munir

Image Processing is an emerging field that is used to extract information from images. In recent years, this field has received immense attention from researchers, especially in…

Abstract

Purpose

Image Processing is an emerging field that is used to extract information from images. In recent years, this field has received immense attention from researchers, especially in the research domains of object detection, Biomedical Imaging and Semantic segmentation. In this study, a bibliometric analysis of publications related to image processing in the Science Expanded Index Extended (SCI-Expanded) has been performed. Several parameters have been analyzed such as annual scientific production, citations per article, most cited documents, top 20 articles, most relevant authors, authors evaluation using y-index, top and most relevant sources (journals) and hot topics.

Design/methodology/approach

The Bibliographic data has been extracted from the Web of Science which is well known and the world's top database of bibliographic citations of multidisciplinary areas that covers the various journals of computer science, engineering, medical and social sciences.

Findings

The research work in image processing is meager in the past decade, however, from 2014 to 2019, it increases dramatically. Recently, the IEEE Access journal is the most relevant source with an average of 115 publications per year. The USA is most productive and its publications are highly cited while China comes in second place. Image Segmentation, Feature Extraction and Medical Image Processing are hot topics in recent years. The National Natural Science Foundation of China provides 8% of all funds for Image Processing. As Image Processing is now becoming one of the most critical fields, the research productivity has enhanced during the past five years and more work is done while the era of 2005–2013 was the area with the least amount of work in this area.

Originality/value

This research is novel in this regard that no previous research focuses on Bibliometric Analysis in the Image Processing domain, which is one of the hot research areas in computer science and engineering.

Article
Publication date: 17 June 2021

Ambica Ghai, Pradeep Kumar and Samrat Gupta

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered…

1158

Abstract

Purpose

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered with to influence public opinion. Since the consumers of online information (misinformation) tend to trust the content when the image(s) supplement the text, image manipulation software is increasingly being used to forge the images. To address the crucial problem of image manipulation, this study focusses on developing a deep-learning-based image forgery detection framework.

Design/methodology/approach

The proposed deep-learning-based framework aims to detect images forged using copy-move and splicing techniques. The image transformation technique aids the identification of relevant features for the network to train effectively. After that, the pre-trained customized convolutional neural network is used to train on the public benchmark datasets, and the performance is evaluated on the test dataset using various parameters.

Findings

The comparative analysis of image transformation techniques and experiments conducted on benchmark datasets from a variety of socio-cultural domains establishes the effectiveness and viability of the proposed framework. These findings affirm the potential applicability of proposed framework in real-time image forgery detection.

Research limitations/implications

This study bears implications for several important aspects of research on image forgery detection. First this research adds to recent discussion on feature extraction and learning for image forgery detection. While prior research on image forgery detection, hand-crafted the features, the proposed solution contributes to stream of literature that automatically learns the features and classify the images. Second, this research contributes to ongoing effort in curtailing the spread of misinformation using images. The extant literature on spread of misinformation has prominently focussed on textual data shared over social media platforms. The study addresses the call for greater emphasis on the development of robust image transformation techniques.

Practical implications

This study carries important practical implications for various domains such as forensic sciences, media and journalism where image data is increasingly being used to make inferences. The integration of image forgery detection tools can be helpful in determining the credibility of the article or post before it is shared over the Internet. The content shared over the Internet by the users has become an important component of news reporting. The framework proposed in this paper can be further extended and trained on more annotated real-world data so as to function as a tool for fact-checkers.

Social implications

In the current scenario wherein most of the image forgery detection studies attempt to assess whether the image is real or forged in an offline mode, it is crucial to identify any trending or potential forged image as early as possible. By learning from historical data, the proposed framework can aid in early prediction of forged images to detect the newly emerging forged images even before they occur. In summary, the proposed framework has a potential to mitigate physical spreading and psychological impact of forged images on social media.

Originality/value

This study focusses on copy-move and splicing techniques while integrating transfer learning concepts to classify forged images with high accuracy. The synergistic use of hitherto little explored image transformation techniques and customized convolutional neural network helps design a robust image forgery detection framework. Experiments and findings establish that the proposed framework accurately classifies forged images, thus mitigating the negative socio-cultural spread of misinformation.

Details

Information Technology & People, vol. 37 no. 2
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 28 February 2023

Tulsi Pawan Fowdur, M.A.N. Shaikh Abdoolla and Lokeshwar Doobur

The purpose of this paper is to perform a comparative analysis of the delay associated in running two real-time machine learning-based applications, namely, a video quality…

Abstract

Purpose

The purpose of this paper is to perform a comparative analysis of the delay associated in running two real-time machine learning-based applications, namely, a video quality assessment (VQA) and a phishing detection application by using the edge, fog and cloud computing paradigms.

Design/methodology/approach

The VQA algorithm was developed using Android Studio and run on a mobile phone for the edge paradigm. For the fog paradigm, it was hosted on a Java server and for the cloud paradigm on the IBM and Firebase clouds. The phishing detection algorithm was embedded into a browser extension for the edge paradigm. For the fog paradigm, it was hosted on a Node.js server and for the cloud paradigm on Firebase.

Findings

For the VQA algorithm, the edge paradigm had the highest response time while the cloud paradigm had the lowest, as the algorithm was computationally intensive. For the phishing detection algorithm, the edge paradigm had the lowest response time, and the cloud paradigm had the highest, as the algorithm had a low computational complexity. Since the determining factor for the response time was the latency, the edge paradigm provided the smallest delay as all processing were local.

Research limitations/implications

The main limitation of this work is that the experiments were performed on a small scale due to time and budget constraints.

Originality/value

A detailed analysis with real applications has been provided to show how the complexity of an application can determine the best computing paradigm on which it can be deployed.

Details

International Journal of Pervasive Computing and Communications, vol. 20 no. 1
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 19 March 2024

Cemalettin Akdoğan, Tolga Özer and Yüksel Oğuz

Nowadays, food problems are likely to arise because of the increasing global population and decreasing arable land. Therefore, it is necessary to increase the yield of…

Abstract

Purpose

Nowadays, food problems are likely to arise because of the increasing global population and decreasing arable land. Therefore, it is necessary to increase the yield of agricultural products. Pesticides can be used to improve agricultural land products. This study aims to make the spraying of cherry trees more effective and efficient with the designed artificial intelligence (AI)-based agricultural unmanned aerial vehicle (UAV).

Design/methodology/approach

Two approaches have been adopted for the AI-based detection of cherry trees: In approach 1, YOLOv5, YOLOv7 and YOLOv8 models are trained with 70, 100 and 150 epochs. In Approach 2, a new method is proposed to improve the performance metrics obtained in Approach 1. Gaussian, wavelet transform (WT) and Histogram Equalization (HE) preprocessing techniques were applied to the generated data set in Approach 2. The best-performing models in Approach 1 and Approach 2 were used in the real-time test application with the developed agricultural UAV.

Findings

In Approach 1, the best F1 score was 98% in 100 epochs with the YOLOv5s model. In Approach 2, the best F1 score and mAP values were obtained as 98.6% and 98.9% in 150 epochs, with the YOLOv5m model with an improvement of 0.6% in the F1 score. In real-time tests, the AI-based spraying drone system detected and sprayed cherry trees with an accuracy of 66% in Approach 1 and 77% in Approach 2. It was revealed that the use of pesticides could be reduced by 53% and the energy consumption of the spraying system by 47%.

Originality/value

An original data set was created by designing an agricultural drone to detect and spray cherry trees using AI. YOLOv5, YOLOv7 and YOLOv8 models were used to detect and classify cherry trees. The results of the performance metrics of the models are compared. In Approach 2, a method including HE, Gaussian and WT is proposed, and the performance metrics are improved. The effect of the proposed method in a real-time experimental application is thoroughly analyzed.

Details

Robotic Intelligence and Automation, vol. 44 no. 1
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 22 January 2024

Sann Ryu

This study aims to examine the visual effects of cause-related marketing (CM) posts on Instagram, with a focus on image resolution and consumer engagement.

Abstract

Purpose

This study aims to examine the visual effects of cause-related marketing (CM) posts on Instagram, with a focus on image resolution and consumer engagement.

Design/methodology/approach

Three studies were conducted through an experimental design. Study 1 (N = 155) uncovered the mediation underlying the effects of image quality (low and high image resolution). Study 2 (N = 160) replicated the findings of the first study and extended the investigation by examining the mediator (fluency) and moderator (visual sensitivity). Study 3 (N = 291) further extended the effects of image resolution by demonstrating its interactive effects with the visual complexity of an Instagram post design in a 2 × 2 factorial experiment.

Findings

The serial mediation analysis demonstrated that high image resolution CM posts yielded more favorable evaluations in terms of brand credibility and information costs saved, subsequently leading to positive brand attitudes, purchase intentions and increased Instagram engagement. Processing fluency mediated image effects on brand credibility, while individual differences in visual sensitivity moderated the image effects. The image resolution effects were greater for visually complex CM posts compared to simple ones.

Originality/value

To one's best knowledge, little to no research has examined the image quality of Instagram posts in the context of CM and the extent to which such visual cues can affect consumers' brand evaluations and engagement on the platform.

Research implications

Despite its practical significance, there exists a notable gap in understanding the specific role of CM posts on Instagram and the impact of visual elements on consumer behaviors. The current research findings aim to bridge the research gap.

Details

Journal of Research in Interactive Marketing, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2040-7122

Keywords

Article
Publication date: 31 August 2023

Hongwei Zhang, Shihao Wang, Hongmin Mi, Shuai Lu, Le Yao and Zhiqiang Ge

The defect detection problem of color-patterned fabric is still a huge challenge due to the lack of manual defect labeling samples. Recently, many fabric defect detection…

115

Abstract

Purpose

The defect detection problem of color-patterned fabric is still a huge challenge due to the lack of manual defect labeling samples. Recently, many fabric defect detection algorithms based on feature engineering and deep learning have been proposed, but these methods have overdetection or miss-detection problems because they cannot adapt to the complex patterns of color-patterned fabrics. The purpose of this paper is to propose a defect detection framework based on unsupervised adversarial learning for image reconstruction to solve the above problems.

Design/methodology/approach

The proposed framework consists of three parts: a generator, a discriminator and an image postprocessing module. The generator is able to extract the features of the image and then reconstruct the image. The discriminator can supervise the generator to repair defects in the samples to improve the quality of image reconstruction. The multidifference image postprocessing module is used to obtain the final detection results of color-patterned fabric defects.

Findings

The proposed framework is compared with state-of-the-art methods on the public dataset YDFID-1(Yarn-Dyed Fabric Image Dataset-version1). The proposed framework is also validated on several classes in the MvTec AD dataset. The experimental results of various patterns/classes on YDFID-1 and MvTecAD demonstrate the effectiveness and superiority of this method in fabric defect detection.

Originality/value

It provides an automatic defect detection solution that is convenient for engineering applications for the inspection process of the color-patterned fabric manufacturing industry. A public dataset is provided for academia.

Details

International Journal of Clothing Science and Technology, vol. 35 no. 6
Type: Research Article
ISSN: 0955-6222

Keywords

Article
Publication date: 29 September 2021

Swetha Parvatha Reddy Chandrasekhara, Mohan G. Kabadi and Srivinay

This study has mainly aimed to compare and contrast two completely different image processing algorithms that are very adaptive for detecting prostate cancer using wearable…

Abstract

Purpose

This study has mainly aimed to compare and contrast two completely different image processing algorithms that are very adaptive for detecting prostate cancer using wearable Internet of Things (IoT) devices. Cancer in these modern times is still considered as one of the most dreaded disease, which is continuously pestering the mankind over a past few decades. According to Indian Council of Medical Research, India alone registers about 11.5 lakh cancer related cases every year and closely up to 8 lakh people die with cancer related issues each year. Earlier the incidence of prostate cancer was commonly seen in men aged above 60 years, but a recent study has revealed that this type of cancer has been on rise even in men between the age groups of 35 and 60 years as well. These findings make it even more necessary to prioritize the research on diagnosing the prostate cancer at an early stage, so that the patients can be cured and can lead a normal life.

Design/methodology/approach

The research focuses on two types of feature extraction algorithms, namely, scale invariant feature transform (SIFT) and gray level co-occurrence matrix (GLCM) that are commonly used in medical image processing, in an attempt to discover and improve the gap present in the potential detection of prostate cancer in medical IoT. Later the results obtained by these two strategies are classified separately using a machine learning based classification model called multi-class support vector machine (SVM). Owing to the advantage of better tissue discrimination and contrast resolution, magnetic resonance imaging images have been considered for this study. The classification results obtained for both the SIFT as well as GLCM methods are then compared to check, which feature extraction strategy provides the most accurate results for diagnosing the prostate cancer.

Findings

The potential of both the models has been evaluated in terms of three aspects, namely, accuracy, sensitivity and specificity. Each model’s result was checked against diversified ranges of training and test data set. It was found that the SIFT-multiclass SVM model achieved a highest performance rate of 99.9451% accuracy, 100% sensitivity and 99% specificity at 40:60 ratio of the training and testing data set.

Originality/value

The SIFT-multi SVM versus GLCM-multi SVM based comparison has been introduced for the first time to perceive the best model to be used for the accurate diagnosis of prostate cancer. The performance of the classification for each of the feature extraction strategies is enumerated in terms of accuracy, sensitivity and specificity.

Details

International Journal of Pervasive Computing and Communications, vol. 20 no. 1
Type: Research Article
ISSN: 1742-7371

Keywords

Open Access
Article
Publication date: 1 December 2023

Francois Du Rand, André Francois van der Merwe and Malan van Tonder

This paper aims to discuss the development of a defect classification system that can be used to detect and classify powder bed surface defects from captured layer images without…

Abstract

Purpose

This paper aims to discuss the development of a defect classification system that can be used to detect and classify powder bed surface defects from captured layer images without the need for specialised computational hardware. The idea is to develop this system by making use of more traditional machine learning (ML) models instead of using computationally intensive deep learning (DL) models.

Design/methodology/approach

The approach that is used by this study is to use traditional image processing and classification techniques that can be applied to captured layer images to detect and classify defects without the need for DL algorithms.

Findings

The study proved that a defect classification algorithm could be developed by making use of traditional ML models with a high degree of accuracy and the images could be processed at higher speeds than typically reported in literature when making use of DL models.

Originality/value

This paper addresses a need that has been identified for a high-speed defect classification algorithm that can detect and classify defects without the need for specialised hardware that is typically used when making use of DL technologies. This is because when developing closed-loop feedback systems for these additive manufacturing machines, it is important to detect and classify defects without inducing additional delays to the control system.

Details

Rapid Prototyping Journal, vol. 29 no. 11
Type: Research Article
ISSN: 1355-2546

Keywords

1 – 10 of 169