Search results

1 – 10 of 318
Article
Publication date: 2 January 2023

Enbo Li, Haibo Feng and Yili Fu

The grasping task of robots in dense cluttered scenes from a single-view has not been solved perfectly, and there is still a problem of low grasping success rate. This study aims…

Abstract

Purpose

The grasping task of robots in dense cluttered scenes from a single-view has not been solved perfectly, and there is still a problem of low grasping success rate. This study aims to propose an end-to-end grasp generation method to solve this problem.

Design/methodology/approach

A new grasp representation method is proposed, which cleverly uses the normal vector of the table surface to derive the grasp baseline vectors, and maps the grasps to the pointed points (PP), so that there is no need to add orthogonal constraints between vectors when using a neural network to predict rotation matrixes of grasps.

Findings

Experimental results show that the proposed method is beneficial to the training of the neural network, and the model trained on synthetic data set can also have high grasping success rate and completion rate in real-world tasks.

Originality/value

The main contribution of this paper is that the authors propose a new grasp representation method, which maps the 6-DoF grasps to a PP and an angle related to the tabletop normal vector, thereby eliminating the need to add orthogonal constraints between vectors when directly predicting grasps using neural networks. The proposed method can generate hundreds of grasps covering the whole surface in about 0.3 s. The experimental results show that the proposed method has obvious superiority compared with other methods.

Details

Industrial Robot: the international journal of robotics research and application, vol. 50 no. 3
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 13 March 2024

Rong Jiang, Bin He, Zhipeng Wang, Xu Cheng, Hongrui Sang and Yanmin Zhou

Compared with traditional methods relying on manual teaching or system modeling, data-driven learning methods, such as deep reinforcement learning and imitation learning, show…

Abstract

Purpose

Compared with traditional methods relying on manual teaching or system modeling, data-driven learning methods, such as deep reinforcement learning and imitation learning, show more promising potential to cope with the challenges brought by increasingly complex tasks and environments, which have become the hot research topic in the field of robot skill learning. However, the contradiction between the difficulty of collecting robot–environment interaction data and the low data efficiency causes all these methods to face a serious data dilemma, which has become one of the key issues restricting their development. Therefore, this paper aims to comprehensively sort out and analyze the cause and solutions for the data dilemma in robot skill learning.

Design/methodology/approach

First, this review analyzes the causes of the data dilemma based on the classification and comparison of data-driven methods for robot skill learning; Then, the existing methods used to solve the data dilemma are introduced in detail. Finally, this review discusses the remaining open challenges and promising research topics for solving the data dilemma in the future.

Findings

This review shows that simulation–reality combination, state representation learning and knowledge sharing are crucial for overcoming the data dilemma of robot skill learning.

Originality/value

To the best of the authors’ knowledge, there are no surveys that systematically and comprehensively sort out and analyze the data dilemma in robot skill learning in the existing literature. It is hoped that this review can be helpful to better address the data dilemma in robot skill learning in the future.

Details

Robotic Intelligence and Automation, vol. 44 no. 2
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 16 August 2021

Shilpa Gite, Ketan Kotecha and Gheorghita Ghinea

This study aims to analyze driver risks in the driving environment. A complete analysis of context aware assistive driving techniques. Context awareness in assistive driving by…

287

Abstract

Purpose

This study aims to analyze driver risks in the driving environment. A complete analysis of context aware assistive driving techniques. Context awareness in assistive driving by probabilistic modeling techniques. Advanced techniques using Spatio-temporal techniques, computer vision and deep learning techniques.

Design/methodology/approach

Autonomous vehicles have been aimed to increase driver safety by introducing vehicle control from the driver to Advanced Driver Assistance Systems (ADAS). The core objective of these systems is to cut down on road accidents by helping the user in various ways. Early anticipation of a particular action would give a prior benefit to the driver to successfully handle the dangers on the road. In this paper, the advancements that have taken place in the use of multi-modal machine learning for assistive driving systems are surveyed. The aim is to help elucidate the recent progress and techniques in the field while also identifying the scope for further research and improvement. The authors take an overview of context-aware driver assistance systems that alert drivers in case of maneuvers by taking advantage of multi-modal human processing to better safety and drivability.

Findings

There has been a huge improvement and investment in ADAS being a key concept for road safety. In such applications, data is processed and information is extracted from multiple data sources, thus requiring training of machine learning algorithms in a multi-modal style. The domain is fast gaining traction owing to its applications across multiple disciplines with crucial gains.

Research limitations/implications

The research is focused on deep learning and computer vision-based techniques to generate a context for assistive driving and it would definitely adopt by the ADAS manufacturers.

Social implications

As context-aware assistive driving would work in real-time and it would save the lives of many drivers, pedestrians.

Originality/value

This paper provides an understanding of context-aware deep learning frameworks for assistive driving. The research is mainly focused on deep learning and computer vision-based techniques to generate a context for assistive driving. It incorporates the latest state-of-the-art techniques using suitable driving context and the driver is alerted. Many automobile manufacturing companies and researchers would refer to this study for their enhancements.

Details

International Journal of Pervasive Computing and Communications, vol. 19 no. 3
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 2 March 2022

Yanli Fan and Liyan Liu

Deep learning (DL) technology is used to design a voice evaluation system to understand the impact of learning aids on DL and mobile platforms on students’ learning behavior.

Abstract

Purpose

Deep learning (DL) technology is used to design a voice evaluation system to understand the impact of learning aids on DL and mobile platforms on students’ learning behavior.

Design/methodology/approach

DL technology is used to design a speech evaluation system.

Findings

The experimental results show that the speech evaluation system designed has a high accuracy rate, the highest agreement rate with manual evaluation of pronunciation is 89.5%, and the correct speech recognition rate is 96.64%. The designed voice evaluation system and the manual voice rating system have a maximum error rate of 2%. The experimental results suggest that it is necessary to further optimize the learning aids for mobile platform. The learning aids of the mobile platform need to be further optimized to promote the improvement of student learning efficiency.

Originality/value

The results show that the speech evaluation system designed has good practical application value, and it provides a certain reference value for the future study of learning tools on DL.

Details

Library Hi Tech, vol. 41 no. 5
Type: Research Article
ISSN: 0737-8831

Keywords

Open Access
Article
Publication date: 15 August 2023

Doreen Nkirote Bundi

The purpose of this study is to examine the state of research into adoption of machine learning systems within the health sector, to identify themes that have been studied and…

1125

Abstract

Purpose

The purpose of this study is to examine the state of research into adoption of machine learning systems within the health sector, to identify themes that have been studied and observe the important gaps in the literature that can inform a research agenda going forward.

Design/methodology/approach

A systematic literature strategy was utilized to identify and analyze scientific papers between 2012 and 2022. A total of 28 articles were identified and reviewed.

Findings

The outcomes reveal that while advances in machine learning have the potential to improve service access and delivery, there have been sporadic growth of literature in this area which is perhaps surprising given the immense potential of machine learning within the health sector. The findings further reveal that themes such as recordkeeping, drugs development and streamlining of treatment have primarily been focused on by the majority of authors in this area.

Research limitations/implications

The search was limited to journal articles published in English, resulting in the exclusion of studies disseminated through alternative channels, such as conferences, and those published in languages other than English. Considering that scholars in developing nations may encounter less difficulty in disseminating their work through alternative channels and that numerous emerging nations employ languages other than English, it is plausible that certain research has been overlooked in the present investigation.

Originality/value

This review provides insights into future research avenues for theory, content and context on adoption of machine learning within the health sector.

Details

Digital Transformation and Society, vol. 3 no. 1
Type: Research Article
ISSN: 2755-0761

Keywords

Article
Publication date: 17 June 2021

Ambica Ghai, Pradeep Kumar and Samrat Gupta

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered…

1177

Abstract

Purpose

Web users rely heavily on online content make decisions without assessing the veracity of the content. The online content comprising text, image, video or audio may be tampered with to influence public opinion. Since the consumers of online information (misinformation) tend to trust the content when the image(s) supplement the text, image manipulation software is increasingly being used to forge the images. To address the crucial problem of image manipulation, this study focusses on developing a deep-learning-based image forgery detection framework.

Design/methodology/approach

The proposed deep-learning-based framework aims to detect images forged using copy-move and splicing techniques. The image transformation technique aids the identification of relevant features for the network to train effectively. After that, the pre-trained customized convolutional neural network is used to train on the public benchmark datasets, and the performance is evaluated on the test dataset using various parameters.

Findings

The comparative analysis of image transformation techniques and experiments conducted on benchmark datasets from a variety of socio-cultural domains establishes the effectiveness and viability of the proposed framework. These findings affirm the potential applicability of proposed framework in real-time image forgery detection.

Research limitations/implications

This study bears implications for several important aspects of research on image forgery detection. First this research adds to recent discussion on feature extraction and learning for image forgery detection. While prior research on image forgery detection, hand-crafted the features, the proposed solution contributes to stream of literature that automatically learns the features and classify the images. Second, this research contributes to ongoing effort in curtailing the spread of misinformation using images. The extant literature on spread of misinformation has prominently focussed on textual data shared over social media platforms. The study addresses the call for greater emphasis on the development of robust image transformation techniques.

Practical implications

This study carries important practical implications for various domains such as forensic sciences, media and journalism where image data is increasingly being used to make inferences. The integration of image forgery detection tools can be helpful in determining the credibility of the article or post before it is shared over the Internet. The content shared over the Internet by the users has become an important component of news reporting. The framework proposed in this paper can be further extended and trained on more annotated real-world data so as to function as a tool for fact-checkers.

Social implications

In the current scenario wherein most of the image forgery detection studies attempt to assess whether the image is real or forged in an offline mode, it is crucial to identify any trending or potential forged image as early as possible. By learning from historical data, the proposed framework can aid in early prediction of forged images to detect the newly emerging forged images even before they occur. In summary, the proposed framework has a potential to mitigate physical spreading and psychological impact of forged images on social media.

Originality/value

This study focusses on copy-move and splicing techniques while integrating transfer learning concepts to classify forged images with high accuracy. The synergistic use of hitherto little explored image transformation techniques and customized convolutional neural network helps design a robust image forgery detection framework. Experiments and findings establish that the proposed framework accurately classifies forged images, thus mitigating the negative socio-cultural spread of misinformation.

Details

Information Technology & People, vol. 37 no. 2
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 17 March 2023

Rui Tian, Ruheng Yin and Feng Gan

Music sentiment analysis helps to promote the diversification of music information retrieval methods. Traditional music emotion classification tasks suffer from high manual…

Abstract

Purpose

Music sentiment analysis helps to promote the diversification of music information retrieval methods. Traditional music emotion classification tasks suffer from high manual workload and low classification accuracy caused by difficulty in feature extraction and inaccurate manual determination of hyperparameter. In this paper, the authors propose an optimized convolution neural network-random forest (CNN-RF) model for music sentiment classification which is capable of optimizing the manually selected hyperparameters to improve the accuracy of music sentiment classification and reduce labor costs and human classification errors.

Design/methodology/approach

A CNN-RF music sentiment classification model is designed based on quantum particle swarm optimization (QPSO). First, the audio data are transformed into a Mel spectrogram, and feature extraction is conducted by a CNN. Second, the music features extracted are processed by RF algorithm to complete a preliminary emotion classification. Finally, to select the suitable hyperparameters for a CNN, the QPSO algorithm is adopted to extract the best hyperparameters and obtain the final classification results.

Findings

The model has gone through experimental validations and achieved a classification accuracy of 97 per cent for different sentiment categories with shortened training time. The proposed method with QPSO achieved 1.2 and 1.6 per cent higher accuracy than that with particle swarm optimization and genetic algorithm, respectively. The proposed model had great potential for music sentiment classification.

Originality/value

The dual contribution of this work comprises the proposed model which integrated two deep learning models and the introduction of a QPSO into model optimization. With these two innovations, the efficiency and accuracy of music emotion recognition and classification have been significantly improved.

Details

Data Technologies and Applications, vol. 57 no. 5
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 15 July 2022

Joy Iong-Zong Chen, Ping-Feng Huang and Chung Sheng Pi

Apart from, the smart edge computing (EC) robot (SECR) provides the tools to manage Internet of things (IoT) services in the edge landscape by means of real-world test-bed…

Abstract

Purpose

Apart from, the smart edge computing (EC) robot (SECR) provides the tools to manage Internet of things (IoT) services in the edge landscape by means of real-world test-bed designed in ECR. Eventually, based on the results from two experiments held in little constrained condition, such as the maximum data size is 2GB, the performance of the proposed techniques demonstrate the effectiveness, scalability and performance efficiency of the proposed IoT model.

Design/methodology/approach

Certainly, the proposed SECR is trying primarily to take over other traditional static robots in a centralized or distributed cloud environment. One aspect of representation of the proposed edge computing algorithms is due to challenge to slow down the consumption of time which happened in an artificial intelligence (AI) robot system. Thus, the developed SECR trained by tiny machine learning (TinyML) techniques to develop a decentralized and dynamic software environment.

Findings

Specifically, the waste time of SECR has actually slowed down when it is embedded with Edge Computing devices in the demonstration of data transmission within different paths. The TinyML is applied to train with image data sets for generating a framework running in the SECR for the recognition which has also proved with a second complete experiment.

Originality/value

The work presented in this paper is the first research effort, and which is focusing on resource allocation and dynamic path selection for edge computing. The developed platform using a decoupled resource management model that manages the allocation of micro node resources independent of the service provisioning performed at the cloud and manager nodes. Besides, the algorithm of the edge computing management is established with different path and pass large data to cloud and receive it. In this work which considered the SECR framework is able to perform the same function as that supports to the multi-dimensional scaling (MDS).

Details

Industrial Robot: the international journal of robotics research and application, vol. 50 no. 4
Type: Research Article
ISSN: 0143-991X

Keywords

Article
Publication date: 15 December 2023

Muhammad Arif Mahmood, Chioibasu Diana, Uzair Sajjad, Sabin Mihai, Ion Tiseanu and Andrei C. Popescu

Porosity is a commonly analyzed defect in the laser-based additive manufacturing processes owing to the enormous thermal gradient caused by repeated melting and solidification…

Abstract

Purpose

Porosity is a commonly analyzed defect in the laser-based additive manufacturing processes owing to the enormous thermal gradient caused by repeated melting and solidification. Currently, the porosity estimation is limited to powder bed fusion. The porosity estimation needs to be explored in the laser melting deposition (LMD) process, particularly analytical models that provide cost- and time-effective solutions compared to finite element analysis. For this purpose, this study aims to formulate two mathematical models for deposited layer dimensions and corresponding porosity in the LMD process.

Design/methodology/approach

In this study, analytical models have been proposed. Initially, deposited layer dimensions, including layer height, width and depth, were calculated based on the operating parameters. These outputs were introduced in the second model to estimate the part porosity. The models were validated with experimental data for Ti6Al4V depositions on Ti6Al4V substrate. A calibration curve (CC) was also developed for Ti6Al4V material and characterized using X-ray computed tomography. The models were also validated with the experimental results adopted from literature. The validated models were linked with the deep neural network (DNN) for its training and testing using a total of 6,703 computations with 1,500 iterations. Here, laser power, laser scanning speed and powder feeding rate were selected inputs, whereas porosity was set as an output.

Findings

The computations indicate that owing to the simultaneous inclusion of powder particulates, the powder elements use a substantial percentage of the laser beam energy for their melting, resulting in laser beam energy attenuation and reducing thermal value at the substrate. The primary operating parameters are directly correlated with the number of layers and total height in CC. Through X-ray computed tomography analyses, the number of layers showed a straightforward correlation with mean sphericity, while a converse relation was identified with the number, mean volume and mean diameter of pores. DNN and analytical models showed 2%–3% and 7%–9% mean absolute deviations, respectively, compared to the experimental results.

Originality/value

This research provides a unique solution for LMD porosity estimation by linking the developed analytical computational models with artificial neural networking. The presented framework predicts the porosity in the LMD-ed parts efficiently.

Article
Publication date: 9 August 2022

Vinay Singh, Iuliia Konovalova and Arpan Kumar Kar

Explainable artificial intelligence (XAI) has importance in several industrial applications. The study aims to provide a comparison of two important methods used for explainable…

Abstract

Purpose

Explainable artificial intelligence (XAI) has importance in several industrial applications. The study aims to provide a comparison of two important methods used for explainable AI algorithms.

Design/methodology/approach

In this study multiple criteria has been used to compare between explainable Ranked Area Integrals (xRAI) and integrated gradient (IG) methods for the explainability of AI algorithms, based on a multimethod phase-wise analysis research design.

Findings

The theoretical part includes the comparison of frameworks of two methods. In contrast, the methods have been compared across five dimensions like functional, operational, usability, safety and validation, from a practical point of view.

Research limitations/implications

A comparison has been made by combining criteria from theoretical and practical points of view, which demonstrates tradeoffs in terms of choices for the user.

Originality/value

Our results show that the xRAI method performs better from a theoretical point of view. However, the IG method shows a good result with both model accuracy and prediction quality.

Details

Benchmarking: An International Journal, vol. 30 no. 9
Type: Research Article
ISSN: 1463-5771

Keywords

Access

Year

Last 12 months (318)

Content type

Article (318)
1 – 10 of 318