Search results

1 – 10 of 89
Open Access
Article
Publication date: 4 April 2024

Yanmin Zhou, Zheng Yan, Ye Yang, Zhipeng Wang, Ping Lu, Philip F. Yuan and Bin He

Vision, audition, olfactory, tactile and taste are five important senses that human uses to interact with the real world. As facing more and more complex environments, a sensing…

Abstract

Purpose

Vision, audition, olfactory, tactile and taste are five important senses that human uses to interact with the real world. As facing more and more complex environments, a sensing system is essential for intelligent robots with various types of sensors. To mimic human-like abilities, sensors similar to human perception capabilities are indispensable. However, most research only concentrated on analyzing literature on single-modal sensors and their robotics application.

Design/methodology/approach

This study presents a systematic review of five bioinspired senses, especially considering a brief introduction of multimodal sensing applications and predicting current trends and future directions of this field, which may have continuous enlightenments.

Findings

This review shows that bioinspired sensors can enable robots to better understand the environment, and multiple sensor combinations can support the robot’s ability to behave intelligently.

Originality/value

The review starts with a brief survey of the biological sensing mechanisms of the five senses, which are followed by their bioinspired electronic counterparts. Their applications in the robots are then reviewed as another emphasis, covering the main application scopes of localization and navigation, objection identification, dexterous manipulation, compliant interaction and so on. Finally, the trends, difficulties and challenges of this research were discussed to help guide future research on intelligent robot sensors.

Details

Robotic Intelligence and Automation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 12 April 2023

Shaobo Liang and Linfeng Yu

As voice search has progressively become a new way of information acquisition and human–computer interaction, this paper aims to explore the users' voice search behavior in…

Abstract

Purpose

As voice search has progressively become a new way of information acquisition and human–computer interaction, this paper aims to explore the users' voice search behavior in human–vehicle interaction.

Design/methodology/approach

This study employed mixed research methods, including questionnaires and interviews. A total of 151 Amazon MTurk volunteers were recruited to complete a questionnaire based on their most recent and impressive voice search experience. After the questionnaire, this paper conducted an online interview with the participants.

Findings

This paper studied users' voice search behavior characteristics in the context of the human–vehicle interaction and analyzed the voice search content, search need, search motivation and user satisfaction. In addition, this paper studied the barriers and suggestions for voice search in human–vehicle interaction through a content analysis of the interviews.

Practical implications

This paper's analysis of users' barriers and suggestions has a specific reference value for optimizing the voice search interaction system and improving the service.

Originality/value

This study is exploratory research that seeks to identify users' voice search needs and tasks and investigate voice search satisfaction in human–vehicle interaction context.

Details

Library Hi Tech, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0737-8831

Keywords

Article
Publication date: 16 April 2024

Jinwei Zhao, Shuolei Feng, Xiaodong Cao and Haopei Zheng

This paper aims to concentrate on recent innovations in flexible wearable sensor technology tailored for monitoring vital signals within the contexts of wearable sensors and…

Abstract

Purpose

This paper aims to concentrate on recent innovations in flexible wearable sensor technology tailored for monitoring vital signals within the contexts of wearable sensors and systems developed specifically for monitoring health and fitness metrics.

Design/methodology/approach

In recent decades, wearable sensors for monitoring vital signals in sports and health have advanced greatly. Vital signals include electrocardiogram, electroencephalogram, electromyography, inertial data, body motions, cardiac rate and bodily fluids like blood and sweating, making them a good choice for sensing devices.

Findings

This report reviewed reputable journal articles on wearable sensors for vital signal monitoring, focusing on multimode and integrated multi-dimensional capabilities like structure, accuracy and nature of the devices, which may offer a more versatile and comprehensive solution.

Originality/value

The paper provides essential information on the present obstacles and challenges in this domain and provide a glimpse into the future directions of wearable sensors for the detection of these crucial signals. Importantly, it is evident that the integration of modern fabricating techniques, stretchable electronic devices, the Internet of Things and the application of artificial intelligence algorithms has significantly improved the capacity to efficiently monitor and leverage these signals for human health monitoring, including disease prediction.

Details

Sensor Review, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0260-2288

Keywords

Open Access
Article
Publication date: 21 April 2022

Warot Moungsouy, Thanawat Tawanbunjerd, Nutcha Liamsomboon and Worapan Kusakunniran

This paper proposes a solution for recognizing human faces under mask-wearing. The lower part of human face is occluded and could not be used in the learning process of face…

2638

Abstract

Purpose

This paper proposes a solution for recognizing human faces under mask-wearing. The lower part of human face is occluded and could not be used in the learning process of face recognition. So, the proposed solution is developed to recognize human faces on any available facial components which could be varied depending on wearing or not wearing a mask.

Design/methodology/approach

The proposed solution is developed based on the FaceNet framework, aiming to modify the existing facial recognition model to improve the performance of both scenarios of mask-wearing and without mask-wearing. Then, simulated masked-face images are computed on top of the original face images, to be used in the learning process of face recognition. In addition, feature heatmaps are also drawn out to visualize majority of parts of facial images that are significant in recognizing faces under mask-wearing.

Findings

The proposed method is validated using several scenarios of experiments. The result shows an outstanding accuracy of 99.2% on a scenario of mask-wearing faces. The feature heatmaps also show that non-occluded components including eyes and nose become more significant for recognizing human faces, when compared with the lower part of human faces which could be occluded under masks.

Originality/value

The convolutional neural network based solution is tuned up for recognizing human faces under a scenario of mask-wearing. The simulated masks on original face images are augmented for training the face recognition model. The heatmaps are then computed to prove that features generated from the top half of face images are correctly chosen for the face recognition.

Details

Applied Computing and Informatics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2634-1964

Keywords

Article
Publication date: 7 March 2024

Nehemia Sugianto, Dian Tjondronegoro and Golam Sorwar

This study proposes a collaborative federated learning (CFL) framework to address personal data transmission and retention issues for artificial intelligence (AI)-enabled video…

Abstract

Purpose

This study proposes a collaborative federated learning (CFL) framework to address personal data transmission and retention issues for artificial intelligence (AI)-enabled video surveillance in public spaces.

Design/methodology/approach

This study examines specific challenges for long-term people monitoring in public spaces and defines AI-enabled video surveillance requirements. Based on the requirements, this study proposes a CFL framework to gradually adapt AI models’ knowledge while reducing personal data transmission and retention. The framework uses three different federated learning strategies to rapidly learn from different new data sources while minimizing personal data transmission and retention to a central machine.

Findings

The findings confirm that the proposed CFL framework can help minimize the use of personal data without compromising the AI model's performance. The gradual learning strategies help develop AI-enabled video surveillance that continuously adapts for long-term deployment in public spaces.

Originality/value

This study makes two specific contributions to advance the development of AI-enabled video surveillance in public spaces. First, it examines specific challenges for long-term people monitoring in public spaces and defines AI-enabled video surveillance requirements. Second, it proposes a CFL framework to minimize data transmission and retention for AI-enabled video surveillance. The study provides comprehensive experimental results to evaluate the effectiveness of the proposed framework in the context of facial expression recognition (FER) which involves large-scale datasets.

Details

Information Technology & People, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 22 April 2024

Ruoxi Zhang and Chenhan Ren

This study aims to construct a sentiment series generation method for danmu comments based on deep learning, and explore the features of sentiment series after clustering.

Abstract

Purpose

This study aims to construct a sentiment series generation method for danmu comments based on deep learning, and explore the features of sentiment series after clustering.

Design/methodology/approach

This study consisted of two main parts: danmu comment sentiment series generation and clustering. In the first part, the authors proposed a sentiment classification model based on BERT fine-tuning to quantify danmu comment sentiment polarity. To smooth the sentiment series, they used methods, such as comprehensive weights. In the second part, the shaped-based distance (SBD)-K-shape method was used to cluster the actual collected data.

Findings

The filtered sentiment series or curves of the microfilms on the Bilibili website could be divided into four major categories. There is an apparently stable time interval for the first three types of sentiment curves, while the fourth type of sentiment curve shows a clear trend of fluctuation in general. In addition, it was found that “disputed points” or “highlights” are likely to appear at the beginning and the climax of films, resulting in significant changes in the sentiment curves. The clustering results show a significant difference in user participation, with the second type prevailing over others.

Originality/value

Their sentiment classification model based on BERT fine-tuning outperformed the traditional sentiment lexicon method, which provides a reference for using deep learning as well as transfer learning for danmu comment sentiment analysis. The BERT fine-tuning–SBD-K-shape algorithm can weaken the effect of non-regular noise and temporal phase shift of danmu text.

Details

The Electronic Library , vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 23 May 2023

Ayoung Suh

Despite the increasing interest in the metaverse—immersive three-dimensional virtual worlds wherein personalized avatars interact with one another—little is known about how users…

913

Abstract

Purpose

Despite the increasing interest in the metaverse—immersive three-dimensional virtual worlds wherein personalized avatars interact with one another—little is known about how users cognitively appraise and emotionally experience it. To fill this gap, the present study explores the emotional, behavioral and social consequences of users' cognitive appraisals, while focusing on social virtual reality (VR) as a representative entry point to the metaverse.

Design/methodology/approach

Drawing on cognitive appraisal theory against the backdrop of a framework for classifying emotions, this study develops and tests a theoretical model to account for users' continuance intention and its consequences in the social VR context based on two-wave panel survey data collected from 216 users of social VR platforms, including AltspaceVR, VRChat, Bigscreen and Rec Room.

Findings

The results of the first survey showed that perceived opportunity was more strongly influenced by technological opportunity than social opportunity, whereas perceived threat was more strongly affected by social threat than technological threat. Integrating the data collected from the first survey with those of the second survey, we also found that users' continuance intention positively influenced both their behavioral engagement and social self-efficacy.

Originality/value

By adopting a longitudinal approach, this study provides insights that may be valuable to researchers and practitioners who seek to use social VR for business purposes. This study also contributes to the metaverse literature by conceptualizing and operationalizing the opportunity and threat factors of social VR and identifying salient emotions that users experience in this context. Finally, this study has practical implications for addressing the social and technological features that may cause adverse user experiences in social VR.

Details

Information Technology & People, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0959-3845

Keywords

Article
Publication date: 12 September 2023

Wei Shi, Jing Zhang and Shaoyi He

With the rapid development of short videos in China, the public has become accustomed to using short videos to express their opinions. This paper aims to solve problems such as…

113

Abstract

Purpose

With the rapid development of short videos in China, the public has become accustomed to using short videos to express their opinions. This paper aims to solve problems such as how to represent the features of different modalities and achieve effective cross-modal feature fusion when analyzing the multi-modal sentiment of Chinese short videos (CSVs).

Design/methodology/approach

This paper aims to propose a sentiment analysis model MSCNN-CPL-CAFF using multi-scale convolutional neural network and cross attention fusion mechanism to analyze the CSVs. The audio-visual and textual data of CSVs themed on “COVID-19, catering industry” are collected from CSV platform Douyin first, and then a comparative analysis is conducted with advanced baseline models.

Findings

The sample number of the weak negative and neutral sentiment is the largest, and the sample number of the positive and weak positive sentiment is relatively small, accounting for only about 11% of the total samples. The MSCNN-CPL-CAFF model has achieved the Acc-2, Acc-3 and F1 score of 85.01%, 74.16 and 84.84%, respectively, which outperforms the highest value of baseline methods in accuracy and achieves competitive computation speed.

Practical implications

This research offers some implications regarding the impact of COVID-19 on catering industry in China by focusing on multi-modal sentiment of CSVs. The methodology can be utilized to analyze the opinions of the general public on social media platform and to categorize them accordingly.

Originality/value

This paper presents a novel deep-learning multimodal sentiment analysis model, which provides a new perspective for public opinion research on the short video platform.

Details

Kybernetes, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 13 March 2024

Rong Jiang, Bin He, Zhipeng Wang, Xu Cheng, Hongrui Sang and Yanmin Zhou

Compared with traditional methods relying on manual teaching or system modeling, data-driven learning methods, such as deep reinforcement learning and imitation learning, show…

Abstract

Purpose

Compared with traditional methods relying on manual teaching or system modeling, data-driven learning methods, such as deep reinforcement learning and imitation learning, show more promising potential to cope with the challenges brought by increasingly complex tasks and environments, which have become the hot research topic in the field of robot skill learning. However, the contradiction between the difficulty of collecting robot–environment interaction data and the low data efficiency causes all these methods to face a serious data dilemma, which has become one of the key issues restricting their development. Therefore, this paper aims to comprehensively sort out and analyze the cause and solutions for the data dilemma in robot skill learning.

Design/methodology/approach

First, this review analyzes the causes of the data dilemma based on the classification and comparison of data-driven methods for robot skill learning; Then, the existing methods used to solve the data dilemma are introduced in detail. Finally, this review discusses the remaining open challenges and promising research topics for solving the data dilemma in the future.

Findings

This review shows that simulation–reality combination, state representation learning and knowledge sharing are crucial for overcoming the data dilemma of robot skill learning.

Originality/value

To the best of the authors’ knowledge, there are no surveys that systematically and comprehensively sort out and analyze the data dilemma in robot skill learning in the existing literature. It is hoped that this review can be helpful to better address the data dilemma in robot skill learning in the future.

Details

Robotic Intelligence and Automation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2754-6969

Keywords

Article
Publication date: 27 March 2024

Yupeng Mou, Yixuan Gong and Zhihua Ding

Artificial intelligence (AI) is experiencing growth and prosperity worldwide because of its convenience and other benefits. However, AI faces challenges related to consumer…

Abstract

Purpose

Artificial intelligence (AI) is experiencing growth and prosperity worldwide because of its convenience and other benefits. However, AI faces challenges related to consumer resistance. Thus, drawing on the user resistance theory, this study explores factors that influence consumers’ resistance to AI and suggests ways to mitigate this negative influence.

Design/methodology/approach

This study tested four hypotheses across four studies by conducting lab experiments. Study 1 used a questionnaire to verify the hypothesis that AI’s “substitute” image leads to consumer resistance to AI; Study 2 focused on the role of perceived threat as an underlying driver of resistance to AI. Studies 3–4 provided process evidence by the way of a measured moderator, testing whether AI with servant communication style and literal language style is resisted less.

Findings

This study showed that AI’s “substitute” image increased users' resistance to AI. This occurs because the substitute image increases consumers’ perceived threat. The study also found that using servant communication and literal language styles in the interaction between AI and consumers can mitigate the negative effects of AI-substituted images.

Originality/value

This study reveals the mechanism of action between AI image and consumers’ resistance and sheds light on how to choose appropriate image and expression styles for AI products, which is important for lowering consumer resistance to AI.

Details

Marketing Intelligence & Planning, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0263-4503

Keywords

1 – 10 of 89