Search results

1 – 10 of over 6000
Article
Publication date: 20 April 2023

Vishva Payghode, Ayush Goyal, Anupama Bhan, Sailesh Suryanarayan Iyer and Ashwani Kumar Dubey

This paper aims to implement and extend the You Only Live Once (YOLO) algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural…

Abstract

Purpose

This paper aims to implement and extend the You Only Live Once (YOLO) algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural network once to detect the objects in an image, which is why it is powerful and fast. Cameras are found at many different crossroads and locations, but video processing of the feed through an object detection algorithm allows determining and tracking what is captured. Video Surveillance has many applications such as Car Tracking and tracking of people related to crime prevention. This paper provides exhaustive comparison between the existing methods and proposed method. Proposed method is found to have highest object detection accuracy.

Design/methodology/approach

The goal of this research is to develop a deep learning framework to automate the task of analyzing video footage through object detection in images. This framework processes video feed or image frames from CCTV, webcam or a DroidCam, which allows the camera in a mobile phone to be used as a webcam for a laptop. The object detection algorithm, with its model trained on a large data set of images, is able to load in each image given as an input, process the image and determine the categories of the matching objects that it finds. As a proof of concept, this research demonstrates the algorithm on images of several different objects. This research implements and extends the YOLO algorithm for detection of objects and activities. The advantage of YOLO is that it only runs a neural network once to detect the objects in an image, which is why it is powerful and fast. Cameras are found at many different crossroads and locations, but video processing of the feed through an object detection algorithm allows determining and tracking what is captured. For video surveillance of traffic cameras, this has many applications, such as car tracking and person tracking for crime prevention. In this research, the implemented algorithm with the proposed methodology is compared against several different prior existing methods in literature. The proposed method was found to have the highest object detection accuracy for object detection and activity recognition, better than other existing methods.

Findings

The results indicate that the proposed deep learning–based model can be implemented in real-time for object detection and activity recognition. The added features of car crash detection, fall detection and social distancing detection can be used to implement a real-time video surveillance system that can help save lives and protect people. Such a real-time video surveillance system could be installed at street and traffic cameras and in CCTV systems. When this system would detect a car crash or a fatal human or pedestrian fall with injury, it can be programmed to send automatic messages to the nearest local police, emergency and fire stations. When this system would detect a social distancing violation, it can be programmed to inform the local authorities or sound an alarm with a warning message to alert the public to maintain their distance and avoid spreading their aerosol particles that may cause the spread of viruses, including the COVID-19 virus.

Originality/value

This paper proposes an improved and augmented version of the YOLOv3 model that has been extended to perform activity recognition, such as car crash detection, human fall detection and social distancing detection. The proposed model is based on a deep learning convolutional neural network model used to detect objects in images. The model is trained using the widely used and publicly available Common Objects in Context data set. The proposed model, being an extension of YOLO, can be implemented for real-time object and activity recognition. The proposed model had higher accuracies for both large-scale and all-scale object detection. This proposed model also exceeded all the other previous methods that were compared in extending and augmenting the object detection to activity recognition. The proposed model resulted in the highest accuracy for car crash detection, fall detection and social distancing detection.

Details

International Journal of Web Information Systems, vol. 19 no. 3/4
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 20 October 2023

Komal Ghafoor, Tauqir Ahmad, Muhammad Aslam and Samyan Wahla

Assistive technology has been developed to assist the visually impaired individuals in their social interactions. Specifically designed to enhance communication skills, facilitate…

Abstract

Purpose

Assistive technology has been developed to assist the visually impaired individuals in their social interactions. Specifically designed to enhance communication skills, facilitate social engagement and improve the overall quality of life, conversational assistive technologies include speech recognition APIs, text-to-speech APIs and various communication tools that are real. Enable real-time interaction. Using natural language processing (NLP) and machine learning algorithms, the technology analyzes spoken language and provides appropriate responses, offering an immersive experience through voice commands, audio feedback and vibration alerts.

Design/methodology/approach

These technologies have demonstrated their ability to promote self-confidence and self-reliance in visually impaired individuals during social interactions. Moreover, they promise to improve social competence and foster better relationships. In short, assistive technology in conversation stands as a promising tool that empowers the visually impaired individuals, elevating the quality of their social engagement.

Findings

The main benefit of assistive communication technology is that it will help visually impaired people overcome communication barriers in social contexts. This technology helps them communicate effectively with acquaintances, family, co-workers and even strangers in public places. By enabling smoother and more natural communication, it works to reduce feelings of isolation and increase overall quality of life.

Originality/value

Research findings include successful activity recognition, aligning with activities on which the VGG-16 model was trained, such as hugging, shaking hands, talking, walking, waving and more. The originality of this study lies in its approach to address the challenges faced by the visually impaired individuals in their social interactions through modern technology. Research adds to the body of knowledge in the area of assistive technologies, which contribute to the empowerment and social inclusion of the visually impaired individuals.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 17 no. 1
Type: Research Article
ISSN: 1756-378X

Keywords

Article
Publication date: 31 August 2012

Xin Hong, Chris D. Nugent, Maurice D. Mulvenna, Suzanne Martin, Steven Devlin and Jonathan G. Wallace

Within smart homes, ambient sensors are used to monitor interactions between users and the home environment. The data produced from the sensors are used as the basis for the…

Abstract

Purpose

Within smart homes, ambient sensors are used to monitor interactions between users and the home environment. The data produced from the sensors are used as the basis for the inference of the users' behaviour information. Partitioning sensor data in response to individual instances of activity is critical for a smart home to be fully functional and to fulfil its roles, such as correctly measuring health status and detecting emergency situations. The purpose of this study is to propose a similarity‐based segmentation approach applied on time series sensor data in an effort to detect and recognise activities within a smart home.

Design/methodology/approach

The paper explores methods for analysing time‐related sensor activation events in an effort to undercover hidden activity events through the use of generic sensor modelling of activity based upon the general knowledge of the activities. Two similarity measures are proposed to compare a time series based sensor sequence and a generic sensor model of an activity. In addition, a framework is developed for automatically analysing sensor streams.

Findings

The results from evaluation of the proposed methodology on a publicly accessible reference dataset show that the proposed methods can detect and recognise multi‐category activities with satisfying accuracy, in addition to the capability of detecting interleaved activities.

Originality/value

The concepts introduced in this paper will improve automatic detection and recognition of daily living activities from timely ordered sensor events based on domain knowledge of the activities.

Details

International Journal of Pervasive Computing and Communications, vol. 8 no. 3
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 10 June 2014

Julia Kantorovitch, Janne Väre, Vesa Pehkonen, Arto Laikari and Heikki Seppälä

The purpose of this paper is to create new ideas for assistive technology products at home, especially products utilizing robotic consumer appliances available in the homes of…

Abstract

Purpose

The purpose of this paper is to create new ideas for assistive technology products at home, especially products utilizing robotic consumer appliances available in the homes of elderly people. The work was founded on a reported increase in household robots as well as an ageing population in the industrialized world.

Design/methodology/approach

Technology should be something that is perceived as belonging to our own world that fits our daily practices. Earlier studies show that in addition to cleaning functions, new household robots could change home routines and people's relationship to them. Taking the previous studies as a starting point, the paper proposes a vacuum cleaner robot as a platform for developing pervasive safety services and describe implementation of a conceptual prototype which brings the feeling of safety to an older person and their relatives by assisting in case of accidents. Moreover, the results are presented of an empirical evaluation of the prototype with end-users.

Findings

It is proved that reasonably priced off-the-shelf components can be used to build the safety product demonstration model. The initial evaluation results, as well as referenced studies show that the acceptance rate of a household robot-based product is high, which is encouraging for further research in this domain. Also the paper could pinpoint areas that will require further work.

Research limitations/implications

To add more practicality to the research and move towards product development, a strong industrial partner involved in household robotics would be needed. For increased reliability and robustness, more research is required in areas of advanced sensing technology and decision algorithms.

Originality/value

A novel concept of a safety product for elderly care based on a vacuum cleaner robot is presented and an attempt is made to increase awareness that there will be a demand for such products.

Details

Journal of Assistive Technologies, vol. 8 no. 2
Type: Research Article
ISSN: 1754-9450

Keywords

Content available

Abstract

Details

International Journal of Web Information Systems, vol. 19 no. 3/4
Type: Research Article
ISSN: 1744-0084

Article
Publication date: 5 June 2017

Sulaimon Adebayo Bashir, Andrei Petrovski and Daniel Doolan

This purpose of this paper is to develop a change detection technique for activity recognition model. The approach aims to detect changes in the initial accuracy of the model…

Abstract

Purpose

This purpose of this paper is to develop a change detection technique for activity recognition model. The approach aims to detect changes in the initial accuracy of the model after training and when the model is deployed for recognizing new unseen activities without access to the ground truth. The changes between the two sessions may occur because of differences in sensor placement, orientation and user characteristics such as age and gender. However, many of the existing approaches for model adaptation in activity recognition are blind methods because they continuously adapt the recognition model without explicit detection of changes in the model performance.

Design/methodology/approach

The approach determines the variation between reference activity data belonging to different classes and newly classified unseen data. If there is coherency between the data, it means the model is correctly classifying the instances; otherwise, a significant variation indicates wrong instances are being classified to different classes. Thus, the approach is formulated as a two-level architectural framework comprising of the off-line phase and the online phase. The off-line phase extracts of Shewart Chart change parameters from the training data set. The online phase performs classification of new samples and the detection of the changes in each class of activity present in the data set by using the change parameters computed earlier.

Findings

The approach is evaluated using a real activity-recognition data set. The results show that there are consistent detections that correlate with the error rate of the model.

Originality/value

The developed approach does not use ground truth to detect classifier performance degradation. Rather, it uses a data discrimination method and a base classifier to detect the changes by using the parameters computed from the reference data of each class to discriminate outliers in the new data being classified to the same class. The approach is the first, to the best of the authors’ knowledge, that addresses the problem of detecting within-user and cross-user variations that lead to concept drift in activity recognition. The approach is also the first to use statistical process control method for change detection in activity recognition, with a robust integrated framework that seamlessly detects variations in the underlying model performance.

Details

International Journal of Pervasive Computing and Communications, vol. 13 no. 2
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 28 January 2022

Jiaqi Li, Guangyi Zhou, Dongfang Li, Mingyuan Zhang and Xuefeng Zhao

Recognizing every worker's working status instead of only describing the existing construction activities in static images or videos as most computer vision-based approaches do;…

Abstract

Purpose

Recognizing every worker's working status instead of only describing the existing construction activities in static images or videos as most computer vision-based approaches do; identifying workers and their activities simultaneously; establishing a connection between workers and their behaviors.

Design/methodology/approach

Taking a reinforcement processing area as a research case, a new method for recognizing each different worker's activity through the position relationship of objects detected by Faster R-CNN is proposed. Firstly, based on four workers and four kinds of high-frequency activities, a Faster R-CNN model is trained. Then, by inputting the video into the model, with the coordinate of the boxes at each moment, the status of each worker can be judged.

Findings

The Faster R-CNN detector shows a satisfying performance with an mAP of 0.9654; with the detected boxes, a connection between the workers and activities is established; Through this connection, the average accuracy of activity recognition reached 0.92; with the proposed method, the labor consumption of each worker can be viewed more intuitively on the visualization graphics.

Originality/value

With this proposed method, the visualization graphics generated will help managers to evaluate the labor consumption of each worker more intuitively. Furthermore, human resources can be allocated more efficiently according to the information obtained. It is especially suitable for some small construction scenarios, in which the recognition model can work for a long time after it is established. This is potentially beneficial for the healthy operation of the entire project, and can also have a positive indirect impact on structural health and safety.

Details

Engineering, Construction and Architectural Management, vol. 30 no. 4
Type: Research Article
ISSN: 0969-9988

Keywords

Article
Publication date: 1 October 2006

Marco Leo, Tiziana D'Orazio, Paolo Spagnolo and Arcangelo Distante

The problem of automatic recognition of human activity is one of the most important and challenging areas of research in computer vision because of the wide range of possible…

Abstract

Purpose

The problem of automatic recognition of human activity is one of the most important and challenging areas of research in computer vision because of the wide range of possible applications, for example surveillance, advanced human‐computer interactions, monitoring. This paper presents statistical computer vision approaches to automatically recognize different human activities.

Design/methodology/approach

The human activity recognition process has three steps: firstly human blobs are segmented by motion analysis; then the human body posture is estimated and, finally a temporal model of the detected posture series is generated by discrete hidden Markov models to identify the activity.

Findings

The system was tested on image sequences acquired in a real archaeological site while some people simulated both legal and illegal actions. Four kinds of activity were automatically classified with a high percentage of correct detections.

Research limitations/implications

The proposed approach provides efficient solutions to some of the most common problems in the human activity recognition research field: high detailed image requirement, sequence alignment and intensive user interaction in the training phase. The main constraint of this framework is that the posture estimation approach is not completely view independent.

Practical implications

Results of time performance tests were very encouraging for the use of the proposed method in real time surveillance applications.

Originality/value

The proposed framework can work using low cost cameras with large view focal lenses. It does not need any a priori knowledge of the scene and no intensive user interaction is required in the early training phase.

Details

Sensor Review, vol. 26 no. 4
Type: Research Article
ISSN: 0260-2288

Keywords

Article
Publication date: 7 September 2015

Usman Naeem, Rabih Bashroush, Richard Anthony, Muhammad Awais Azam, Abdel Rahman Tawil, Sin Wee Lee and M.L. Dennis Wong

This paper aims to focus on applying a range of traditional classification- and semantic reasoning-based techniques to recognise activities of daily life (ADLs). ADL recognition

261

Abstract

Purpose

This paper aims to focus on applying a range of traditional classification- and semantic reasoning-based techniques to recognise activities of daily life (ADLs). ADL recognition plays an important role in tracking functional decline among elderly people who suffer from Alzheimer’s disease. Accurate recognition enables smart environments to support and assist the elderly to lead an independent life for as long as possible. However, the ability to represent the complex structure of an ADL in a flexible manner remains a challenge.

Design/methodology/approach

This paper presents an ADL recognition approach, which uses a hierarchical structure for the representation and modelling of the activities, its associated tasks and their relationships. This study describes an approach in constructing ADLs based on a task-specific and intention-oriented plan representation language called Asbru. The proposed method is particularly flexible and adaptable for caregivers to be able to model daily schedules for Alzheimer’s patients.

Findings

A proof of concept prototype evaluation has been conducted for the validation of the proposed ADL recognition engine, which has comparable recognition results with existing ADL recognition approaches.

Originality/value

The work presented in this paper is novel, as the developed ADL recognition approach takes into account all relationships and dependencies within the modelled ADLs. This is very useful when conducting activity recognition with very limited features.

Details

International Journal of Pervasive Computing and Communications, vol. 11 no. 3
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 8 March 2021

Neethu P.S., Suguna R. and Palanivel Rajan S.

This paper aims to propose a novel methodology for classifying the gestures using support vector machine (SVM) classification method. Initially, the Red Green Blue color hand…

270

Abstract

Purpose

This paper aims to propose a novel methodology for classifying the gestures using support vector machine (SVM) classification method. Initially, the Red Green Blue color hand gesture image is converted into YCbCr image in preprocessing stage and then palm with finger region is segmented by threshold process. Then, distance transformation method is applied on the palm with finger segmented image. Further, the center point (centroid) of palm region is detected and the fingertips are detected using SVM classification algorithm based on the detected centroids of the detected palm region.

Design/methodology/approach

Gesture is a physical indication of the body to convey information. Though any bodily movement can be considered a gesture, generally it originates from the movement of hand or face or combination of both. Combined gestures are quiet complex and difficult for a machine to classify. This paper proposes a novel methodology for classifying the gestures using SVM classification method. Initially, the color hand gesture image is converted into YCbCr image in preprocessing stage and then palm with finger region is segmented by threshold process. Then, distance transformation method is applied on the palm with finger segmented image. Further, the center point of the palm region is detected and the fingertips are detected using SVM classification algorithm. The proposed hand gesture image classification system is applied and tested on “Jochen Triesch,” “Sebastien Marcel” and “11Khands” data set hand gesture images to evaluate the efficiency of the proposed system. The performance of the proposed system is analyzed with respect to sensitivity, specificity, accuracy and recognition rate. The simulation results of the proposed method on these different data sets are compared with the conventional methods.

Findings

This paper proposes a novel methodology for classifying the gestures using SVM classification method. Distance transform method is used to detect the center point of the segmented palm region. The proposed hand gesture detection methodology achieves 96.5% of sensitivity, 97.1% of specificity, 96.9% of accuracy and 99.3% of recognition rate on “Jochen Triesch” data set. The proposed hand gesture detection methodology achieves 94.6% of sensitivity, 95.4% of specificity, 95.3% of accuracy and 97.8% of recognition rate on “Sebastien Marcel” data set. The proposed hand gesture detection methodology achieves 97% of sensitivity, 98% of specificity, 98.1% of accuracy and 98.8% of recognition rate on “11Khands” data set. The proposed hand gesture detection methodology consumes 0.52 s as recognition time on “Jochen Triesch” data set images, 0.71 s as recognition time on “Sebastien Marcel” data set images and 0.22 s as recognition time on “11Khands” data set images. It is very clear that the proposed hand gesture detection methodology consumes less recognition rate on “11Khands” data set when compared with other data set images. Hence, this data set is very suitable for real-time hand gesture applications with multi background environments.

Originality/value

The modern world requires more numbers of automated systems for improving our daily routine activities in an efficient manner. This present day technology emerges touch screen methodology for operating or functioning many devices or machines with or without wire connections. This also makes impact on automated vehicles where the vehicles can be operated without any interfacing with the driver. This is possible through hand gesture recognition system. This hand gesture recognition system captures the real-time hand gestures, a physical movement of human hand, as a digital image and recognizes them with the pre stored set of hand gestures.

Details

Circuit World, vol. 48 no. 2
Type: Research Article
ISSN: 0305-6120

Keywords

1 – 10 of over 6000