Search results
1 – 10 of 370Tarun Jaiswal, Manju Pandey and Priyanka Tripathi
The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional…
Abstract
Purpose
The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Typical convolutional neural networks (CNNs) are unable to capture both local and global contextual information effectively and apply a uniform operation to all pixels in an image. To address this, we propose an innovative approach that integrates a dynamic convolution operation at the encoder stage, improving image encoding quality and disease detection. In addition, a decoder based on the gated recurrent unit (GRU) is used for language modeling, and an attention network is incorporated to enhance consistency. This novel combination allows for improved feature extraction, mimicking the expertise of radiologists by selectively focusing on important areas and producing coherent captions with valuable clinical information.
Design/methodology/approach
In this study, we have presented a new report generation approach that utilizes dynamic convolution applied Resnet-101 (DyCNN) as an encoder (Verelst and Tuytelaars, 2019) and GRU as a decoder (Dey and Salemt, 2017; Pan et al., 2020), along with an attention network (see Figure 1). This integration innovatively extends the capabilities of image encoding and sequential caption generation, representing a shift from conventional CNN architectures. With its ability to dynamically adapt receptive fields, the DyCNN excels at capturing features of varying scales within the CXR images. This dynamic adaptability significantly enhances the granularity of feature extraction, enabling precise representation of localized abnormalities and structural intricacies. By incorporating this flexibility into the encoding process, our model can distil meaningful and contextually rich features from the radiographic data. While the attention mechanism enables the model to selectively focus on different regions of the image during caption generation. The attention mechanism enhances the report generation process by allowing the model to assign different importance weights to different regions of the image, mimicking human perception. In parallel, the GRU-based decoder adds a critical dimension to the process by ensuring a smooth, sequential generation of captions.
Findings
The findings of this study highlight the significant advancements achieved in chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Experiments conducted using the IU-Chest X-ray datasets showed that the proposed model outperformed other state-of-the-art approaches. The model achieved notable scores, including a BLEU_1 score of 0.591, a BLEU_2 score of 0.347, a BLEU_3 score of 0.277 and a BLEU_4 score of 0.155. These results highlight the efficiency and efficacy of the model in producing precise radiology reports, enhancing image interpretation and clinical decision-making.
Originality/value
This work is the first of its kind, which employs DyCNN as an encoder to extract features from CXR images. In addition, GRU as the decoder for language modeling was utilized and the attention mechanisms into the model architecture were incorporated.
Details
Keywords
Prajakta Chandrakant Kandarkar and V. Ravi
Industry 4.0 has put forward a smart perspective on managing supply chain networks and their operations. The current manufacturing system is primarily data-driven. Industries are…
Abstract
Purpose
Industry 4.0 has put forward a smart perspective on managing supply chain networks and their operations. The current manufacturing system is primarily data-driven. Industries are deploying new emerging technologies in their operations to build a competitive edge in the business environment; however, the true potential of smart manufacturing has not yet been fully unveiled. This research aims to extensively analyse emerging technologies and their interconnection with smart manufacturing in developing smarter supply chains.
Design/methodology/approach
This research endeavours to establish a conceptual framework for a smart supply chain. A real case study on a smart factory is conducted to demonstrate the validity of this framework for building smarter supply chains. A comparative analysis is carried out between conventional and smart supply chains to ascertain the advantages of smart supply chains. In addition, a thorough investigation of the several factors needed to transition from smart to smarter supply chains is undertaken.
Findings
The integration of smart technology exemplifies the ability to improve the efficiency of supply chain operations. Research findings indicate that transitioning to a smart factory radically enhances productivity, quality assurance, data privacy and labour efficiency. The outcomes of this research will help academic and industrial sectors critically comprehend technological breakthroughs and their applications in smart supply chains.
Originality/value
This study highlights the implications of incorporating smart technologies into supply chain operations, specifically in smart purchasing, smart factory operations, smart warehousing and smart customer performance. A paradigm transition from conventional, smart to smarter supply chains offers a comprehensive perspective on the evolving dynamics in automation, optimisation and manufacturing technology domains, ultimately leading to the emergence of Industry 5.0.
Details
Keywords
Hong Zhou, Binwei Gao, Shilong Tang, Bing Li and Shuyu Wang
The number of construction dispute cases has maintained a high growth trend in recent years. The effective exploration and management of construction contract risk can directly…
Abstract
Purpose
The number of construction dispute cases has maintained a high growth trend in recent years. The effective exploration and management of construction contract risk can directly promote the overall performance of the project life cycle. The miss of clauses may result in a failure to match with standard contracts. If the contract, modified by the owner, omits key clauses, potential disputes may lead to contractors paying substantial compensation. Therefore, the identification of construction project contract missing clauses has heavily relied on the manual review technique, which is inefficient and highly restricted by personnel experience. The existing intelligent means only work for the contract query and storage. It is urgent to raise the level of intelligence for contract clause management. Therefore, this paper aims to propose an intelligent method to detect construction project contract missing clauses based on Natural Language Processing (NLP) and deep learning technology.
Design/methodology/approach
A complete classification scheme of contract clauses is designed based on NLP. First, construction contract texts are pre-processed and converted from unstructured natural language into structured digital vector form. Following the initial categorization, a multi-label classification of long text construction contract clauses is designed to preliminary identify whether the clause labels are missing. After the multi-label clause missing detection, the authors implement a clause similarity algorithm by creatively integrating the image detection thought, MatchPyramid model, with BERT to identify missing substantial content in the contract clauses.
Findings
1,322 construction project contracts were tested. Results showed that the accuracy of multi-label classification could reach 93%, the accuracy of similarity matching can reach 83%, and the recall rate and F1 mean of both can reach more than 0.7. The experimental results verify the feasibility of intelligently detecting contract risk through the NLP-based method to some extent.
Originality/value
NLP is adept at recognizing textual content and has shown promising results in some contract processing applications. However, the mostly used approaches of its utilization for risk detection in construction contract clauses predominantly are rule-based, which encounter challenges when handling intricate and lengthy engineering contracts. This paper introduces an NLP technique based on deep learning which reduces manual intervention and can autonomously identify and tag types of contractual deficiencies, aligning with the evolving complexities anticipated in future construction contracts. Moreover, this method achieves the recognition of extended contract clause texts. Ultimately, this approach boasts versatility; users simply need to adjust parameters such as segmentation based on language categories to detect omissions in contract clauses of diverse languages.
Details
Keywords
Koraljka Golub, Osma Suominen, Ahmed Taiye Mohammed, Harriet Aagaard and Olof Osterman
In order to estimate the value of semi-automated subject indexing in operative library catalogues, the study aimed to investigate five different automated implementations of an…
Abstract
Purpose
In order to estimate the value of semi-automated subject indexing in operative library catalogues, the study aimed to investigate five different automated implementations of an open source software package on a large set of Swedish union catalogue metadata records, with Dewey Decimal Classification (DDC) as the target classification system. It also aimed to contribute to the body of research on aboutness and related challenges in automated subject indexing and evaluation.
Design/methodology/approach
On a sample of over 230,000 records with close to 12,000 distinct DDC classes, an open source tool Annif, developed by the National Library of Finland, was applied in the following implementations: lexical algorithm, support vector classifier, fastText, Omikuji Bonsai and an ensemble approach combing the former four. A qualitative study involving two senior catalogue librarians and three students of library and information studies was also conducted to investigate the value and inter-rater agreement of automatically assigned classes, on a sample of 60 records.
Findings
The best results were achieved using the ensemble approach that achieved 66.82% accuracy on the three-digit DDC classification task. The qualitative study confirmed earlier studies reporting low inter-rater agreement but also pointed to the potential value of automatically assigned classes as additional access points in information retrieval.
Originality/value
The paper presents an extensive study of automated classification in an operative library catalogue, accompanied by a qualitative study of automated classes. It demonstrates the value of applying semi-automated indexing in operative information retrieval systems.
Details
Keywords
Weak repeatability is observed in handcrafted keypoints, leading to tracking failures in visual simultaneous localization and mapping (SLAM) systems under challenging scenarios…
Abstract
Purpose
Weak repeatability is observed in handcrafted keypoints, leading to tracking failures in visual simultaneous localization and mapping (SLAM) systems under challenging scenarios such as illumination change, rapid rotation and large angle of view variation. In contrast, learning-based keypoints exhibit higher repetition but entail considerable computational costs. This paper proposes an innovative algorithm for keypoint extraction, aiming to strike an equilibrium between precision and efficiency. This paper aims to attain accurate, robust and versatile visual localization in scenes of formidable complexity.
Design/methodology/approach
SiLK-SLAM initially refines the cutting-edge learning-based extractor, SiLK, and introduces an innovative postprocessing algorithm for keypoint homogenization and operational efficiency. Furthermore, SiLK-SLAM devises a reliable relocalization strategy called PCPnP, leveraging progressive and consistent sampling, thereby bolstering its robustness.
Findings
Empirical evaluations conducted on TUM, KITTI and EuRoC data sets substantiate SiLK-SLAM’s superior localization accuracy compared to ORB-SLAM3 and other methods. Compared to ORB-SLAM3, SiLK-SLAM demonstrates an enhancement in localization accuracy even by 70.99%, 87.20% and 85.27% across the three data sets. The relocalization experiments demonstrate SiLK-SLAM’s capability in producing precise and repeatable keypoints, showcasing its robustness in challenging environments.
Originality/value
The SiLK-SLAM achieves exceedingly elevated localization accuracy and resilience in formidable scenarios, holding paramount importance in enhancing the autonomy of robots navigating intricate environments. Code is available at https://github.com/Pepper-FlavoredChewingGum/SiLK-SLAM.
Details
Keywords
The advent of robotics and automation technologies was augmenting firm initiatives to attain competitive advantage. From a resource-based view perspective, human-led capabilities…
Abstract
Purpose
The advent of robotics and automation technologies was augmenting firm initiatives to attain competitive advantage. From a resource-based view perspective, human-led capabilities were important to operate with technology resource base of an organisation. This was evident for both manufacturing as well as services firms. However, employees as an individual confronted technology anxiety (TA) when they were working with new technologies like robotics and automation technologies. Thus, the purpose of this paper was to examine the factors causing TA.
Design/methodology/approach
Given the novelty of this research study context a qualitative exploratory method was designed. For this research study, the data collected was through in-depth interviews conducted through open-ended semi-structured questionnaire. The data was collected from 62 frontline employees who were working with robotics and automation-based technologies in manufacturing firms. The authors applied thematic content analysis on collected data for analysis.
Findings
Technology anxieties ranged from fear of complete inability to learn new technologies, failure to learn new technologies properly, incapability to implement the learned skills and job loss to younger technology savvy employees. Finally, there was anxiety over job loss as automation and robotic technologies over the years was expected to erode the employment of human workforce altogether.
Research limitations/implications
The author undertook the research study based upon the TA perspective advocated by Meuter et al. (2003) and Yang and Forney (2013). Furthermore, this research study in the context of robotics and automation-based technologies in the manufacturing sector applied the mental accounting theory (Thaler, 1999) and technology self-efficacy perspective (Huffman et al., 2013).
Practical implications
Managers involved in the implementation of robotics and automation-based technologies were required to address TA of employees. Fear of job loss had to be addressed specially in a country like India. Anxiety regarding the ability to learn to work with robotics and automation technologies also was needed to be addressed by managers through adequate training and time for preparation. The benefits of robotics and automation-based technologies for employees as well as organisations have to be ascertained through open communication between the management and the frontline employees.
Originality/value
To the best of the authors’ knowledge, this was one of the first empirical research studies which deliberated regarding TA in the context of frontline workers working with robotics and automation-based technologies in the manufacturing sector. This research study was based upon a combination of varied perspectives ranging from micro foundations theory, TA, mental accounting theory and technology self-efficacy perspective.
Details
Keywords
Yi-Hung Liu, Sheng-Fong Chen and Dan-Wei (Marian) Wen
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case…
Abstract
Purpose
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case report information can assist the general public in appropriately managing their diseases. Therefore, this paper aims to introduce a novel deep learning-based method that allows non-professionals to make inquiries using ordinary vocabulary, retrieving the most relevant case reports for accurate and effective health information.
Design/methodology/approach
The dataset of case reports was collected from both the patient-generated research network and the digital medical journal repository. To enhance the accuracy of obtaining relevant case reports, the authors propose a retrieval approach that combines BERT and BiLSTM methods. The authors identified representative health-related case reports and analyzed the retrieval performance, as well as user judgments.
Findings
This study aims to provide the necessary functionalities to deliver relevant health case reports based on input from ordinary terms. The proposed framework includes features for health management, user feedback acquisition and ranking by weights to obtain the most pertinent case reports.
Originality/value
This study contributes to health information systems by analyzing patients' experiences and treatments with the case report retrieval model. The results of this study can provide immense benefit to the general public who intend to find treatment decisions and experiences from relevant case reports.
Details
Keywords
Hamid Hassani, Azadeh Mohebi, M.J. Ershadi and Ammar Jalalimanesh
The purpose of this research is to provide a framework in which new data quality dimensions are defined. The new dimensions provide new metrics for the assessment of lecture video…
Abstract
Purpose
The purpose of this research is to provide a framework in which new data quality dimensions are defined. The new dimensions provide new metrics for the assessment of lecture video indexing. As lecture video indexing involves various steps, the proposed framework containing new dimensions, introduces new integrated approach for evaluating an indexing method or algorithm from the beginning to the end.
Design/methodology/approach
The emphasis in this study is on the fifth step of design science research methodology (DSRM), known as evaluation. That is, the methods that are developed in the field of lecture video indexing as an artifact, should be evaluated from different aspects. In this research, nine dimensions of data quality including accuracy, value-added, relevancy, completeness, appropriate amount of data, concise, consistency, interpretability and accessibility have been redefined based on previous studies and nominal group technique (NGT).
Findings
The proposed dimensions are implemented as new metrics to evaluate a newly developed lecture video indexing algorithm, LVTIA and numerical values have been obtained based on the proposed definitions for each dimension. In addition, the new dimensions are compared with each other in terms of various aspects. The comparison shows that each dimension that is used for assessing lecture video indexing, is able to reflect a different weakness or strength of an indexing method or algorithm.
Originality/value
Despite development of different methods for indexing lecture videos, the issue of data quality and its various dimensions have not been studied. Since data with low quality can affect the process of scientific lecture video indexing, the issue of data quality in this process requires special attention.
Details
Keywords
Yu Zhou, Jiaxin Liu and Dongliang Lei
This paper aims to investigate whether the two dominant financial reporting regimes, US Generally Accepted Accounting Principles (US GAAP) and International Financial Reporting…
Abstract
Purpose
This paper aims to investigate whether the two dominant financial reporting regimes, US Generally Accepted Accounting Principles (US GAAP) and International Financial Reporting Standards (IFRS), are associated with audit pricing and audit report lags.
Design/methodology/approach
In 2007, the US SEC eliminated the requirement for foreign registrants to reconcile their financial statements to US GAAP from IFRS. In this post-reconciliation setting in the USA, the authors use panel ordinary least square regressions to examine a sample of foreign firms cross-listed in the USA reporting under IFRS and US domestic firms reporting under US GAAP during the fiscal year 2007–2019.
Findings
The authors find that the firms reporting under IFRS have longer audit report lags than firms reporting under US GAAP. In addition, the authors find that firms reporting under IFRS pay higher audit fees than their US GAAP counterparts. The results are robust after controlling for the firm- and country-specific characteristics as well as using propensity-score matching.
Originality/value
To the best of the authors’ knowledge, this study is the first to provide empirical evidence that the differences between the two reporting regimes are associated with auditor behavior, possibly through additional audit efforts and audit complexity associated with auditing the principle-based IFRS relative to the rule-based US GAAP.
Details
Keywords
S. Thavasi and T. Revathi
With so many placement opportunities around the students in their final or prefinal year, they start to feel the strain of the season. The students feel the need to be aware of…
Abstract
Purpose
With so many placement opportunities around the students in their final or prefinal year, they start to feel the strain of the season. The students feel the need to be aware of their position and how to increase their chances of being hired. Hence, a system to guide their career is one of the needs of the day.
Design/methodology/approach
The job role prediction system utilizes machine learning techniques such as Naïve Bayes, K-Nearest Neighbor, Support Vector machines (SVM) and Artificial Neural Networks (ANN) to suggest a student’s job role based on their academic performance and course outcomes (CO), out of which ANN performs better. The system uses the Mepco Schlenk Engineering College curriculum, placement and students’ Assessment data sets, in which the CO and syllabus are used to determine the skills that the student has gained from their courses. The necessary skills for a job position are then extracted from the job advertisements. The system compares the student’s skills with the required skills for the job role based on the placement prediction result.
Findings
The system predicts placement possibilities with an accuracy of 93.33 and 98% precision. Also, the skill analysis for students gives the students information about their skill-set strengths and weaknesses.
Research limitations/implications
For skill-set analysis, only the direct assessment of the students is considered. Indirect assessment shall also be considered for future scope.
Practical implications
The model is adaptable and flexible (customizable) to any type of academic institute or universities.
Social implications
The research will be very much useful for the students community to bridge the gap between the academic and industrial needs.
Originality/value
Several works are done for career guidance for the students. However, these career guidance methodologies are designed only using the curriculum and students’ basic personal information. The proposed system will consider the students’ academic performance through direct assessment, along with their curriculum and basic personal information.
Details