Search results
1 – 10 of 231Tarun Jaiswal, Manju Pandey and Priyanka Tripathi
The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional…
Abstract
Purpose
The purpose of this study is to investigate and demonstrate the advancements achieved in the field of chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Typical convolutional neural networks (CNNs) are unable to capture both local and global contextual information effectively and apply a uniform operation to all pixels in an image. To address this, we propose an innovative approach that integrates a dynamic convolution operation at the encoder stage, improving image encoding quality and disease detection. In addition, a decoder based on the gated recurrent unit (GRU) is used for language modeling, and an attention network is incorporated to enhance consistency. This novel combination allows for improved feature extraction, mimicking the expertise of radiologists by selectively focusing on important areas and producing coherent captions with valuable clinical information.
Design/methodology/approach
In this study, we have presented a new report generation approach that utilizes dynamic convolution applied Resnet-101 (DyCNN) as an encoder (Verelst and Tuytelaars, 2019) and GRU as a decoder (Dey and Salemt, 2017; Pan et al., 2020), along with an attention network (see Figure 1). This integration innovatively extends the capabilities of image encoding and sequential caption generation, representing a shift from conventional CNN architectures. With its ability to dynamically adapt receptive fields, the DyCNN excels at capturing features of varying scales within the CXR images. This dynamic adaptability significantly enhances the granularity of feature extraction, enabling precise representation of localized abnormalities and structural intricacies. By incorporating this flexibility into the encoding process, our model can distil meaningful and contextually rich features from the radiographic data. While the attention mechanism enables the model to selectively focus on different regions of the image during caption generation. The attention mechanism enhances the report generation process by allowing the model to assign different importance weights to different regions of the image, mimicking human perception. In parallel, the GRU-based decoder adds a critical dimension to the process by ensuring a smooth, sequential generation of captions.
Findings
The findings of this study highlight the significant advancements achieved in chest X-ray image captioning through the utilization of dynamic convolutional encoder–decoder networks (DyCNN). Experiments conducted using the IU-Chest X-ray datasets showed that the proposed model outperformed other state-of-the-art approaches. The model achieved notable scores, including a BLEU_1 score of 0.591, a BLEU_2 score of 0.347, a BLEU_3 score of 0.277 and a BLEU_4 score of 0.155. These results highlight the efficiency and efficacy of the model in producing precise radiology reports, enhancing image interpretation and clinical decision-making.
Originality/value
This work is the first of its kind, which employs DyCNN as an encoder to extract features from CXR images. In addition, GRU as the decoder for language modeling was utilized and the attention mechanisms into the model architecture were incorporated.
Details
Keywords
Meng Zhu and Xiaolong Xu
Intent detection (ID) and slot filling (SF) are two important tasks in natural language understanding. ID is to identify the main intent of a paragraph of text. The goal of SF is…
Abstract
Purpose
Intent detection (ID) and slot filling (SF) are two important tasks in natural language understanding. ID is to identify the main intent of a paragraph of text. The goal of SF is to extract the information that is important to the intent from the input sentence. However, most of the existing methods use sentence-level intention recognition, which has the risk of error propagation, and the relationship between intention recognition and SF is not explicitly modeled. Aiming at this problem, this paper proposes a collaborative model of ID and SF for intelligent spoken language understanding called ID-SF-Fusion.
Design/methodology/approach
ID-SF-Fusion uses Bidirectional Encoder Representation from Transformers (BERT) and Bidirectional Long Short-Term Memory (BiLSTM) to extract effective word embedding and context vectors containing the whole sentence information respectively. Fusion layer is used to provide intent–slot fusion information for SF task. In this way, the relationship between ID and SF task is fully explicitly modeled. This layer takes the result of ID and slot context vectors as input to obtain the fusion information which contains both ID result and slot information. Meanwhile, to further reduce error propagation, we use word-level ID for the ID-SF-Fusion model. Finally, two tasks of ID and SF are realized by joint optimization training.
Findings
We conducted experiments on two public datasets, Airline Travel Information Systems (ATIS) and Snips. The results show that the Intent ACC score and Slot F1 score of ID-SF-Fusion on ATIS and Snips are 98.0 per cent and 95.8 per cent, respectively, and the two indicators on Snips dataset are 98.6 per cent and 96.7 per cent, respectively. These models are superior to slot-gated, SF-ID NetWork, stack-Prop and other models. In addition, ablation experiments were performed to further analyze and discuss the proposed model.
Originality/value
This paper uses word-level intent recognition and introduces intent information into the SF process, which is a significant improvement on both data sets.
Details
Keywords
Songlin Bao, Tiantian Li and Bin Cao
In the era of big data, various industries are generating large amounts of text data every day. Simplifying and summarizing these data can effectively serve users and improve…
Abstract
Purpose
In the era of big data, various industries are generating large amounts of text data every day. Simplifying and summarizing these data can effectively serve users and improve efficiency. Recently, zero-shot prompting in large language models (LLMs) has demonstrated remarkable performance on various language tasks. However, generating a very “concise” multi-document summary is a difficult task for it. When conciseness is specified in the zero-shot prompting, the generated multi-document summary still contains some unimportant information, even with the few-shot prompting. This paper aims to propose a LLMs prompting for multi-document summarization task.
Design/methodology/approach
To overcome this challenge, the authors propose chain-of-event (CoE) prompting for multi-document summarization (MDS) task. In this prompting, the authors take events as the center and propose a four-step summary reasoning process: specific event extraction; event abstraction and generalization; common event statistics; and summary generation. To further improve the performance of LLMs, the authors extend CoE prompting with the example of summary reasoning.
Findings
Summaries generated by CoE prompting are more abstractive, concise and accurate. The authors evaluate the authors’ proposed prompting on two data sets. The experimental results over ChatGLM2-6b show that the authors’ proposed CoE prompting consistently outperforms other typical promptings across all data sets.
Originality/value
This paper proposes CoE prompting to solve MDS tasks by the LLMs. CoE prompting can not only identify the key events but also ensure the conciseness of the summary. By this method, users can access the most relevant and important information quickly, improving their decision-making processes.
Details
Keywords
Syihabuddin Syihabuddin, Nurul Murtadho, Yusring Sanusi Baso, Hikmah Maulani and Shofa Musthofa Khalid
Assessing whether a book is relevant or suitable for use in teaching materials is not an easy and haphazard matter, various methods and theories have been offered by researchers…
Abstract
Purpose
Assessing whether a book is relevant or suitable for use in teaching materials is not an easy and haphazard matter, various methods and theories have been offered by researchers in studying this matter. Taking a study of the context of textbooks, researchers found the urgency that textbooks are a foundation for education, socialization and transmission of knowledge and its construction. Researchers offer another approach, namely by using praxeology as a study tool so that the goals of the textbooks previously intended are fulfilled.
Design/methodology/approach
The researcher uses a qualitative approach through grounded theory. Grounded theory procedures are designed to develop a well-integrated set of concepts that provide a thorough theoretical explanation of the social phenomena under study. A grounded theory must explain as well as describe. It may also implicitly provide some degree of predictability, but only with respect to certain conditions (Corbin and Strauss, 1990). Document analysis in conducting this research study. Document analysis itself examines systematic procedures for reviewing or evaluating documents, both printed and electronic materials.
Findings
Two issues regarding gender acquisition have been investigated in L2 Arabic acquisition studies; the order in which L2 Arabic learners acquire certain grammatical features of the gender system and the effect of L1 on the acquisition of some grammatical features from L2 grammatical gender. Arabic has a two-gender system that classifies all nouns, animate and inanimate, as masculine or feminine. Verbs, nouns, adjectives, personal, demonstrative and relative pronouns related to nouns in the syntactic structure of sentences show gender agreement.
Research limitations/implications
In practice, as a book intended for non-speakers, the book is presented using a general view of linguistic theory. In relation to the gender agreement, the presentation of the book begins and is inserted with the concepts of nouns and verbs. Returning to the praxeology context, First, The Know How (Praxis) explains practice (i.e. the tasks performed and the techniques used). Second, To Know Why or Knowledge (logos) which explains and justifies practice from a technological and theoretical point of view. Answering the first concept, the exercise presented in the book is a concept with three clusters explained at the beginning of the discussion. And the second concept, explained with a task design approach which includes word categorization by separating masculine and feminine word forms.
Practical implications
Practically, this research obtains perspectives studied from a textbook, namely the Arabic gender agreement is presented with various examples of noun contexts; textbook authors present book concepts in a particular way with regard to curriculum features and this task design affects student performance, and which approach is more effective for developing student understanding. Empirically, the material is in line with the formulation of competency standards for non-Arabic speakers in Indonesia.
Originality/value
With this computational search, the researcher found a novelty that was considered accurate by taking the praxeology context as a review in the analysis of non-speaking Arabic textbooks, especially in the year 2022 (last data collection in September) there has been no study on this context. So then, the researcher finds other interests in that praxeology can examine more broadly parts of the task of the contents of the book with the approach of relevant linguistic theories.
Details
Keywords
B. Vasavi, P. Dileep and Ulligaddala Srinivasarao
Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use…
Abstract
Purpose
Aspect-based sentiment analysis (ASA) is a task of sentiment analysis that requires predicting aspect sentiment polarity for a given sentence. Many traditional techniques use graph-based mechanisms, which reduce prediction accuracy and introduce large amounts of noise. The other problem with graph-based mechanisms is that for some context words, the feelings change depending on the aspect, and therefore it is impossible to draw conclusions on their own. ASA is challenging because a given sentence can reveal complicated feelings about multiple aspects.
Design/methodology/approach
This research proposed an optimized attention-based DL model known as optimized aspect and self-attention aware long short-term memory for target-based semantic analysis (OAS-LSTM-TSA). The proposed model goes through three phases: preprocessing, aspect extraction and classification. Aspect extraction is done using a double-layered convolutional neural network (DL-CNN). The optimized aspect and self-attention embedded LSTM (OAS-LSTM) is used to classify aspect sentiment into three classes: positive, neutral and negative.
Findings
To detect and classify sentiment polarity of the aspect using the optimized aspect and self-attention embedded LSTM (OAS-LSTM) model. The results of the proposed method revealed that it achieves a high accuracy of 95.3 per cent for the restaurant dataset and 96.7 per cent for the laptop dataset.
Originality/value
The novelty of the research work is the addition of two effective attention layers in the network model, loss function reduction and accuracy enhancement, using a recent efficient optimization algorithm. The loss function in OAS-LSTM is minimized using the adaptive pelican optimization algorithm, thus increasing the accuracy rate. The performance of the proposed method is validated on four real-time datasets, Rest14, Lap14, Rest15 and Rest16, for various performance metrics.
Details
Keywords
Karen M. DSouza and Aaron M. French
Purveyors of fake news perpetuate information that can harm society, including businesses. Social media's reach quickly amplifies distortions of fake news. Research has not yet…
Abstract
Purpose
Purveyors of fake news perpetuate information that can harm society, including businesses. Social media's reach quickly amplifies distortions of fake news. Research has not yet fully explored the mechanisms of such adversarial behavior or the adversarial techniques of machine learning that might be deployed to detect fake news. Debiasing techniques are also explored to combat against the generation of fake news using adversarial data. The purpose of this paper is to present the challenges and opportunities in fake news detection.
Design/methodology/approach
First, this paper provides an overview of adversarial behaviors and current machine learning techniques. Next, it describes the use of long short-term memory (LSTM) to identify fake news in a corpus of articles. Finally, it presents the novel adversarial behavior approach to protect targeted business datasets from attacks.
Findings
This research highlights the need for a corpus of fake news that can be used to evaluate classification methods. Adversarial debiasing using IBM's Artificial Intelligence Fairness 360 (AIF360) toolkit can improve the disparate impact of unfavorable characteristics of a dataset. Debiasing also demonstrates significant potential to reduce fake news generation based on the inherent bias in the data. These findings provide avenues for further research on adversarial collaboration and robust information systems.
Originality/value
Adversarial debiasing of datasets demonstrates that by reducing bias related to protected attributes, such as sex, race and age, businesses can reduce the potential of exploitation to generate fake news through adversarial data.
Details
Keywords
Yi-Hung Liu, Sheng-Fong Chen and Dan-Wei (Marian) Wen
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case…
Abstract
Purpose
Online medical repositories provide a platform for users to share information and dynamically access abundant electronic health data. It is important to determine whether case report information can assist the general public in appropriately managing their diseases. Therefore, this paper aims to introduce a novel deep learning-based method that allows non-professionals to make inquiries using ordinary vocabulary, retrieving the most relevant case reports for accurate and effective health information.
Design/methodology/approach
The dataset of case reports was collected from both the patient-generated research network and the digital medical journal repository. To enhance the accuracy of obtaining relevant case reports, the authors propose a retrieval approach that combines BERT and BiLSTM methods. The authors identified representative health-related case reports and analyzed the retrieval performance, as well as user judgments.
Findings
This study aims to provide the necessary functionalities to deliver relevant health case reports based on input from ordinary terms. The proposed framework includes features for health management, user feedback acquisition and ranking by weights to obtain the most pertinent case reports.
Originality/value
This study contributes to health information systems by analyzing patients' experiences and treatments with the case report retrieval model. The results of this study can provide immense benefit to the general public who intend to find treatment decisions and experiences from relevant case reports.
Details
Keywords
Victor Diogho Heuer de Carvalho and Ana Paula Cabral Seixas Costa
This article presents two Brazilian Portuguese corpora collected from different media concerning public security issues in a specific location. The primary motivation is…
Abstract
Purpose
This article presents two Brazilian Portuguese corpora collected from different media concerning public security issues in a specific location. The primary motivation is supporting analyses, so security authorities can make appropriate decisions about their actions.
Design/methodology/approach
The corpora were obtained through web scraping from a newspaper's website and tweets from a Brazilian metropolitan region. Natural language processing was applied considering: text cleaning, lemmatization, summarization, part-of-speech and dependencies parsing, named entities recognition, and topic modeling.
Findings
Several results were obtained based on the methodology used, highlighting some: an example of a summarization using an automated process; dependency parsing; the most common topics in each corpus; the forty named entities and the most common slogans were extracted, highlighting those linked to public security.
Research limitations/implications
Some critical tasks were identified for the research perspective, related to the applied methodology: the treatment of noise from obtaining news on their source websites, passing through textual elements quite present in social network posts such as abbreviations, emojis/emoticons, and even writing errors; the treatment of subjectivity, to eliminate noise from irony and sarcasm; the search for authentic news of issues within the target domain. All these tasks aim to improve the process to enable interested authorities to perform accurate analyses.
Practical implications
The corpora dedicated to the public security domain enable several analyses, such as mining public opinion on security actions in a given location; understanding criminals' behaviors reported in the news or even on social networks and drawing their attitudes timeline; detecting movements that may cause damage to public property and people welfare through texts from social networks; extracting the history and repercussions of police actions, crossing news with records on social networks; among many other possibilities.
Originality/value
The work on behalf of the corpora reported in this text represents one of the first initiatives to create textual bases in Portuguese, dedicated to Brazil's specific public security domain.
Details
Keywords
While sustainability experts point to interrelated social, economic and environmental goals, students may think about sustainability primarily as natural resources. To prepare…
Abstract
Purpose
While sustainability experts point to interrelated social, economic and environmental goals, students may think about sustainability primarily as natural resources. To prepare students to tackle global challenges to well-being, this paper aims to show that educators need to assess and address students’ shortcomings in considering socioeconomic dimensions.
Design/methodology/approach
This study coded essays on the meaning and components of sustainability written by 93 undergraduate and graduate students in environmental policy, business and engineering courses at US and Austrian universities. Then, the study reviewed a teaching strategy using diverse experts, case studies and assignments. Finally, the analysis evaluated students’ final projects proposing sustainability legislation with social, economic and environmental dimensions.
Findings
Students usually connect sustainability with limited natural resources affecting current and future generations, but seldom think that sustainability means acting on prominent socioeconomic challenges like poverty, food insecurity, pandemics and violence. Teaching in diverse courses through multidimensional case studies and legislation broadened and deepened students’ understanding and preparedness to act.
Originality/value
Despite experts’ attention to the interconnected Sustainable Development Goals, educators and policymakers need information on whether students associate sustainability with socioeconomic challenges. Open-response questions can reveal gaps in the respondents’ sustainability beliefs. In a wide range of courses, teaching can use diverse experts and multidimensional case studies and legislative assignments.
Details
Keywords
Xiang Zheng, Mingjie Li, Ze Wan and Yan Zhang
This study aims to extract knowledge of ancient Chinese scientific and technological documents bibliographic summaries (STDBS) and provide the knowledge graph (KG) comprehensively…
Abstract
Purpose
This study aims to extract knowledge of ancient Chinese scientific and technological documents bibliographic summaries (STDBS) and provide the knowledge graph (KG) comprehensively and systematically. By presenting the relationship among content, discipline, and author, this study focuses on providing services for knowledge discovery of ancient Chinese scientific and technological documents.
Design/methodology/approach
This study compiles ancient Chinese STDBS and designs a knowledge mining and graph visualization framework. The authors define the summaries' entities, attributes, and relationships for knowledge representation, use deep learning techniques such as BERT-BiLSTM-CRF models and rules for knowledge extraction, unify the representation of entities for knowledge fusion, and use Neo4j and other visualization techniques for KG construction and application. This study presents the generation, distribution, and evolution of ancient Chinese agricultural scientific and technological knowledge in visualization graphs.
Findings
The knowledge mining and graph visualization framework is feasible and effective. The BERT-BiLSTM-CRF model has domain adaptability and accuracy. The knowledge generation of ancient Chinese agricultural scientific and technological documents has distinctive time features. The knowledge distribution is uneven and concentrated, mainly concentrated on C1-Planting and cultivation, C2-Silkworm, and C3-Mulberry and water conservancy. The knowledge evolution is apparent, and differentiation and integration coexist.
Originality/value
This study is the first to visually present the knowledge connotation and association of ancient Chinese STDBS. It solves the problems of the lack of in-depth knowledge mining and connotation visualization of ancient Chinese STDBS.
Details