Search results

1 – 10 of over 14000
Article
Publication date: 11 December 2018

Nassim Abdeldjallal Otmani, Malik Si-Mohammed, Catherine Comparot and Pierre-Jean Charrel

The purpose of this study is to propose a framework for extracting medical information from the Web using domain ontologies. Patient–Doctor conversations have become prevalent on…

Abstract

Purpose

The purpose of this study is to propose a framework for extracting medical information from the Web using domain ontologies. Patient–Doctor conversations have become prevalent on the Web. For instance, solutions like HealthTap or AskTheDoctors allow patients to ask doctors health-related questions. However, most online health-care consumers still struggle to express their questions efficiently due mainly to the expert/layman language and knowledge discrepancy. Extracting information from these layman descriptions, which typically lack expert terminology, is challenging. This hinders the efficiency of the underlying applications such as information retrieval. Herein, an ontology-driven approach is proposed, which aims at extracting information from such sparse descriptions using a meta-model.

Design/methodology/approach

A meta-model is designed to bridge the gap between the vocabulary of the medical experts and the consumers of the health services. The meta-model is mapped with SNOMED-CT to access the comprehensive medical vocabulary, as well as with WordNet to improve the coverage of layman terms during information extraction. To assess the potential of the approach, an information extraction prototype based on syntactical patterns is implemented.

Findings

The evaluation of the approach on the gold standard corpus defined in Task1 of ShARe CLEF 2013 showed promising results, an F-score of 0.79 for recognizing medical concepts in real-life medical documents.

Originality/value

The originality of the proposed approach lies in the way information is extracted. The context defined through a meta-model proved to be efficient for the task of information extraction, especially from layman descriptions.

Details

International Journal of Web Information Systems, vol. 15 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 30 November 2018

Sudarsana Desul, Madurai Meenachi N., Thejas Venkatesh, Vijitha Gunta, Gowtham R. and Magapu Sai Baba

Ontology of a domain mainly consists of a set of concepts and their semantic relations. It is typically constructed and maintained by using ontology editors with substantial human…

Abstract

Purpose

Ontology of a domain mainly consists of a set of concepts and their semantic relations. It is typically constructed and maintained by using ontology editors with substantial human intervention. It is desirable to perform the task automatically, which has led to the development of ontology learning techniques. One of the main challenges of ontology learning from the text is to identify key concepts from the documents. A wide range of techniques for key concept extraction have been proposed but are having the limitations of low accuracy, poor performance, not so flexible and applicability to a specific domain. The propose of this study is to explore a new method to extract key concepts and to apply them to literature in the nuclear domain.

Design/methodology/approach

In this article, a novel method for key concept extraction is proposed and applied to the documents from the nuclear domain. A hybrid approach was used, which includes a combination of domain, syntactic name entity knowledge and statistical based methods. The performance of the developed method has been evaluated from the data obtained using two out of three voting logic from three domain experts by using 120 documents retrieved from SCOPUS database.

Findings

The work reported pertains to extracting concepts from the set of selected documents and aids the search for documents relating to given concepts. The results of a case study indicated that the method developed has demonstrated better metrics than Text2Onto and CFinder. The method described has the capability of extracting valid key concepts from a set of candidates with long phrases.

Research limitations/implications

The present study is restricted to literature coming out in the English language and applied to the documents from nuclear domain. It has the potential to extend to other domains also.

Practical implications

The work carried out in the current study has the potential of leading to updating International Nuclear Information System thesaurus for ontology in the nuclear domain. This can lead to efficient search methods.

Originality/value

This work is the first attempt to automatically extract key concepts from the nuclear documents. The proposed approach will address and fix the most of the problems that are existed in the current methods and thereby increase the performance.

Details

The Electronic Library, vol. 37 no. 1
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 15 August 2016

Ioana Barbantan, Mihaela Porumb, Camelia Lemnaru and Rodica Potolea

Improving healthcare services by developing assistive technologies includes both the health aid devices and the analysis of the data collected by them. The acquired data modeled…

Abstract

Purpose

Improving healthcare services by developing assistive technologies includes both the health aid devices and the analysis of the data collected by them. The acquired data modeled as a knowledge base give more insight into each patient’s health status and needs. Therefore, the ultimate goal of a health-care system is obtaining recommendations provided by an assistive decision support system using such knowledge base, benefiting the patients, the physicians and the healthcare industry. This paper aims to define the knowledge flow for a medical assistive decision support system by structuring raw medical data and leveraging the knowledge contained in the data proposing solutions for efficient data search, medical investigation or diagnosis and medication prediction and relationship identification.

Design/methodology/approach

The solution this paper proposes for implementing a medical assistive decision support system can analyze any type of unstructured medical documents which are processed by applying Natural Language Processing (NLP) tasks followed by semantic analysis, leading to the medical concept identification, thus imposing a structure on the input documents. The structured information is filtered and classified such that custom decisions regarding patients’ health status can be made. The current research focuses on identifying the relationships between medical concepts as defined by the REMed (Relation Extraction from Medical documents) solution that aims at finding the patterns that lead to the classification of concept pairs into concept-to-concept relations.

Findings

This paper proposed the REMed solution expressed as a multi-class classification problem tackled using the support vector machine classifier. Experimentally, this paper determined the most appropriate setup for the multi-class classification problem which is a combination of lexical, context, syntactic and grammatical features, as each feature category is good at representing particular relations, but not all. The best results we obtained are expressed as F1-measure of 74.9 per cent which is 1.4 per cent better than the results reported by similar systems.

Research limitations/implications

The difficulty to discriminate between TrIP and TrAP relations revolves around the hierarchical relationship between the two classes as TrIP is a particular type (an instance) of TrAP. The intuition behind this behavior was that the classifier cannot discern the correct relations because of the bias toward the majority classes. The analysis was conducted by using only sentences from electronic health record that contain at least two medical concepts. This limitation was introduced by the availability of the annotated data with reported results, as relations were defined at sentence level.

Originality/value

The originality of the proposed solution lies in the methodology to extract valuable information from the medical records via semantic searches; concept-to-concept relation identification; and recommendations for diagnosis, treatment and further investigations. The REMed solution introduces a learning-based approach for the automatic discovery of relations between medical concepts. We propose an original list of features: lexical – 3, context – 6, grammatical – 4 and syntactic – 4. The similarity feature introduced in this paper has a significant influence on the classification, and, to the best of the authors’ knowledge, it has not been used as feature in similar solutions.

Details

International Journal of Web Information Systems, vol. 12 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 9 February 2015

Bhaskar Sinha, Somnath Chandra and Megha Garg

The purpose of this explorative research study is to focus on the implementation of semantic Web technology on agriculture domain of e-governance data. The study contributes to an…

1145

Abstract

Purpose

The purpose of this explorative research study is to focus on the implementation of semantic Web technology on agriculture domain of e-governance data. The study contributes to an understanding of problems and difficulties in implantations of unstructured and unformatted unique datasets of multilingual local language-based electronic dictionary (IndoWordnet).

Design/methodology/approach

An approach to an implementation in the perspective of conceptual logical concept to realization of agriculture-based terms and terminology extracted from linked multilingual IndoWordNet while maintaining the support and specification of the World Wide Web Consortium (W3C) standard of semantic Web technology to generate ontology and uniform unicode structured datasets.

Findings

The findings reveal the fact about partial support of extraction of terms, relations and concepts while linking to IndoWordNet, resulting in the form of SynSets, lexical relations of Words and relations between themselves. This helped in generation of ontology, hierarchical modeling and creation of structured metadata datasets.

Research limitations/implications

IndoWordNet has limitations, as it is not fully revised version due to diversified cultural base in India, and the new version is yet to be released in due time span. As mentioned in Section 5, implications of these ideas and experiments will have good impact in doing more exploration and better applications using such wordnet.

Practical implications

Language developer tools and frameworks have been used to get tagged annotated raw data processed and get intermediate results, which provides as a source for the generation of ontology and dynamic metadata.

Social implications

The results are expected to be applied for other e-governance applications. Better use of applications in social and government departments.

Originality/value

The authors have worked out experimental facts and raw information source datasets, revealing satisfactory results such as SynSets, sensecount, semantic and lexical relations, class concepts hierarchy and other related output, which helped in developing ontology of domain interest and, hence, creation of a dynamic metadata which can be globally used to facilitate various applications support.

Details

Journal of Knowledge Management, vol. 19 no. 1
Type: Research Article
ISSN: 1367-3270

Keywords

Article
Publication date: 14 May 2020

Di Wu, Lei Wu, Alexis Palmer, Dr Kinshuk and Peng Zhou

Interaction content is created during online learning interaction for the exchanged information to convey experience and share knowledge. Prior studies have mainly focused on the…

Abstract

Purpose

Interaction content is created during online learning interaction for the exchanged information to convey experience and share knowledge. Prior studies have mainly focused on the quantity of online learning interaction content (OLIC) from the perspective of types or frequency, resulting in a limited analysis of the quality of OLIC. Domain concepts as the highest form of interaction are shown as entities or things that are particularly relevant to the educational domain of an online course. The purpose of this paper is to explore a new method to evaluate the quality of OLIC using domain concepts.

Design/methodology/approach

This paper proposes a novel approach to automatically evaluate the quality of OLIC regarding relevance, completeness and usefulness. A sample of OLIC corpus is classified and evaluated based on domain concepts and textual features.

Findings

Experimental results show that random forest classifiers not only outperform logistic regression and support vector machines but also their performance is improved by considering the quality dimensions of relevance and completeness. In addition, domain concepts contribute to improving the performance of evaluating OLIC.

Research limitations/implications

This paper adopts a limited sample to train the classification models. It has great benefits in monitoring students’ knowledge performance, supporting teachers’ decision-making and even enhancing the efficiency of school management.

Originality/value

This study extends the research of domain concepts in quality evaluation, especially in the online learning domain. It also has great potential for other domains.

Details

The Electronic Library , vol. 38 no. 3
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 10 March 2014

Junping Qiu and Wen Lou

– The purpose of this study is to construct a Chinese information science resource ontology and to explore a new method for semiautomatic ontology construction.

2455

Abstract

Purpose

The purpose of this study is to construct a Chinese information science resource ontology and to explore a new method for semiautomatic ontology construction.

Design/methodology/approach

More than 8,290 articles indexed in the Chinese Social Science Citation Index (CSSCI), covering the years 2001 to 2010, were included in this study. Statistical analysis, co-occurrence analysis, and semantic similarity methods were applied to the selected articles. The ontology was built using existing construction principles and methods, as well as categories and hierarchy definitions based on CSSCI indexing fields.

Findings

Seven categories were found to be relevant for the Chinese information science resource ontology, which, in this study, consists of a three-tier architecture, 78,291 instances, and 182,109 pairs of semantic relations. These results indicate the following: further improvements are required in ontology construction methods; resource ontology is a breakthrough concept in ontology studies; the combination of semantic similarities and co-occurrence analysis can quantitatively describe relationships between concepts.

Originality/value

This study pioneers the resource ontology concept. It is one of the first to combine informetric methods with semantic similarity to reveal deep relationships in textual data.

Details

Aslib Journal of Information Management, vol. 66 no. 2
Type: Research Article
ISSN: 2050-3806

Keywords

Article
Publication date: 7 August 2019

Marie Tirvaudey, Robin Bouclier, Jean-Charles Passieux and Ludovic Chamoin

The purpose of this paper is to further simplify the use of NURBS in industrial environnements. Although isogeometric analysis (IGA) has been the object of intensive studies over…

Abstract

Purpose

The purpose of this paper is to further simplify the use of NURBS in industrial environnements. Although isogeometric analysis (IGA) has been the object of intensive studies over the past decade, its massive deployment in industrial analysis still appears quite marginal. This is partly due to its implementation, which is not straightforward with respect to the elementary structure of finite element (FE) codes. This often discourages industrial engineers from adopting isogeometric capabilities in their well-established simulation environment.

Design/methodology/approach

Based on the concept of Bézier and Lagrange extractions, a novel method is proposed to implement IGA from an existing industrial FE code with the aim of bringing human implementation effort to the minimal possible level (only using standard input-output of finite element analysis (FEA) codes, avoid code-dependent subroutines implementation). An approximate global link to go from Lagrange polynomials to non-uniform-rational-B-splines functions is formulated, which enables the whole FE routines to be untouched during the implementation.

Findings

As a result, only the linear system resolution step is bypassed: the resolution is performed in an external script after projecting the FE system onto the reduced, more regular and isogeometric basis. The novel procedure is successfully validated through different numerical experiments involving linear and nonlinear isogeometric analyses using the standard input/output of the industrial FE software Code_Aster.

Originality/value

A non-invasive implementation of IGA into FEA software is proposed. The whole FE routines are untouched during the novel implementation procedure; a focus is made on the IGA solution of nonlinear problems from existing FEA software; technical details on the approach are provided by means of illustrative examples and step-by-step implementation; the methodology is evaluated on a range of two- and three-dimensional elasticity and elastoplasticity benchmarks solved using the commercial software Code_Aster.

Details

Engineering Computations, vol. 37 no. 1
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 6 February 2024

Somayeh Tamjid, Fatemeh Nooshinfard, Molouk Sadat Hosseini Beheshti, Nadjla Hariri and Fahimeh Babalhavaeji

The purpose of this study is to develop a domain independent, cost-effective, time-saving and semi-automated ontology generation framework that could extract taxonomic concepts

Abstract

Purpose

The purpose of this study is to develop a domain independent, cost-effective, time-saving and semi-automated ontology generation framework that could extract taxonomic concepts from unstructured text corpus. In the human disease domain, ontologies are found to be extremely useful for managing the diversity of technical expressions in favour of information retrieval objectives. The boundaries of these domains are expanding so fast that it is essential to continuously develop new ontologies or upgrade available ones.

Design/methodology/approach

This paper proposes a semi-automated approach that extracts entities/relations via text mining of scientific publications. Text mining-based ontology (TmbOnt)-named code is generated to assist a user in capturing, processing and establishing ontology elements. This code takes a pile of unstructured text files as input and projects them into high-valued entities or relations as output. As a semi-automated approach, a user supervises the process, filters meaningful predecessor/successor phrases and finalizes the demanded ontology-taxonomy. To verify the practical capabilities of the scheme, a case study was performed to drive glaucoma ontology-taxonomy. For this purpose, text files containing 10,000 records were collected from PubMed.

Findings

The proposed approach processed over 3.8 million tokenized terms of those records and yielded the resultant glaucoma ontology-taxonomy. Compared with two famous disease ontologies, TmbOnt-driven taxonomy demonstrated a 60%–100% coverage ratio against famous medical thesauruses and ontology taxonomies, such as Human Disease Ontology, Medical Subject Headings and National Cancer Institute Thesaurus, with an average of 70% additional terms recommended for ontology development.

Originality/value

According to the literature, the proposed scheme demonstrated novel capability in expanding the ontology-taxonomy structure with a semi-automated text mining approach, aiming for future fully-automated approaches.

Details

The Electronic Library , vol. 42 no. 2
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 11 May 2012

Sumit Lodhia and Nigel Martin

Following the Australian Government's Garnaut Climate Change Review (CCR), the implementation of a joint business and climate change agenda is weighing heavily on the minds of…

Abstract

Purpose

Following the Australian Government's Garnaut Climate Change Review (CCR), the implementation of a joint business and climate change agenda is weighing heavily on the minds of those executives whose firms fit within the Emissions‐Intensive Trade‐Exposed Industry (EITEI) sectors. The purpose of this paper is to analyse and explain the major concerns that confront EITEI firms as the government moves Australia towards a low carbon economy.

Design/methodology/approach

The paper adopts an economic regulation perspective that focuses on public and private interests, coupled with the leximancer software package, which was used to analyse submissions made by EITEI firms to the Garnaut CCR.

Findings

The authors observed that the impact of costs on business and trade performance, future emissions trading schemes, investment in low emissions technologies, world greenhouse gas production levels in emissions‐intensive industries, and conflicting government policies form the foundations of serious corporate‐level concerns and uncertainties.

Research limitations/implications

The paper highlights that private interests, as expressed in the analysed submissions, intersect with the public interest and need to be addressed seriously.

Practical implications

Suggestions for a cooperative approach to addressing climate change that would involve businesses and governments are also put forward.

Originality/value

The paper utilises an economic regulation perspective to explain a practical issue and has implications for future climate change policy development.

Details

Sustainability Accounting, Management and Policy Journal, vol. 3 no. 1
Type: Research Article
ISSN: 2040-8021

Keywords

Article
Publication date: 13 March 2009

Ranjit Bose

Advanced analytics‐driven data analyses allow enterprises to have a complete or “360 degrees” view of their operations and customers. The insight that they gain from such analyses…

13546

Abstract

Purpose

Advanced analytics‐driven data analyses allow enterprises to have a complete or “360 degrees” view of their operations and customers. The insight that they gain from such analyses is then used to direct, optimize, and automate their decision making to successfully achieve their organizational goals. Data, text, and web mining technologies are some of the key contributors to making advanced analytics possible. This paper aims to investigate these three mining technologies in terms of how they are used and the issues that are related to their effective implementation and management within the broader context of predictive or advanced analytics.

Design/methodology/approach

A range of recently published research literature on business intelligence (BI); predictive analytics; and data, text and web mining is reviewed to explore their current state, issues and challenges learned from their practice.

Findings

The findings are reported in two parts. The first part discusses a framework for BI using the data, text, and web mining technologies for advanced analytics; and the second part identifies and discusses the opportunities and challenges the business managers dealing with these technologies face for gaining competitive advantages for their businesses.

Originality/value

The study findings are intended to assist business managers to effectively understand the issues and emerging technologies behind advanced analytics implementation.

Details

Industrial Management & Data Systems, vol. 109 no. 2
Type: Research Article
ISSN: 0263-5577

Keywords

1 – 10 of over 14000