Search results
1 – 10 of over 14000Nassim Abdeldjallal Otmani, Malik Si-Mohammed, Catherine Comparot and Pierre-Jean Charrel
The purpose of this study is to propose a framework for extracting medical information from the Web using domain ontologies. Patient–Doctor conversations have become prevalent on…
Abstract
Purpose
The purpose of this study is to propose a framework for extracting medical information from the Web using domain ontologies. Patient–Doctor conversations have become prevalent on the Web. For instance, solutions like HealthTap or AskTheDoctors allow patients to ask doctors health-related questions. However, most online health-care consumers still struggle to express their questions efficiently due mainly to the expert/layman language and knowledge discrepancy. Extracting information from these layman descriptions, which typically lack expert terminology, is challenging. This hinders the efficiency of the underlying applications such as information retrieval. Herein, an ontology-driven approach is proposed, which aims at extracting information from such sparse descriptions using a meta-model.
Design/methodology/approach
A meta-model is designed to bridge the gap between the vocabulary of the medical experts and the consumers of the health services. The meta-model is mapped with SNOMED-CT to access the comprehensive medical vocabulary, as well as with WordNet to improve the coverage of layman terms during information extraction. To assess the potential of the approach, an information extraction prototype based on syntactical patterns is implemented.
Findings
The evaluation of the approach on the gold standard corpus defined in Task1 of ShARe CLEF 2013 showed promising results, an F-score of 0.79 for recognizing medical concepts in real-life medical documents.
Originality/value
The originality of the proposed approach lies in the way information is extracted. The context defined through a meta-model proved to be efficient for the task of information extraction, especially from layman descriptions.
Details
Keywords
Sudarsana Desul, Madurai Meenachi N., Thejas Venkatesh, Vijitha Gunta, Gowtham R. and Magapu Sai Baba
Ontology of a domain mainly consists of a set of concepts and their semantic relations. It is typically constructed and maintained by using ontology editors with substantial human…
Abstract
Purpose
Ontology of a domain mainly consists of a set of concepts and their semantic relations. It is typically constructed and maintained by using ontology editors with substantial human intervention. It is desirable to perform the task automatically, which has led to the development of ontology learning techniques. One of the main challenges of ontology learning from the text is to identify key concepts from the documents. A wide range of techniques for key concept extraction have been proposed but are having the limitations of low accuracy, poor performance, not so flexible and applicability to a specific domain. The propose of this study is to explore a new method to extract key concepts and to apply them to literature in the nuclear domain.
Design/methodology/approach
In this article, a novel method for key concept extraction is proposed and applied to the documents from the nuclear domain. A hybrid approach was used, which includes a combination of domain, syntactic name entity knowledge and statistical based methods. The performance of the developed method has been evaluated from the data obtained using two out of three voting logic from three domain experts by using 120 documents retrieved from SCOPUS database.
Findings
The work reported pertains to extracting concepts from the set of selected documents and aids the search for documents relating to given concepts. The results of a case study indicated that the method developed has demonstrated better metrics than Text2Onto and CFinder. The method described has the capability of extracting valid key concepts from a set of candidates with long phrases.
Research limitations/implications
The present study is restricted to literature coming out in the English language and applied to the documents from nuclear domain. It has the potential to extend to other domains also.
Practical implications
The work carried out in the current study has the potential of leading to updating International Nuclear Information System thesaurus for ontology in the nuclear domain. This can lead to efficient search methods.
Originality/value
This work is the first attempt to automatically extract key concepts from the nuclear documents. The proposed approach will address and fix the most of the problems that are existed in the current methods and thereby increase the performance.
Details
Keywords
Ioana Barbantan, Mihaela Porumb, Camelia Lemnaru and Rodica Potolea
Improving healthcare services by developing assistive technologies includes both the health aid devices and the analysis of the data collected by them. The acquired data modeled…
Abstract
Purpose
Improving healthcare services by developing assistive technologies includes both the health aid devices and the analysis of the data collected by them. The acquired data modeled as a knowledge base give more insight into each patient’s health status and needs. Therefore, the ultimate goal of a health-care system is obtaining recommendations provided by an assistive decision support system using such knowledge base, benefiting the patients, the physicians and the healthcare industry. This paper aims to define the knowledge flow for a medical assistive decision support system by structuring raw medical data and leveraging the knowledge contained in the data proposing solutions for efficient data search, medical investigation or diagnosis and medication prediction and relationship identification.
Design/methodology/approach
The solution this paper proposes for implementing a medical assistive decision support system can analyze any type of unstructured medical documents which are processed by applying Natural Language Processing (NLP) tasks followed by semantic analysis, leading to the medical concept identification, thus imposing a structure on the input documents. The structured information is filtered and classified such that custom decisions regarding patients’ health status can be made. The current research focuses on identifying the relationships between medical concepts as defined by the REMed (Relation Extraction from Medical documents) solution that aims at finding the patterns that lead to the classification of concept pairs into concept-to-concept relations.
Findings
This paper proposed the REMed solution expressed as a multi-class classification problem tackled using the support vector machine classifier. Experimentally, this paper determined the most appropriate setup for the multi-class classification problem which is a combination of lexical, context, syntactic and grammatical features, as each feature category is good at representing particular relations, but not all. The best results we obtained are expressed as F1-measure of 74.9 per cent which is 1.4 per cent better than the results reported by similar systems.
Research limitations/implications
The difficulty to discriminate between TrIP and TrAP relations revolves around the hierarchical relationship between the two classes as TrIP is a particular type (an instance) of TrAP. The intuition behind this behavior was that the classifier cannot discern the correct relations because of the bias toward the majority classes. The analysis was conducted by using only sentences from electronic health record that contain at least two medical concepts. This limitation was introduced by the availability of the annotated data with reported results, as relations were defined at sentence level.
Originality/value
The originality of the proposed solution lies in the methodology to extract valuable information from the medical records via semantic searches; concept-to-concept relation identification; and recommendations for diagnosis, treatment and further investigations. The REMed solution introduces a learning-based approach for the automatic discovery of relations between medical concepts. We propose an original list of features: lexical – 3, context – 6, grammatical – 4 and syntactic – 4. The similarity feature introduced in this paper has a significant influence on the classification, and, to the best of the authors’ knowledge, it has not been used as feature in similar solutions.
Details
Keywords
Bhaskar Sinha, Somnath Chandra and Megha Garg
The purpose of this explorative research study is to focus on the implementation of semantic Web technology on agriculture domain of e-governance data. The study contributes to an…
Abstract
Purpose
The purpose of this explorative research study is to focus on the implementation of semantic Web technology on agriculture domain of e-governance data. The study contributes to an understanding of problems and difficulties in implantations of unstructured and unformatted unique datasets of multilingual local language-based electronic dictionary (IndoWordnet).
Design/methodology/approach
An approach to an implementation in the perspective of conceptual logical concept to realization of agriculture-based terms and terminology extracted from linked multilingual IndoWordNet while maintaining the support and specification of the World Wide Web Consortium (W3C) standard of semantic Web technology to generate ontology and uniform unicode structured datasets.
Findings
The findings reveal the fact about partial support of extraction of terms, relations and concepts while linking to IndoWordNet, resulting in the form of SynSets, lexical relations of Words and relations between themselves. This helped in generation of ontology, hierarchical modeling and creation of structured metadata datasets.
Research limitations/implications
IndoWordNet has limitations, as it is not fully revised version due to diversified cultural base in India, and the new version is yet to be released in due time span. As mentioned in Section 5, implications of these ideas and experiments will have good impact in doing more exploration and better applications using such wordnet.
Practical implications
Language developer tools and frameworks have been used to get tagged annotated raw data processed and get intermediate results, which provides as a source for the generation of ontology and dynamic metadata.
Social implications
The results are expected to be applied for other e-governance applications. Better use of applications in social and government departments.
Originality/value
The authors have worked out experimental facts and raw information source datasets, revealing satisfactory results such as SynSets, sensecount, semantic and lexical relations, class concepts hierarchy and other related output, which helped in developing ontology of domain interest and, hence, creation of a dynamic metadata which can be globally used to facilitate various applications support.
Details
Keywords
Di Wu, Lei Wu, Alexis Palmer, Dr Kinshuk and Peng Zhou
Interaction content is created during online learning interaction for the exchanged information to convey experience and share knowledge. Prior studies have mainly focused on the…
Abstract
Purpose
Interaction content is created during online learning interaction for the exchanged information to convey experience and share knowledge. Prior studies have mainly focused on the quantity of online learning interaction content (OLIC) from the perspective of types or frequency, resulting in a limited analysis of the quality of OLIC. Domain concepts as the highest form of interaction are shown as entities or things that are particularly relevant to the educational domain of an online course. The purpose of this paper is to explore a new method to evaluate the quality of OLIC using domain concepts.
Design/methodology/approach
This paper proposes a novel approach to automatically evaluate the quality of OLIC regarding relevance, completeness and usefulness. A sample of OLIC corpus is classified and evaluated based on domain concepts and textual features.
Findings
Experimental results show that random forest classifiers not only outperform logistic regression and support vector machines but also their performance is improved by considering the quality dimensions of relevance and completeness. In addition, domain concepts contribute to improving the performance of evaluating OLIC.
Research limitations/implications
This paper adopts a limited sample to train the classification models. It has great benefits in monitoring students’ knowledge performance, supporting teachers’ decision-making and even enhancing the efficiency of school management.
Originality/value
This study extends the research of domain concepts in quality evaluation, especially in the online learning domain. It also has great potential for other domains.
Details
Keywords
Junping Qiu and Wen Lou
– The purpose of this study is to construct a Chinese information science resource ontology and to explore a new method for semiautomatic ontology construction.
Abstract
Purpose
The purpose of this study is to construct a Chinese information science resource ontology and to explore a new method for semiautomatic ontology construction.
Design/methodology/approach
More than 8,290 articles indexed in the Chinese Social Science Citation Index (CSSCI), covering the years 2001 to 2010, were included in this study. Statistical analysis, co-occurrence analysis, and semantic similarity methods were applied to the selected articles. The ontology was built using existing construction principles and methods, as well as categories and hierarchy definitions based on CSSCI indexing fields.
Findings
Seven categories were found to be relevant for the Chinese information science resource ontology, which, in this study, consists of a three-tier architecture, 78,291 instances, and 182,109 pairs of semantic relations. These results indicate the following: further improvements are required in ontology construction methods; resource ontology is a breakthrough concept in ontology studies; the combination of semantic similarities and co-occurrence analysis can quantitatively describe relationships between concepts.
Originality/value
This study pioneers the resource ontology concept. It is one of the first to combine informetric methods with semantic similarity to reveal deep relationships in textual data.
Details
Keywords
Marie Tirvaudey, Robin Bouclier, Jean-Charles Passieux and Ludovic Chamoin
The purpose of this paper is to further simplify the use of NURBS in industrial environnements. Although isogeometric analysis (IGA) has been the object of intensive studies over…
Abstract
Purpose
The purpose of this paper is to further simplify the use of NURBS in industrial environnements. Although isogeometric analysis (IGA) has been the object of intensive studies over the past decade, its massive deployment in industrial analysis still appears quite marginal. This is partly due to its implementation, which is not straightforward with respect to the elementary structure of finite element (FE) codes. This often discourages industrial engineers from adopting isogeometric capabilities in their well-established simulation environment.
Design/methodology/approach
Based on the concept of Bézier and Lagrange extractions, a novel method is proposed to implement IGA from an existing industrial FE code with the aim of bringing human implementation effort to the minimal possible level (only using standard input-output of finite element analysis (FEA) codes, avoid code-dependent subroutines implementation). An approximate global link to go from Lagrange polynomials to non-uniform-rational-B-splines functions is formulated, which enables the whole FE routines to be untouched during the implementation.
Findings
As a result, only the linear system resolution step is bypassed: the resolution is performed in an external script after projecting the FE system onto the reduced, more regular and isogeometric basis. The novel procedure is successfully validated through different numerical experiments involving linear and nonlinear isogeometric analyses using the standard input/output of the industrial FE software Code_Aster.
Originality/value
A non-invasive implementation of IGA into FEA software is proposed. The whole FE routines are untouched during the novel implementation procedure; a focus is made on the IGA solution of nonlinear problems from existing FEA software; technical details on the approach are provided by means of illustrative examples and step-by-step implementation; the methodology is evaluated on a range of two- and three-dimensional elasticity and elastoplasticity benchmarks solved using the commercial software Code_Aster.
Details
Keywords
Somayeh Tamjid, Fatemeh Nooshinfard, Molouk Sadat Hosseini Beheshti, Nadjla Hariri and Fahimeh Babalhavaeji
The purpose of this study is to develop a domain independent, cost-effective, time-saving and semi-automated ontology generation framework that could extract taxonomic concepts…
Abstract
Purpose
The purpose of this study is to develop a domain independent, cost-effective, time-saving and semi-automated ontology generation framework that could extract taxonomic concepts from unstructured text corpus. In the human disease domain, ontologies are found to be extremely useful for managing the diversity of technical expressions in favour of information retrieval objectives. The boundaries of these domains are expanding so fast that it is essential to continuously develop new ontologies or upgrade available ones.
Design/methodology/approach
This paper proposes a semi-automated approach that extracts entities/relations via text mining of scientific publications. Text mining-based ontology (TmbOnt)-named code is generated to assist a user in capturing, processing and establishing ontology elements. This code takes a pile of unstructured text files as input and projects them into high-valued entities or relations as output. As a semi-automated approach, a user supervises the process, filters meaningful predecessor/successor phrases and finalizes the demanded ontology-taxonomy. To verify the practical capabilities of the scheme, a case study was performed to drive glaucoma ontology-taxonomy. For this purpose, text files containing 10,000 records were collected from PubMed.
Findings
The proposed approach processed over 3.8 million tokenized terms of those records and yielded the resultant glaucoma ontology-taxonomy. Compared with two famous disease ontologies, TmbOnt-driven taxonomy demonstrated a 60%–100% coverage ratio against famous medical thesauruses and ontology taxonomies, such as Human Disease Ontology, Medical Subject Headings and National Cancer Institute Thesaurus, with an average of 70% additional terms recommended for ontology development.
Originality/value
According to the literature, the proposed scheme demonstrated novel capability in expanding the ontology-taxonomy structure with a semi-automated text mining approach, aiming for future fully-automated approaches.
Details
Keywords
Following the Australian Government's Garnaut Climate Change Review (CCR), the implementation of a joint business and climate change agenda is weighing heavily on the minds of…
Abstract
Purpose
Following the Australian Government's Garnaut Climate Change Review (CCR), the implementation of a joint business and climate change agenda is weighing heavily on the minds of those executives whose firms fit within the Emissions‐Intensive Trade‐Exposed Industry (EITEI) sectors. The purpose of this paper is to analyse and explain the major concerns that confront EITEI firms as the government moves Australia towards a low carbon economy.
Design/methodology/approach
The paper adopts an economic regulation perspective that focuses on public and private interests, coupled with the leximancer software package, which was used to analyse submissions made by EITEI firms to the Garnaut CCR.
Findings
The authors observed that the impact of costs on business and trade performance, future emissions trading schemes, investment in low emissions technologies, world greenhouse gas production levels in emissions‐intensive industries, and conflicting government policies form the foundations of serious corporate‐level concerns and uncertainties.
Research limitations/implications
The paper highlights that private interests, as expressed in the analysed submissions, intersect with the public interest and need to be addressed seriously.
Practical implications
Suggestions for a cooperative approach to addressing climate change that would involve businesses and governments are also put forward.
Originality/value
The paper utilises an economic regulation perspective to explain a practical issue and has implications for future climate change policy development.
Details
Keywords
Advanced analytics‐driven data analyses allow enterprises to have a complete or “360 degrees” view of their operations and customers. The insight that they gain from such analyses…
Abstract
Purpose
Advanced analytics‐driven data analyses allow enterprises to have a complete or “360 degrees” view of their operations and customers. The insight that they gain from such analyses is then used to direct, optimize, and automate their decision making to successfully achieve their organizational goals. Data, text, and web mining technologies are some of the key contributors to making advanced analytics possible. This paper aims to investigate these three mining technologies in terms of how they are used and the issues that are related to their effective implementation and management within the broader context of predictive or advanced analytics.
Design/methodology/approach
A range of recently published research literature on business intelligence (BI); predictive analytics; and data, text and web mining is reviewed to explore their current state, issues and challenges learned from their practice.
Findings
The findings are reported in two parts. The first part discusses a framework for BI using the data, text, and web mining technologies for advanced analytics; and the second part identifies and discusses the opportunities and challenges the business managers dealing with these technologies face for gaining competitive advantages for their businesses.
Originality/value
The study findings are intended to assist business managers to effectively understand the issues and emerging technologies behind advanced analytics implementation.
Details