Search results

1 – 10 of over 1000
Article
Publication date: 5 June 2024

Azanzi Jiomekong and Sanju Tiwari

This paper aims to curate open research knowledge graph (ORKG) with papers related to ontology learning and define an approach using ORKG as a computer-assisted tool to organize…

Abstract

Purpose

This paper aims to curate open research knowledge graph (ORKG) with papers related to ontology learning and define an approach using ORKG as a computer-assisted tool to organize key-insights extracted from research papers.

Design/methodology/approach

Action research was used to explore, test and evaluate the use of the Open Research Knowledge Graph as a computer assistant tool for knowledge acquisition from scientific papers.

Findings

To extract, structure and describe research contributions, the granularity of information should be decided; to facilitate the comparison of scientific papers, one should design a common template that will be used to describe the state of the art of a domain.

Originality/value

This approach is currently used to document “food information engineering,” “tabular data to knowledge graph matching” and “question answering” research problems and the “neurosymbolic AI” domain. More than 200 papers are ingested in ORKG. From these papers, more than 800 contributions are documented and these contributions are used to build over 100 comparison tables. At the end of this work, we found that ORKG is a valuable tool that can reduce the working curve of state-of-the-art research.

Article
Publication date: 29 May 2024

Lino Gonzalez-Garcia, Gema González-Carreño, Ana María Rivas Machota and Juan Padilla Fernández-Vega

Knowledge graphs (KGs) are structured knowledge bases that represent real-world entities and are used in a variety of applications. Many of them are created and curated from a…

Abstract

Purpose

Knowledge graphs (KGs) are structured knowledge bases that represent real-world entities and are used in a variety of applications. Many of them are created and curated from a combination of automated and manual processes. Microdata embedded in Web pages for purposes of facilitating indexing and search engine optimization are a potential source to augment KGs under some assumptions of complementarity and quality that have not been thoroughly explored to date. In that direction, this paper aims to report results on a study that evaluates the potential of using microdata extracted from the Web to augment the large, open and manually curated Wikidata KG for the domain of touristic information. As large corpora of Web text is currently being leveraged via large language models (LLMs), these are used to compare the effectiveness of the microdata enhancement method.

Design/methodology/approach

The Schema.org taxonomy was used as the source to determine the annotation types to be collected. Here, the authors focused on tourism-related pages as a case study, selecting the relevant Schema.org concepts as point of departure. The large CommonCrawl resource was used to select those annotations from a large recent sample of the World Wide Web. The extracted annotations were processed and matched with Wikidata to estimate the degree to which microdata produced for SEO might become a valuable resource to complement KGs or vice versa. The Web pages themselves can also serve as a context to produce additional metadata elements using them as context in pipelines of an existing LLMs. That way, both the annotations and the contents itself can be used as sources.

Findings

The samples extracted revealed a concentration of metadata annotations in only a few of the relevant Schema.org attributes and also revealed the possible influence of authoring tools in a significant fraction of microdata produced. The analysis of the overlapping of attributes in the sample with those of Wikidata showed the potential of the technique, limited by the disbalance of the presence of attributes. The combination of those with the use of LLMs to produce additional annotations demonstrates the feasibility of the approach in the population of existing Wikidata locations. However, in both cases, the effectiveness appears to be lower in the cases of less content in the KG, which are arguably the most relevant when considering the scenario of an automated population approach.

Originality/value

The research reports novel empirical findings on the way touristic annotations with a SEO orientation are being produced in the wild and provides an assessment of their potential to complement KGs, or reuse information from those graphs. It also provides insights on the potential of using LLMs for the task.

Details

The Electronic Library , vol. 42 no. 3
Type: Research Article
ISSN: 0264-0473

Keywords

Article
Publication date: 19 July 2024

Kuoyi Lin, Xiaoyang Kan and Meilian Liu

This study develops and validates an innovative approach for extracting knowledge from online user reviews by integrating textual content and emojis. Recognizing the pivotal role…

Abstract

Purpose

This study develops and validates an innovative approach for extracting knowledge from online user reviews by integrating textual content and emojis. Recognizing the pivotal role emojis play in enhancing the expressiveness and emotional depth of digital communication, this study aims to address the significant gap in existing sentiment analysis models, which have largely overlooked the contribution of emojis in interpreting user preferences and sentiments. By constructing a comprehensive model that synergizes emotional and semantic information conveyed through emojis and text, this study seeks to provide a more nuanced understanding of user preferences, thereby enhancing the accuracy and depth of knowledge extraction from online reviews. The goal is to offer a robust framework that enables more effective and empathetic engagement with user-generated content on digital platforms, paving the way for improved service delivery, product development and customer satisfaction through informed insights into consumer behavior and sentiments.

Design/methodology/approach

This study uses a structured methodology to integrate and analyze text and emojis from online reviews for effective knowledge extraction, focusing on user preferences and sentiments. This methodology consists of four key stages. First, this study leverages high-frequency noun analysis to identify and extract product attributes mentioned in online user reviews. By focusing on nouns that appear frequently, the authors can systematically discern the primary features or aspects of products that users discuss, thereby providing a foundation for a more detailed sentiment and preference analysis. Second, a foundational sentiment dictionary is established that incorporates sentiment-bearing words, intensifiers and negation terms to analyze the textual part of the reviews. This dictionary is used to assign sentiment scores to phrases and sentences within reviews, allowing the quantification of textual sentiments based on the presence and combination of these predefined lexical items. Third, an emoticon sentiment dictionary is developed to address the emotional content conveyed through emojis. This dictionary categorizes emojis based on their associated sentiments, thus enabling the quantification of emotional expressions in reviews. The sentiment scores derived from the emojis are then integrated with those from the textual analysis. This integration considers the weights of text- and emoji-based emotions to compute a comprehensive attribute sentiment score that reflects a nuanced understanding of user sentiments and preferences. Finally, the authors conduct an empirical study to validate the effectiveness of the proposed methodology in mining user preferences from online reviews by applying the approach to a data set of online reviews and evaluating its ability to accurately identify product attributes and user sentiments. The validation process assessed the reliability and accuracy of the methodology in extracting meaningful insights from the complex interplay between text and emojis. This study offers a holistic and nuanced framework for knowledge extraction from online reviews, capturing both explicit and implicit sentiments expressed by users through text and emojis. By integrating these elements, this study seeks to provide a comprehensive understanding of user preferences, contributing to improved consumer insight and strategic decision-making for businesses and researchers.

Findings

The application of the proposed methodology for integrating emojis with text in online reviews yields significant findings that underscore the feasibility and value of extracting realistic user knowledge to gain insights from user-generated content. The analysis successfully captured consumer preferences, which are instrumental in informing service decisions and driving innovation. This achievement is largely attributed to the development and utilization of a comprehensive emotion-sentiment dictionary tailored to interpret the complex interplay between textual and emoji-based expressions in online reviews. By implementing a sentiment calculation model that intricately combines textual sentiment analysis with emoji sentiment analysis, this study was able to accurately determine the final attribute emotion for various product features discussed in the reviews. This model effectively characterized the emotional knowledge of online users and provided a nuanced understanding of their sentiments and preferences. The emotional knowledge extracted is not only quantifiable but also rich in context, offering deeper insights into consumer behavior and attitudes. Furthermore, a case analysis is conducted to rigorously test the validity of the proposed model in a real-world scenario. This practical examination revealed that the model is not only capable of accurately extracting and analyzing user preferences but is also adaptable to different contexts and product categories. The case analysis highlights the robustness and flexibility of the model, demonstrating its potential to enhance the precision of knowledge extraction processes significantly. Overall, the results confirm the effectiveness of the proposed approach in integrating text and emojis for comprehensive knowledge extraction from online reviews. The findings validate the model’s capability to offer actionable insights into consumer preferences, thereby supporting more informed and strategic decision-making by businesses. This study contributes to the broader field of sentiment analysis by showcasing the untapped potential of emojis as valuable indicators of user sentiments, opening new avenues for research and applications in digital marketing and consumer behavior analysis.

Originality/value

This study introduces a pioneering approach to extract knowledge from Web user interactions, notably through the integration of online reviews that incorporate both textual content and emoticons. This innovative methodology stands out because it holistically considers the dual channels of communication, text and emojis, to comprehensively mine Web user preferences. The key contribution of this study lies in its novel insights into the extraction of consumer preferences, advancing beyond traditional text-based analysis to embrace nuanced expressions conveyed through emoticons. The originality of this study is underpinned by its acknowledgment of emoticons as a significant and untapped source of sentiment and preference indicators in online reviews. By effectively merging emoticon analysis and emoji emotion scoring with textual sentiment analysis, this study enriches the understanding of Web user preferences and enhances the accuracy and depth of consumer preference insights. This dual-analysis approach represents a significant leap forward in sentiment analysis, setting a new standard for how digital communication can be leveraged to derive meaningful insights into consumer behavior. Furthermore, the results have practical implications to businesses and marketers. The insights gained from this integrated analytical approach offer a more granular and emotionally nuanced view of customer feedback, which can inform more effective marketing strategies, product development and customer service practices. By pioneering this comprehensive method of knowledge extraction, this study paves the way for future research and practice to interpret and respond more accurately to the complex landscape of online consumer expressions. This study’s originality and value lie in its innovative method of capturing and analyzing the rich tapestry of Web user communication, offering a ground-breaking perspective on consumer preference extraction that promises to enhance both academic research and practical applications in the digital era.

Details

Journal of Knowledge Management, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1367-3270

Keywords

Article
Publication date: 12 February 2024

Hamid Reza Saeidnia, Elaheh Hosseini, Shadi Abdoli and Marcel Ausloos

The study aims to analyze the synergy of artificial intelligence (AI), with scientometrics, webometrics and bibliometrics to unlock and to emphasize the potential of the…

Abstract

Purpose

The study aims to analyze the synergy of artificial intelligence (AI), with scientometrics, webometrics and bibliometrics to unlock and to emphasize the potential of the applications and benefits of AI algorithms in these fields.

Design/methodology/approach

By conducting a systematic literature review, our aim is to explore the potential of AI in revolutionizing the methods used to measure and analyze scholarly communication, identify emerging research trends and evaluate the impact of scientific publications. To achieve this, we implemented a comprehensive search strategy across reputable databases such as ProQuest, IEEE Explore, EBSCO, Web of Science and Scopus. Our search encompassed articles published from January 1, 2000, to September 2022, resulting in a thorough review of 61 relevant articles.

Findings

(1) Regarding scientometrics, the application of AI yields various distinct advantages, such as conducting analyses of publications, citations, research impact prediction, collaboration, research trend analysis and knowledge mapping, in a more objective and reliable framework. (2) In terms of webometrics, AI algorithms are able to enhance web crawling and data collection, web link analysis, web content analysis, social media analysis, web impact analysis and recommender systems. (3) Moreover, automation of data collection, analysis of citations, disambiguation of authors, analysis of co-authorship networks, assessment of research impact, text mining and recommender systems are considered as the potential of AI integration in the field of bibliometrics.

Originality/value

This study covers the particularly new benefits and potential of AI-enhanced scientometrics, webometrics and bibliometrics to highlight the significant prospects of the synergy of this integration through AI.

Details

Library Hi Tech, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0737-8831

Keywords

Article
Publication date: 29 May 2024

Ishrat Ayub Sofi, Taseef Ayub Sofi, Aasif Ahmad Mir and Ajra Bhat

Access to patent-related information is facilitated in large part by repositories of patents. Additionally, patent repositories support transparency and knowledge exchange, both…

Abstract

Purpose

Access to patent-related information is facilitated in large part by repositories of patents. Additionally, patent repositories support transparency and knowledge exchange, both of which can spark new alliances and collaborations. In addition to serving as a source of inspiration for future inventions, they allow companies, researchers and inventors to look up current patents and prevent infringement. Globally, the scientific and academic communities are becoming increasingly interested in open-access repositories. Countries throughout the world have kept up their repositories because of their significance. A directory of open access repositories (OpenDOAR) is a reliable source with minimally inaccurate or dubious content, having been meticulously chosen and validated. It acts as a global registration hub, enabling the visibility and accessibility of research contributions. Hence, this study aims to look into the current status of open-access repositories for archiving “Patents”, at the global level in OpenDOAR by analysing the different characteristic features of repositories.

Design/methodology/approach

The advanced search strategy of the directory of open-access repositories (www.opendoar.org/) is used to extract the data. The data extraction process was carried out on 28th March 2023. The study limited its search to “Patents” only, among the different content types listed in it. A total of 253 repositories were retrieved that contained the selected content type. However, the advanced search feature was combined one by one with other available parameters to retrieve the data. The gathered data was saved in MS Excel for further analysis. Moreover, the open access policies, open archives initiative protocol for metadata harvesting (OAI-PMH) and language interface of repositories were manually looked up from each repository/record information. To present the findings, charts and tables were used to visualize the gathered data effectively.

Findings

The study shows that repositories have increased over the years, with the highest number established in 2022. The UK has emerged as the most prominent country contributing to the development of repositories for archiving patents. The majority of the repositories are institutional, and DSpace is the most commonly used software for their creation. While Web 2.0 tools are not widely used, however, a significant number of repositories have incorporated RSS feeds, Atom and social media. Open access policies play a vital role in managing the content archived in the repositories, and only a small percentage of the repositories were found to be following them. However, the majority of the repositories have shown OAI-PMH compliance. English is the most commonly preferred interface language by repositories for archiving patents. These findings suggest that there is still significant room for improvement in the development and management of repositories, and adherence to open-access policies could play a crucial role in ensuring their sustainability and usefulness in the future.

Originality/value

To the best of the author’s knowledge, the study is the first of its type that examines the global landscape of open-access patent repositories.

Details

Information Discovery and Delivery, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2398-6247

Keywords

Article
Publication date: 8 December 2022

Khurram Shahzad and Shakeel Ahmad Khan

This study aims to investigate the current practices being implemented against the dissemination of fake online news, identify the relationship of new media literacy (NML) with…

Abstract

Purpose

This study aims to investigate the current practices being implemented against the dissemination of fake online news, identify the relationship of new media literacy (NML) with fake news epidemic control and find out the challenges in identifying valid sources of information.

Design/methodology/approach

To accomplish constructed objectives of this study, a systematic literature review (SLR) was conducted. The authors carried out the “Preferred Reporting Items for the Systematic Review and Meta-analysis” guidelines as a research methodology. The data were retrieved from ten world’s leading digital databases and online tools. A total of 25 key studies published in impact factor (IF) journals were included for systematic review vis-à-vis standard approaches.

Findings

This study revealed trending practices to control fake news consisted of critical information literacy, civic education, new thinking patterns, fact-checkers, automatic fake news detection tools, employment of ethical norms and deep learning via neural networks. Results of the synthesized studies revealed that media literacy, web literacy, digital literation, social media literacy skills and NML assisted acted as frontline soldiers in combating the fake news war. The findings of this research also exhibited different challenges to control fake news perils.

Research limitations/implications

This study provides pertinent theoretical contributions in the body of existing knowledge through the addition of valuable literature by conducting in-depth systematic review of 25 IF articles on a need-based topic.

Practical implications

This scholarly contribution is fruitful and practically productive for the policymakers belonging to different spectrums to effectively control web-based fake news epidemic.

Social implications

This intellectual piece is a benchmark to address fake news calamities to save the social system and to educate citizens from harms of false online stories on social networking websites.

Originality/value

This study vivifies new vistas via a reinvigorated outlook to address fake news perils embedded in dynamic, rigorous and heuristic strategies for redefining a predetermined set of social values.

Details

Global Knowledge, Memory and Communication, vol. 73 no. 6/7
Type: Research Article
ISSN: 2514-9342

Keywords

Article
Publication date: 7 December 2022

Peyman Jafary, Davood Shojaei, Abbas Rajabifard and Tuan Ngo

Building information modeling (BIM) is a striking development in the architecture, engineering and construction (AEC) industry, which provides in-depth information on different…

1117

Abstract

Purpose

Building information modeling (BIM) is a striking development in the architecture, engineering and construction (AEC) industry, which provides in-depth information on different stages of the building lifecycle. Real estate valuation, as a fully interconnected field with the AEC industry, can benefit from 3D technical achievements in BIM technologies. Some studies have attempted to use BIM for real estate valuation procedures. However, there is still a limited understanding of appropriate mechanisms to utilize BIM for valuation purposes and the consequent impact that BIM can have on decreasing the existing uncertainties in the valuation methods. Therefore, the paper aims to analyze the literature on BIM for real estate valuation practices.

Design/methodology/approach

This paper presents a systematic review to analyze existing utilizations of BIM for real estate valuation practices, discovers the challenges, limitations and gaps of the current applications and presents potential domains for future investigations. Research was conducted on the Web of Science, Scopus and Google Scholar databases to find relevant references that could contribute to the study. A total of 52 publications including journal papers, conference papers and proceedings, book chapters and PhD and master's theses were identified and thoroughly reviewed. There was no limitation on the starting date of research, but the end date was May 2022.

Findings

Four domains of application have been identified: (1) developing machine learning-based valuation models using the variables that could directly be captured through BIM and industry foundation classes (IFC) data instances of building objects and their attributes; (2) evaluating the capacity of 3D factors extractable from BIM and 3D GIS in increasing the accuracy of existing valuation models; (3) employing BIM for accurate estimation of components of cost approach-based valuation practices; and (4) extraction of useful visual features for real estate valuation from BIM representations instead of 2D images through deep learning and computer vision.

Originality/value

This paper contributes to research efforts on utilization of 3D modeling in real estate valuation practices. In this regard, this paper presents a broad overview of the current applications of BIM for valuation procedures and provides potential ways forward for future investigations.

Details

Engineering, Construction and Architectural Management, vol. 31 no. 4
Type: Research Article
ISSN: 0969-9988

Keywords

Open Access
Article
Publication date: 5 August 2024

James Christopher Westland and Jian Mou

Internet search is a $120bn business that answers lists of search terms or keywords with relevant links to Internet webpages. Only a few companies have sufficient scale to compete…

Abstract

Purpose

Internet search is a $120bn business that answers lists of search terms or keywords with relevant links to Internet webpages. Only a few companies have sufficient scale to compete and thus economics of the process are paramount. This study aims to develop a detailed industry-specific modeling of the economics of internet search.

Design/methodology/approach

The current research develops a stochastic model of the process of Internet indexing, search and retrieval in order to predict expected costs and revenues of particular configurations and usages.

Findings

The models define behavior and economics of parameters that are not directly observable, where it is difficult to empirically determine the distributions and economics.

Originality/value

The model may be used to guide the economics of large search engine operations, including the advertising platforms that depend on them and largely fund them.

Details

Journal of Electronic Business & Digital Economics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2754-4214

Keywords

Article
Publication date: 14 August 2024

Simon Knight, Isabella Bowdler, Heather Ford and Jianlong Zhou

Informational conflict and uncertainty are common features across a range of sources, topics and tasks. Search engines and their presentation of results via search engine results…

Abstract

Purpose

Informational conflict and uncertainty are common features across a range of sources, topics and tasks. Search engines and their presentation of results via search engine results pages (SERPs) often underpinned by knowledge graphs (KGs) are commonly used across tasks. Yet, it is not clear how search does, or could, represent the informational conflict that exists across and within returned results. The purpose of this paper is to review KG and SERP designs for representation of uncertainty or disagreement.

Design/methodology/approach

The authors address the aim through a systematic analysis of material regarding uncertainty and disagreement in KG and SERP contexts. Specifically, the authors focus on the material representation – user interface design features – that have been developed in the context of uncertainty and disagreement representation for KGs and SERPs.

Findings

Searches identified n = 136 items as relevant, with n = 4 sets of visual materials identified from these for analysis of their design features. Design elements were extracted against sets of design principles, highlighting tensions in the design of such features.

Originality/value

The authors conclude by highlighting two key challenges for interface design and recommending six design principles in representing uncertainty and conflict in SERPs. Given the important role technologies play in mediating information access and learning, addressing the representation of uncertainty and disagreement in the representation of information is crucial.

Details

Information and Learning Sciences, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2398-5348

Keywords

Book part
Publication date: 23 April 2024

Emerson Norabuena-Figueroa, Roger Rurush-Asencio, K. P. Jaheer Mukthar, Jose Sifuentes-Stratti and Elia Ramírez-Asís

The development of information technologies has led to a considerable transformation in human resource management from conventional or commonly known as personnel management to…

Abstract

The development of information technologies has led to a considerable transformation in human resource management from conventional or commonly known as personnel management to modern one. Data mining technology, which has been widely used in several applications, including those that function on the web, includes clustering algorithms as a key component. Web intelligence is a recent academic field that calls for sophisticated analytics and machine learning techniques to facilitate information discovery, particularly on the web. Human resource data gathered from the web are typically enormous, highly complex, dynamic, and unstructured. Traditional clustering methods need to be upgraded because they are ineffective. Standard clustering algorithms are enhanced and expanded with optimization capabilities to address this difficulty by swarm intelligence, a subset of nature-inspired computing. We collect the initial raw human resource data and preprocess the data wherein data cleaning, data normalization, and data integration takes place. The proposed K-C-means-data driven cuckoo bat optimization algorithm (KCM-DCBOA) is used for clustering of the human resource data. The feature extraction is done using principal component analysis (PCA) and the classification of human resource data is done using support vector machine (SVM). Other approaches from the literature were contrasted with the suggested approach. According to the experimental findings, the suggested technique has extremely promising features in terms of the quality of clustering and execution time.

Details

Technological Innovations for Business, Education and Sustainability
Type: Book
ISBN: 978-1-83753-106-6

Keywords

1 – 10 of over 1000