Search results

1 – 10 of 12
Open Access
Article
Publication date: 6 December 2022

Worapan Kusakunniran, Sarattha Karnjanapreechakorn, Pitipol Choopong, Thanongchai Siriapisith, Nattaporn Tesavibul, Nopasak Phasukkijwatana, Supalert Prakhunhungsit and Sutasinee Boonsopon

This paper aims to propose a solution for detecting and grading diabetic retinopathy (DR) in retinal images using a convolutional neural network (CNN)-based approach. It could…

1244

Abstract

Purpose

This paper aims to propose a solution for detecting and grading diabetic retinopathy (DR) in retinal images using a convolutional neural network (CNN)-based approach. It could classify input retinal images into a normal class or an abnormal class, which would be further split into four stages of abnormalities automatically.

Design/methodology/approach

The proposed solution is developed based on a newly proposed CNN architecture, namely, DeepRoot. It consists of one main branch, which is connected by two side branches. The main branch is responsible for the primary feature extractor of both high-level and low-level features of retinal images. Then, the side branches further extract more complex and detailed features from the features outputted from the main branch. They are designed to capture details of small traces of DR in retinal images, using modified zoom-in/zoom-out and attention layers.

Findings

The proposed method is trained, validated and tested on the Kaggle dataset. The regularization of the trained model is evaluated using unseen data samples, which were self-collected from a real scenario from a hospital. It achieves a promising performance with a sensitivity of 98.18% under the two classes scenario.

Originality/value

The new CNN-based architecture (i.e. DeepRoot) is introduced with the concept of a multi-branch network. It could assist in solving a problem of an unbalanced dataset, especially when there are common characteristics across different classes (i.e. four stages of DR). Different classes could be outputted at different depths of the network.

Details

Applied Computing and Informatics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2634-1964

Keywords

Open Access
Article
Publication date: 6 July 2020

Basma Makhlouf Shabou, Julien Tièche, Julien Knafou and Arnaud Gaudinat

This paper aims to describe an interdisciplinary and innovative research conducted in Switzerland, at the Geneva School of Business Administration HES-SO and supported by the…

4232

Abstract

Purpose

This paper aims to describe an interdisciplinary and innovative research conducted in Switzerland, at the Geneva School of Business Administration HES-SO and supported by the State Archives of Neuchâtel (Office des archives de l'État de Neuchâtel, OAEN). The problem to be addressed is one of the most classical ones: how to extract and discriminate relevant data in a huge amount of diversified and complex data record formats and contents. The goal of this study is to provide a framework and a proof of concept for a software that helps taking defensible decisions on the retention and disposal of records and data proposed to the OAEN. For this purpose, the authors designed two axes: the archival axis, to propose archival metrics for the appraisal of structured and unstructured data, and the data mining axis to propose algorithmic methods as complementary or/and additional metrics for the appraisal process.

Design/methodology/approach

Based on two axes, this exploratory study designs and tests the feasibility of archival metrics that are paired to data mining metrics, to advance, as much as possible, the digital appraisal process in a systematic or even automatic way. Under Axis 1, the authors have initiated three steps: first, the design of a conceptual framework to records data appraisal with a detailed three-dimensional approach (trustworthiness, exploitability, representativeness). In addition, the authors defined the main principles and postulates to guide the operationalization of the conceptual dimensions. Second, the operationalization proposed metrics expressed in terms of variables supported by a quantitative method for their measurement and scoring. Third, the authors shared this conceptual framework proposing the dimensions and operationalized variables (metrics) with experienced professionals to validate them. The expert’s feedback finally gave the authors an idea on: the relevance and the feasibility of these metrics. Those two aspects may demonstrate the acceptability of such method in a real-life archival practice. In parallel, Axis 2 proposes functionalities to cover not only macro analysis for data but also the algorithmic methods to enable the computation of digital archival and data mining metrics. Based on that, three use cases were proposed to imagine plausible and illustrative scenarios for the application of such a solution.

Findings

The main results demonstrate the feasibility of measuring the value of data and records with a reproducible method. More specifically, for Axis 1, the authors applied the metrics in a flexible and modular way. The authors defined also the main principles needed to enable computational scoring method. The results obtained through the expert’s consultation on the relevance of 42 metrics indicate an acceptance rate above 80%. In addition, the results show that 60% of all metrics can be automated. Regarding Axis 2, 33 functionalities were developed and proposed under six main types: macro analysis, microanalysis, statistics, retrieval, administration and, finally, the decision modeling and machine learning. The relevance of metrics and functionalities is based on the theoretical validity and computational character of their method. These results are largely satisfactory and promising.

Originality/value

This study offers a valuable aid to improve the validity and performance of archival appraisal processes and decision-making. Transferability and applicability of these archival and data mining metrics could be considered for other types of data. An adaptation of this method and its metrics could be tested on research data, medical data or banking data.

Details

Records Management Journal, vol. 30 no. 2
Type: Research Article
ISSN: 0956-5698

Keywords

Open Access
Article
Publication date: 18 April 2023

Worapan Kusakunniran, Pairash Saiviroonporn, Thanongchai Siriapisith, Trongtum Tongdee, Amphai Uraiverotchanakorn, Suphawan Leesakul, Penpitcha Thongnarintr, Apichaya Kuama and Pakorn Yodprom

The cardiomegaly can be determined by the cardiothoracic ratio (CTR) which can be measured in a chest x-ray image. It is calculated based on a relationship between a size of heart…

2707

Abstract

Purpose

The cardiomegaly can be determined by the cardiothoracic ratio (CTR) which can be measured in a chest x-ray image. It is calculated based on a relationship between a size of heart and a transverse dimension of chest. The cardiomegaly is identified when the ratio is larger than a cut-off threshold. This paper aims to propose a solution to calculate the ratio for classifying the cardiomegaly in chest x-ray images.

Design/methodology/approach

The proposed method begins with constructing lung and heart segmentation models based on U-Net architecture using the publicly available datasets with the groundtruth of heart and lung masks. The ratio is then calculated using the sizes of segmented lung and heart areas. In addition, Progressive Growing of GANs (PGAN) is adopted here for constructing the new dataset containing chest x-ray images of three classes including male normal, female normal and cardiomegaly classes. This dataset is then used for evaluating the proposed solution. Also, the proposed solution is used to evaluate the quality of chest x-ray images generated from PGAN.

Findings

In the experiments, the trained models are applied to segment regions of heart and lung in chest x-ray images on the self-collected dataset. The calculated CTR values are compared with the values that are manually measured by human experts. The average error is 3.08%. Then, the models are also applied to segment regions of heart and lung for the CTR calculation, on the dataset computed by PGAN. Then, the cardiomegaly is determined using various attempts of different cut-off threshold values. With the standard cut-off at 0.50, the proposed method achieves 94.61% accuracy, 88.31% sensitivity and 94.20% specificity.

Originality/value

The proposed solution is demonstrated to be robust across unseen datasets for the segmentation, CTR calculation and cardiomegaly classification, including the dataset generated from PGAN. The cut-off value can be adjusted to be lower than 0.50 for increasing the sensitivity. For example, the sensitivity of 97.04% can be achieved at the cut-off of 0.45. However, the specificity is decreased from 94.20% to 79.78%.

Details

Applied Computing and Informatics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2634-1964

Keywords

Open Access
Article
Publication date: 1 March 2019

Juan Pablo Sarmiento, Suzanne Polak and Vicente Sandoval

The purpose of this paper is to analyze the evidence-based research strategy (EBRS) used to evaluate eight projects that applied the neighborhood approach for disaster risk…

2795

Abstract

Purpose

The purpose of this paper is to analyze the evidence-based research strategy (EBRS) used to evaluate eight projects that applied the neighborhood approach for disaster risk reduction (NA-DRR) in informal urban settlements in Colombia, Guatemala, Haiti, Honduras, Jamaica and Peru, between 2012 and 2017.

Design/methodology/approach

The study covers the first five of the seven EBRS stages: first, identify relevant interventions; second, prepare evaluation questions; third, select evidence sources and implement a search strategy; fourth, appraise evidences and identify gaps; fifth, create an evaluation design to include an extensive literature review, followed by a mixed research method with surveys, focus groups and interviews; disaster risk modeling; georeferencing analysis; and engineering inspections. The last two stages: sixth, apply the evidence, and seventh, evaluate the evidence application, will be addressed in a near future.

Findings

Even though the reference to “evidence” is frequent in the DRR field, it is largely based on descriptive processes, anecdotal references, best practices, lessons learned and case studies, and particularly deficient on the subject of informal and precariousness settlements. The evaluation allowed a deep and broad analysis of NA-DRR in urban informal settlements, comparing it with other DRR strategies implemented by different stakeholders in fragile urban settings, assessing the effectiveness and sustainability of the various DRR interventions.

Originality/value

The abundant data, information and knowledge generated will serve as foundation for forthcoming thematic peer-reviewed publications informing evidence-based DRR research, policy and practice, with emphasis on informal and precariousness settlements in particular.

Details

Disaster Prevention and Management: An International Journal, vol. 28 no. 3
Type: Research Article
ISSN: 0965-3562

Keywords

Open Access
Article
Publication date: 12 April 2020

Eliana Andréa Severo, Marcia Marisa Santanna Perin, Julio Cesar Ferro De Guimarães and Elaine Taufer

Environmental problems and natural resources scarcity are changing contemporary organizations management. The current society quest sustainable companies, mostly concern with the…

2593

Abstract

Purpose

Environmental problems and natural resources scarcity are changing contemporary organizations management. The current society quest sustainable companies, mostly concern with the consumption and efficient management of natural resources; those innovative and sustainable companies have the capacity to create innovations and beneficial outcomes for the environment and society. The purpose of this paper is to analyze the relevance of sustainable innovation on products and services innovation, in companies in the northern region of Rio Grande do Sul (Brazil).

Design/methodology/approach

In the research, the authors applied a descriptive and quantitative method, through exploratory factor analysis (EFA), with the use of varimax rotation and multiple linear regression. The final sample of the survey consists of 107 respondents.

Findings

The results indicate that sustainable innovation (SI) has an influence on products and services innovations in the organizations, moreover the process innovations can provide reduced energy consumption and waste emissions, indicating the awareness regarding the environmental issues.

Research limitations/implications

It is emphasized that environmental issues must be linked to investments in environmental education projects in organizations, thus enabling a systemic and effective vision on this issue.

Practical implications

This research presents managerial and academic contributions, as it has developed a scale to measure the importance of SI on products and services innovation.

Originality/value

The study developed a measurement model, with observable variables based on the specialized literature. The measurement model consists of the constructs of product/service innovation and SI, which were statistically validated through the tests of normality, reliability and EFA.

Details

Revista de Gestão, vol. 27 no. 4
Type: Research Article
ISSN: 1809-2276

Keywords

Open Access
Article
Publication date: 30 July 2021

Thiago Ianatoni Camargo, André Luiz Maranhão de Souza-Leão and Bruno Melo Moura

Fans have been characterized as specialized consumers who often express disagreements with the entertainment industry's decisions, especially when it comes to the original content…

1073

Abstract

Purpose

Fans have been characterized as specialized consumers who often express disagreements with the entertainment industry's decisions, especially when it comes to the original content of the works that serve as the basis for the development of media products, evidencing a kind of consumer resistance. Under a Foucauldian perspective aligned with the consumer culture theory (CCT), power relations are established in a dynamic of power exercise and resistance to power. Based on this, the authors pose the following research question: how do fans of media products resist the changes made by the entertainment industry in relation to their canons?

Design/methodology/approach

The authors adopted the Foucault's genealogy of power as a method, analyzing the comments posted on the Westeros.Org website, the main discussion forum of fans of A Song of Ice and Fire (ASoIaF) book series and Game of Thrones (GoT) TV series.

Findings

The findings reveal ways of resistance in relation to the adaptation of the media text permeated by an entertainment dispositif, which considers the adaptation legitimate, and a fannish dispositif, which criticizes the way this adaptation was made. However, their empirical categories reveal that they are forged not only from singularities but also from overlaps. The authors conclude, therefore, that this process occurs in an agonist way, in which conflicts are fought as a reciprocal incitement revealing a productive and ethical relationship.

Originality/value

The agonism shows how consumers can simultaneously be led to incorporate and resist to discourses and market practices. This demonstrates how resistance is not necessarily a force opposed to another, but a dynamic of reciprocal negotiation.

Details

Revista de Gestão, vol. 29 no. 1
Type: Research Article
ISSN: 1809-2276

Keywords

Open Access
Article
Publication date: 21 June 2022

Abhishek Das and Mihir Narayan Mohanty

In time and accurate detection of cancer can save the life of the person affected. According to the World Health Organization (WHO), breast cancer occupies the most frequent…

Abstract

Purpose

In time and accurate detection of cancer can save the life of the person affected. According to the World Health Organization (WHO), breast cancer occupies the most frequent incidence among all the cancers whereas breast cancer takes fifth place in the case of mortality numbers. Out of many image processing techniques, certain works have focused on convolutional neural networks (CNNs) for processing these images. However, deep learning models are to be explored well.

Design/methodology/approach

In this work, multivariate statistics-based kernel principal component analysis (KPCA) is used for essential features. KPCA is simultaneously helpful for denoising the data. These features are processed through a heterogeneous ensemble model that consists of three base models. The base models comprise recurrent neural network (RNN), long short-term memory (LSTM) and gated recurrent unit (GRU). The outcomes of these base learners are fed to fuzzy adaptive resonance theory mapping (ARTMAP) model for decision making as the nodes are added to the F_2ˆa layer if the winning criteria are fulfilled that makes the ARTMAP model more robust.

Findings

The proposed model is verified using breast histopathology image dataset publicly available at Kaggle. The model provides 99.36% training accuracy and 98.72% validation accuracy. The proposed model utilizes data processing in all aspects, i.e. image denoising to reduce the data redundancy, training by ensemble learning to provide higher results than that of single models. The final classification by a fuzzy ARTMAP model that controls the number of nodes depending upon the performance makes robust accurate classification.

Research limitations/implications

Research in the field of medical applications is an ongoing method. More advanced algorithms are being developed for better classification. Still, the scope is there to design the models in terms of better performance, practicability and cost efficiency in the future. Also, the ensemble models may be chosen with different combinations and characteristics. Only signal instead of images may be verified for this proposed model. Experimental analysis shows the improved performance of the proposed model. This method needs to be verified using practical models. Also, the practical implementation will be carried out for its real-time performance and cost efficiency.

Originality/value

The proposed model is utilized for denoising and to reduce the data redundancy so that the feature selection is done using KPCA. Training and classification are performed using heterogeneous ensemble model designed using RNN, LSTM and GRU as base classifiers to provide higher results than that of single models. Use of adaptive fuzzy mapping model makes the final classification accurate. The effectiveness of combining these methods to a single model is analyzed in this work.

Details

Applied Computing and Informatics, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2634-1964

Keywords

Open Access
Article
Publication date: 20 July 2020

E.N. Osegi

In this paper, an emerging state-of-the-art machine intelligence technique called the Hierarchical Temporal Memory (HTM) is applied to the task of short-term load forecasting…

Abstract

In this paper, an emerging state-of-the-art machine intelligence technique called the Hierarchical Temporal Memory (HTM) is applied to the task of short-term load forecasting (STLF). A HTM Spatial Pooler (HTM-SP) stage is used to continually form sparse distributed representations (SDRs) from a univariate load time series data, a temporal aggregator is used to transform the SDRs into a sequential bivariate representation space and an overlap classifier makes temporal classifications from the bivariate SDRs through time. The comparative performance of HTM on several daily electrical load time series data including the Eunite competition dataset and the Polish power system dataset from 2002 to 2004 are presented. The robustness performance of HTM is also further validated using hourly load data from three more recent electricity markets. The results obtained from experimenting with the Eunite and Polish dataset indicated that HTM will perform better than the existing techniques reported in the literature. In general, the robustness test also shows that the error distribution performance of the proposed HTM technique is positively skewed for most of the years considered and with kurtosis values mostly lower than a base value of 3 indicating a reasonable level of outlier rejections.

Details

Applied Computing and Informatics, vol. 17 no. 2
Type: Research Article
ISSN: 2634-1964

Keywords

Content available
Article
Publication date: 1 May 2001

Vincent de P. Roper

80

Abstract

Details

New Library World, vol. 102 no. 4/5
Type: Research Article
ISSN: 0307-4803

Content available
Article
Publication date: 8 July 2022

Vania Vidal, Valéria Magalhães Pequeno, Narciso Moura Arruda Júnior and Marco Antonio Casanova

Enterprise knowledge graphs (EKG) in resource description framework (RDF) consolidate and semantically integrate heterogeneous data sources into a comprehensive dataspace…

Abstract

Purpose

Enterprise knowledge graphs (EKG) in resource description framework (RDF) consolidate and semantically integrate heterogeneous data sources into a comprehensive dataspace. However, to make an external relational data source accessible through an EKG, an RDF view of the underlying relational database, called an RDB2RDF view, must be created. The RDB2RDF view should be materialized in situations where live access to the data source is not possible, or the data source imposes restrictions on the type of query forms and the number of results. In this case, a mechanism for maintaining the materialized view data up-to-date is also required. The purpose of this paper is to address the problem of the efficient maintenance of externally materialized RDB2RDF views.

Design/methodology/approach

This paper proposes a formal framework for the incremental maintenance of externally materialized RDB2RDF views, in which the server computes and publishes changesets, indicating the difference between the two states of the view. The EKG system can then download the changesets and synchronize the externally materialized view. The changesets are computed based solely on the update and the source database state and require no access to the content of the view.

Findings

The central result of this paper shows that changesets computed according to the formal framework correctly maintain the externally materialized RDB2RDF view. The experiments indicate that the proposed strategy supports live synchronization of large RDB2RDF views and that the time taken to compute the changesets with the proposed approach was almost three orders of magnitude smaller than partial rematerialization and three orders of magnitude smaller than full rematerialization.

Originality/value

The main idea that differentiates the proposed approach from previous work on incremental view maintenance is to explore the object-preserving property of typical RDB2RDF views so that the solution can deal with views with duplicates. The algorithms for the incremental maintenance of relational views with duplicates published in the literature require querying the materialized view data to precisely compute the changesets. By contrast, the approach proposed in this paper requires no access to view data. This is important when the view is maintained externally, because accessing a remote data source may be too slow.

Details

International Journal of Web Information Systems, vol. 18 no. 5/6
Type: Research Article
ISSN: 1744-0084

Keywords

1 – 10 of 12