Search results
1 – 10 of 181Cláudio Lima and Ronaldo Santos Mello
NoSQL databases do not require a default schema associated with the data. Even that, they are categorized by data models. A model associated with the data can promote better…
Abstract
Purpose
NoSQL databases do not require a default schema associated with the data. Even that, they are categorized by data models. A model associated with the data can promote better strategies for persistence and manipulation of data in the target database. Based on this motivation, the purpose of this paper is to present an approach for logical design of NoSQL document databases that consists a process that converts a conceptual modeling into efficient logical representations for a NoSQL document database. The authors also evaluate their approach and demonstrate that the generated NoSQL logical structures reduce the amount of data items accessed by queries.
Design/methodology/approach
This paper presents an approach for logical design of NoSQL document database schemas based on a conceptual schema. The authors generate compact and redundancy-free schemas and define appropriate representations in a NoSQL document logical model. The estimated volume of data and workload information can be considered to generate optimized NoSQL document structures.
Findings
This approach was evaluated through a case study with an experimental evaluation in the e-commerce application domain. The results demonstrate that the authors’ workload-based conversion process improves query performance on NoSQL documents by reducing the number of database accesses.
Originality/value
Unlike related work, the reported approach covers all typical conceptual constructs, details a conversion process between conceptual schemas and logical representations for NoSQL document database category and, additionally, considers the estimated database workload to perform optimizations in the logical structure. An experimental evaluation shows that the proposed approach is promising.
Details
Keywords
The paper aims to focus on so‐called NoSQL databases in the context of cloud computing.
Abstract
Purpose
The paper aims to focus on so‐called NoSQL databases in the context of cloud computing.
Design/methodology/approach
Architectures and basic features of these databases are studied, particularly their horizontal scalability and concurrency model, that is mostly weaker than ACID transactions in relational SQL‐like database systems.
Findings
Some characteristics like a data model and querying capabilities of NoSQL databases are discussed in more detail.
Originality/value
The paper shows vary different data models and query possibilities in a common terminology enabling comparison and categorization of NoSQL databases.
Details
Keywords
Aya Khaled Youssef Sayed Mohamed, Dagmar Auer, Daniel Hofer and Josef Küng
Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are…
Abstract
Purpose
Data protection requirements heavily increased due to the rising awareness of data security, legal requirements and technological developments. Today, NoSQL databases are increasingly used in security-critical domains. Current survey works on databases and data security only consider authorization and access control in a very general way and do not regard most of today’s sophisticated requirements. Accordingly, the purpose of this paper is to discuss authorization and access control for relational and NoSQL database models in detail with respect to requirements and current state of the art.
Design/methodology/approach
This paper follows a systematic literature review approach to study authorization and access control for different database models. Starting with a research on survey works on authorization and access control in databases, the study continues with the identification and definition of advanced authorization and access control requirements, which are generally applicable to any database model. This paper then discusses and compares current database models based on these requirements.
Findings
As no survey works consider requirements for authorization and access control in different database models so far, the authors define their requirements. Furthermore, the authors discuss the current state of the art for the relational, key-value, column-oriented, document-based and graph database models in comparison to the defined requirements.
Originality/value
This paper focuses on authorization and access control for various database models, not concrete products. This paper identifies today’s sophisticated – yet general – requirements from the literature and compares them with research results and access control features of current products for the relational and NoSQL database models.
Details
Keywords
Anastasija Nikiforova, Artjoms Daskevics and Otmane Azeroual
Nowadays, there are billions interconnected devices forming Cyber-Physical Systems (CPS), Internet of Things (IoT) and Industrial Internet of Things (IIoT) ecosystems. With an…
Abstract
Nowadays, there are billions interconnected devices forming Cyber-Physical Systems (CPS), Internet of Things (IoT) and Industrial Internet of Things (IIoT) ecosystems. With an increasing number of devices and systems in use, amount and the value of data, the risks of security breaches increase. One of these risks is posed by open data sources, which are databases that are not properly protected. These poorly protected databases are accessible to external actors, which poses a serious risk to the data holder and the results of data-related activities such as analysis, forecasting, monitoring, decision-making, policy development, and the whole contemporary society. This chapter aims at examining the state of the security of open data databases representing both relational databases and NoSQL, with a particular focus on a later category.
Details
Keywords
Sandeep Kumar Singh and Mamata Jenamani
The purpose of this paper is to design a supply chain database schema for Cassandra to store real-time data generated by Radio Frequency IDentification technology in a…
Abstract
Purpose
The purpose of this paper is to design a supply chain database schema for Cassandra to store real-time data generated by Radio Frequency IDentification technology in a traceability system.
Design/methodology/approach
The real-time data generated in such traceability systems are of high frequency and volume, making it difficult to handle by traditional relational database technologies. To overcome this difficulty, a NoSQL database repository based on Casandra is proposed. The efficacy of the proposed schema is compared with two such databases, document-based MongoDB and column family-based Cassandra, which are suitable for storing traceability data.
Findings
The proposed Cassandra-based data repository outperforms the traditional Structured Query Language-based and MongoDB system from the literature in terms of concurrent reading, and works at par with respect to writing and updating of tracing queries.
Originality/value
The proposed schema is able to store the real-time data generated in a supply chain with low latency. To test the performance of the Cassandra-based data repository, a test-bed is designed in the lab and supply chain operations of Indian Public Distribution System are simulated to generate data.
Details
Keywords
Tobias Blanke, Michael Bryant and Reto Speck
In 2010 the European Holocaust Research Infrastructure (EHRI) was funded to support research into the Holocaust. The project follows on from significant efforts in the past to…
Abstract
Purpose
In 2010 the European Holocaust Research Infrastructure (EHRI) was funded to support research into the Holocaust. The project follows on from significant efforts in the past to develop and record the collections of the Holocaust in several national initiatives. The purpose of this paper is to introduce the efforts by EHRI to create a flexible research environment using graph databases. The authors concentrate on the added features and design decisions to enable efficient processing of collection information as a graph.
Design/methodology/approach
The paper concentrates on the specific customisations EHRI had to develop, as the graph database approach is new, and the authors could not rely on existing solutions. The authors describe the serialisations of collections in the graph to provide for efficient processing. Because the EHRI infrastructure is highly distributed, the authors also had to invest a lot of effort into reliable distributed access control mechanisms. Finally, the authors analyse the user-facing work on a portal and a virtual research environment (VRE) in order to discover, share and analyse Holocaust material.
Findings
Using the novel graph database approach, the authors first present how we can model collection information as graphs and why this is effective. Second, we show how we make collection information persistent and describe the complex access management system we have developed. Third, we outline how we integrate user interaction with the data through a VRE.
Originality/value
Scholars require specialised access to information. The authors present the results of the work to develop integrated research with collections on the Holocaust researchers and the proposals for a socio-technical ecosystem based on graph database technologies. The use of graph databases is new and the authors needed to work on several innovative customisations to make them work in the domain.
Details
Keywords
Nathalia Rose Silva da Purificação, Vinícius Barbosa Henrique, Amilton Amorim, Andrea Carneiro and Guilherme Henrique Barros de Souza
The purpose of the study is to compare methodologies for mapping a historic building, with image capture by smartphones and drones, using photogrammetric techniques for…
Abstract
Purpose
The purpose of the study is to compare methodologies for mapping a historic building, with image capture by smartphones and drones, using photogrammetric techniques for three-dimensional (3D) modeling of the structure. Processes and products are also analyzed, as well as possibilities for storing and visualizing data for structuring a cadastre of historical and artistic heritage are studied.
Design/methodology/approach
For mapping with smartphones, the overlapping of photographs was guaranteed, with data acquisition using three different cameras, on the same date as the aerial survey. The models were made from different combinations of camera use. For storage, a conceptual model based on ISO 19.152:2012 is proposed, which was implemented in the MongoDB, resulting in a database for storage. The visualization was carried out on the Cesium ion platform.
Findings
The results indicate that the terrestrial 3D reconstruction using smartphones is an efficient alternative to the historical and artistic cadastre, presenting texture quality superior to the aerial survey in a shorter production time. When dealing with the conceptual model, the LADM (Land Administration Domain Model) standardization guarantees interoperability and facilitates data exchange. In addition, it proved to be flexible for the creation of thematic profiles, supporting their effective storage. The insertion of data in the visualization platform was simple and effective, and it even generated sharing links for visualization of the models.
Originality/value
The study analyses a low-cost method with the use of easily accessible devices, with a combination of methodologies and applied techniques. The data storage and visualization method is also simple and flexible, suitable for application in the cadastre of historical heritage.
Details
Keywords
Akponanabofa Henry Oti, Peter Farrell, Fonbeyin Henry Abanda, Paul McMahon, Abdul-Majeed Mahamadu, Dingayo Mzyece, Adeyemi Ayodele Akintola and Nawal Prinja
The relatively low capital cost and contributions to mitigating global warming have favoured the continuous construction and operation of nuclear power plants (NPPs) across the…
Abstract
Purpose
The relatively low capital cost and contributions to mitigating global warming have favoured the continuous construction and operation of nuclear power plants (NPPs) across the world. One critical phase in the operation of nuclear plants for ensuring the safety and security of radioactive products and by-products is decommissioning. With the advent of digital twinning in the building information modelling (BIM) methodology, efficiency and safety can be improved from context-focus access to regulations pertaining to demolition of structures and the cleaning-up of radioactivity inherent in nuclear stations. The purpose of this study, therefore, is to propose a BIM-driven framework to achieve a more regulation-aware and safer decommissioning of nuclear power plants.
Design/methodology/approach
The framework considers task requirements, and landscape and environmental factors in modelling demolition scenarios that characterise decommissioning processes. The framework integrates decommissioning rules/regulations in a BIM linked non-structured query system to model items and decommissioning tasks, which are implemented based on context-focussed retrieval of decommissioning rules and regulations. The concept’s efficacy is demonstrated using example cases of digitalised NPPs.
Findings
This approach contributes to enhancing improvements in nuclear plant decommissioning with potential for appropriate activity sequencing, risk reduction and ensuring safety.
Originality/value
A BIM-driven framework hinged on querying non-structured databases to provide context-focussed access to nuclear rules and regulations and to aiding decommissioning is new.
Details
Keywords
Priyadarshini R., Latha Tamilselvan and Rajendran N.
The purpose of this paper is to propose a fourfold semantic similarity that results in more accuracy compared to the existing literature. The change detection in the URL and the…
Abstract
Purpose
The purpose of this paper is to propose a fourfold semantic similarity that results in more accuracy compared to the existing literature. The change detection in the URL and the recommendation of the source documents is facilitated by means of a framework in which the fourfold semantic similarity is implied. The latest trends in technology emerge with the continuous growth of resources on the collaborative web. This interactive and collaborative web pretense big challenges in recent technologies like cloud and big data.
Design/methodology/approach
The enormous growth of resources should be accessed in a more efficient manner, and this requires clustering and classification techniques. The resources on the web are described in a more meaningful manner.
Findings
It can be descripted in the form of metadata that is constituted by resource description framework (RDF). Fourfold similarity is proposed compared to three-fold similarity proposed in the existing literature. The fourfold similarity includes the semantic annotation based on the named entity recognition in the user interface, domain-based concept matching and improvised score-based classification of domain-based concept matching based on ontology, sequence-based word sensing algorithm and RDF-based updating of triples. The aggregation of all these similarity measures including the components such as semantic user interface, semantic clustering, and sequence-based classification and semantic recommendation system with RDF updating in change detection.
Research limitations/implications
The existing work suggests that linking resources semantically increases the retrieving and searching ability. Previous literature shows that keywords can be used to retrieve linked information from the article to determine the similarity between the documents using semantic analysis.
Practical implications
These traditional systems also lack in scalability and efficiency issues. The proposed study is to design a model that pulls and prioritizes knowledge-based content from the Hadoop distributed framework. This study also proposes the Hadoop-based pruning system and recommendation system.
Social implications
The pruning system gives an alert about the dynamic changes in the article (virtual document). The changes in the document are automatically updated in the RDF document. This helps in semantic matching and retrieval of the most relevant source with the virtual document.
Originality/value
The recommendation and detection of changes in the blogs are performed semantically using n-triples and automated data structures. User-focussed and choice-based crawling that is proposed in this system also assists the collaborative filtering. Consecutively collaborative filtering recommends the user focussed source documents. The entire clustering and retrieval system is deployed in multi-node Hadoop in the Amazon AWS environment and graphs are plotted and analyzed.
Details
Keywords
Anuoluwapo Ajayi, Lukumon Oyedele, Juan Manuel Davila Delgado, Lukman Akanbi, Muhammad Bilal, Olugbenga Akinade and Oladimeji Olawale
The purpose of this paper is to highlight the use of the big data technologies for health and safety risks analytics in the power infrastructure domain with large data sets of…
Abstract
Purpose
The purpose of this paper is to highlight the use of the big data technologies for health and safety risks analytics in the power infrastructure domain with large data sets of health and safety risks, which are usually sparse and noisy.
Design/methodology/approach
The study focuses on using the big data frameworks for designing a robust architecture for handling and analysing (exploratory and predictive analytics) accidents in power infrastructure. The designed architecture is based on a well coherent health risk analytics lifecycle. A prototype of the architecture interfaced various technology artefacts was implemented in the Java language to predict the likelihoods of health hazards occurrence. A preliminary evaluation of the proposed architecture was carried out with a subset of an objective data, obtained from a leading UK power infrastructure company offering a broad range of power infrastructure services.
Findings
The proposed architecture was able to identify relevant variables and improve preliminary prediction accuracies and explanatory capacities. It has also enabled conclusions to be drawn regarding the causes of health risks. The results represent a significant improvement in terms of managing information on construction accidents, particularly in power infrastructure domain.
Originality/value
This study carries out a comprehensive literature review to advance the health and safety risk management in construction. It also highlights the inability of the conventional technologies in handling unstructured and incomplete data set for real-time analytics processing. The study proposes a technique in big data technology for finding complex patterns and establishing the statistical cohesion of hidden patterns for optimal future decision making.
Details