Search results

1 – 10 of over 102000
Article
Publication date: 21 December 2021

Laouni Djafri

This work can be used as a building block in other settings such as GPU, Map-Reduce, Spark or any other. Also, DDPML can be deployed on other distributed systems such as P2P…

393

Abstract

Purpose

This work can be used as a building block in other settings such as GPU, Map-Reduce, Spark or any other. Also, DDPML can be deployed on other distributed systems such as P2P networks, clusters, clouds computing or other technologies.

Design/methodology/approach

In the age of Big Data, all companies want to benefit from large amounts of data. These data can help them understand their internal and external environment and anticipate associated phenomena, as the data turn into knowledge that can be used for prediction later. Thus, this knowledge becomes a great asset in companies' hands. This is precisely the objective of data mining. But with the production of a large amount of data and knowledge at a faster pace, the authors are now talking about Big Data mining. For this reason, the authors’ proposed works mainly aim at solving the problem of volume, veracity, validity and velocity when classifying Big Data using distributed and parallel processing techniques. So, the problem that the authors are raising in this work is how the authors can make machine learning algorithms work in a distributed and parallel way at the same time without losing the accuracy of classification results. To solve this problem, the authors propose a system called Dynamic Distributed and Parallel Machine Learning (DDPML) algorithms. To build it, the authors divided their work into two parts. In the first, the authors propose a distributed architecture that is controlled by Map-Reduce algorithm which in turn depends on random sampling technique. So, the distributed architecture that the authors designed is specially directed to handle big data processing that operates in a coherent and efficient manner with the sampling strategy proposed in this work. This architecture also helps the authors to actually verify the classification results obtained using the representative learning base (RLB). In the second part, the authors have extracted the representative learning base by sampling at two levels using the stratified random sampling method. This sampling method is also applied to extract the shared learning base (SLB) and the partial learning base for the first level (PLBL1) and the partial learning base for the second level (PLBL2). The experimental results show the efficiency of our solution that the authors provided without significant loss of the classification results. Thus, in practical terms, the system DDPML is generally dedicated to big data mining processing, and works effectively in distributed systems with a simple structure, such as client-server networks.

Findings

The authors got very satisfactory classification results.

Originality/value

DDPML system is specially designed to smoothly handle big data mining classification.

Details

Data Technologies and Applications, vol. 56 no. 4
Type: Research Article
ISSN: 2514-9288

Keywords

Article
Publication date: 20 November 2017

Shanyong Wang, Jun Li and Dingtao Zhao

The purpose of this paper is to apply an extended technology acceptance model to examine the medical data analyst’s intention to use medical big data processing technique.

Abstract

Purpose

The purpose of this paper is to apply an extended technology acceptance model to examine the medical data analyst’s intention to use medical big data processing technique.

Design/methodology/approach

Questionnaire survey method was used to collect data from 293 medical data analysts and analyzed with the assistance of structural equation modeling.

Findings

The results indicate that the perceived usefulness, social influence and attitude are important to the intention to use medical big data processing technique, and the direct effect of perceived usefulness on intention to use is greater than social influence and attitude. The perceived usefulness is influenced by perceived ease of use. Attitude is influenced by perceived usefulness, and attitude acts as a mediator between perceived usefulness and usage intention. Unexpectedly, attitude is not influenced by perceived ease of use and social influence.

Originality/value

This research examines the medical data analyst’s intention to use medical big data processing technique and provides several implications for using medical big data processing technique.

Article
Publication date: 6 January 2022

Ahmad Latifian

Big data has posed problems for businesses, the Information Technology (IT) sector and the science community. The problems posed by big data can be effectively addressed using…

Abstract

Purpose

Big data has posed problems for businesses, the Information Technology (IT) sector and the science community. The problems posed by big data can be effectively addressed using cloud computing and associated distributed computing technology. Cloud computing and big data are two significant past-year problems that allow high-efficiency and competitive computing tools to be delivered as IT services. The paper aims to examine the role of the cloud as a tool for managing big data in various aspects to help businesses.

Design/methodology/approach

This paper delivers solutions in the cloud for storing, compressing, analyzing and processing big data. Hence, articles were divided into four categories: articles on big data storage, articles on big data processing, articles on analyzing and finally, articles on data compression in cloud computing. This article is based on a systematic literature review. Also, it is based on a review of 19 published papers on big data.

Findings

From the results, it can be inferred that cloud computing technology has features that can be useful for big data management. Challenging issues are raised in each section. For example, in storing big data, privacy and security issues are challenging.

Research limitations/implications

There were limitations to this systematic review. The first limitation is that only English articles were reviewed. Also, articles that matched the keywords were used. Finally, in this review, authoritative articles were reviewed, and slides and tutorials were avoided.

Practical implications

The research presents new insight into the business value of cloud computing in interfirm collaborations.

Originality/value

Previous research has often examined other aspects of big data in the cloud. This article takes a new approach to the subject. It allows big data researchers to comprehend the various aspects of big data management in the cloud. In addition, setting an agenda for future research saves time and effort for readers searching for topics within big data.

Details

Kybernetes, vol. 51 no. 6
Type: Research Article
ISSN: 0368-492X

Keywords

Book part
Publication date: 18 July 2022

Teena Pareek, Kiran Sood and Simon Grima

Introduction: New ideas and concepts of big data have emerged in recent years in response to the astounding growth of data in many industries. Furthermore, the phenomenal increase…

Abstract

Introduction: New ideas and concepts of big data have emerged in recent years in response to the astounding growth of data in many industries. Furthermore, the phenomenal increase in the use of the internet and social media has added enormous amounts of data to conventional data processing systems. Still, it has also created challenges for traditional data processing.

Purpose: A significant characteristic of the insurance sector is critically dependent on information. This sector generates a great deal of structured and unstructured data, which traditional data processing techniques cannot handle. As compared to conventional insurance data processing and decision-making requirements, this lesson shows an analysis of data technology’s value additions.

Research methodology: The author assesses the primary use of cases for data in the insurance industry via a case study analysis. From the perspective of the insurance sector, this chapter examines the concepts, technologies, and tools of big data. A few analytical reviews by the insurance company are also provided, which justified several gains gained either through inefficient processing of massive, diverse data sets or by supporting better decisions.

Findings: This chapter demonstrates the importance of adopting new business models that allow insurers to move beyond understand and protect and become more predictive and preventative by using the tools and technologies of big data technology.

Details

Big Data Analytics in the Insurance Market
Type: Book
ISBN: 978-1-80262-638-4

Keywords

Article
Publication date: 30 August 2022

Yang Liu, Wei Fang, Taiwen Feng and Na Gao

Based on organizational information processing theory, this research explores how big data analytics capability (BDAC) contributes to green supply chain integration (GSCI) and the…

1073

Abstract

Purpose

Based on organizational information processing theory, this research explores how big data analytics capability (BDAC) contributes to green supply chain integration (GSCI) and the contingency role that data-driven decision culture plays.

Design/methodology/approach

Using the two-wave survey data collected from 317 Chinese manufacturing firms, the authors validate the hypotheses.

Findings

The results show that big data managerial capability has positive impacts on three dimensions of GSCI, while big data technical capability has positive impacts on green internal and customer integration. Moreover, green internal integration mediates the impacts of big data technical capability and managerial capability on green supplier and customer integration. Finally, data-driven decision culture alleviates the positive impacts of big data technical and managerial capability on green internal integration.

Practical implications

The findings suggest that firms can leverage big data technical and managerial capability to enhance information processing capability for achieving a higher degree of GSCI. Further, the critical role of data-driven decision culture in affecting the link between BDAC and GSCI should not be overlooked.

Originality/value

This research contributes to literature on green supply chain management by revealing the role of BDAC in improving GSCI.

Details

Industrial Management & Data Systems, vol. 122 no. 11
Type: Research Article
ISSN: 0263-5577

Keywords

Article
Publication date: 8 February 2021

Gianluca Solazzo, Gianluca Elia and Giuseppina Passiante

This study aims to investigate the Big Social Data (BSD) paradigm, which still lacks a clear and shared definition, and causes a lack of clarity and understanding about its…

Abstract

Purpose

This study aims to investigate the Big Social Data (BSD) paradigm, which still lacks a clear and shared definition, and causes a lack of clarity and understanding about its beneficial opportunities for practitioners. In the knowledge management (KM) domain, a clear characterization of the BSD paradigm can lead to more effective and efficient KM strategies, processes and systems that leverage a huge amount of structured and unstructured data sources.

Design/methodology/approach

The study adopts a systematic literature review (SLR) methodology based on a mixed analysis approach (unsupervised machine learning and human-based) applied to 199 research articles on BSD topics extracted from Scopus and Web of Science. In particular, machine learning processing has been implemented by using topic extraction and hierarchical clustering techniques.

Findings

The paper provides a threefold contribution: a conceptualization and a consensual definition of the BSD paradigm through the identification of four key conceptual pillars (i.e. sources, properties, technology and value exploitation); a characterization of the taxonomy of BSD data type that extends previous works on this topic; a research agenda for future research studies on BSD and its applications along with a KM perspective.

Research limitations/implications

The main limits of the research rely on the list of articles considered for the literature review that could be enlarged by considering further sources (in addition to Scopus and Web of Science) and/or further languages (in addition to English) and/or further years (the review considers papers published until 2018). Research implications concern the development of a research agenda organized along with five thematic issues, which can feed future research to deepen the paradigm of BSD and explore linkages with the KM field.

Practical implications

Practical implications concern the usage of the proposed definition of BSD to purposefully design applications and services based on BSD in knowledge-intensive domains to generate value for citizens, individuals, companies and territories.

Originality/value

The original contribution concerns the definition of the big data social paradigm built through an SLR the combines machine learning processing and human-based processing. Moreover, the research agenda deriving from the study contributes to investigate the BSD paradigm in the wider domain of KM.

Details

Journal of Knowledge Management, vol. 25 no. 7
Type: Research Article
ISSN: 1367-3270

Keywords

Article
Publication date: 8 February 2016

Zhihua Li, Zianfei Tang and Yihua Yang

The high-efficient processing of mass data is a primary issue in building and maintaining security video surveillance system. This paper aims to focus on the architecture of…

Abstract

Purpose

The high-efficient processing of mass data is a primary issue in building and maintaining security video surveillance system. This paper aims to focus on the architecture of security video surveillance system, which was based on Hadoop parallel processing technology in big data environment.

Design/methodology/approach

A hardware framework of security video surveillance network cascaded system (SVSNCS) was constructed on the basis of Internet of Things, network cascade technology and Hadoop platform. Then, the architecture model of SVSNCS was proposed using the Hadoop and big data processing platform.

Findings

Finally, we suggested the procedure of video processing according to the cascade network characteristics.

Originality/value

Our paper, which focused on the architecture of security video surveillance system in big data environment on the basis of Hadoop parallel processing technology, provided high-quality video surveillance services for security area.

Details

World Journal of Engineering, vol. 13 no. 1
Type: Research Article
ISSN: 1708-5284

Keywords

Article
Publication date: 23 March 2023

Mohd Naz’ri Mahrin, Anusuyah Subbarao, Suriayati Chuprat and Nur Azaliah Abu Bakar

Cloud computing promises dependable services offered through next-generation data centres based on virtualization technologies for computation, network and storage. Big Data

Abstract

Purpose

Cloud computing promises dependable services offered through next-generation data centres based on virtualization technologies for computation, network and storage. Big Data Applications have been made viable by cloud computing technologies due to the tremendous expansion of data. Disaster management is one of the areas where big data applications are rapidly being deployed. This study looks at how big data is being used in conjunction with cloud computing to increase disaster risk reduction (DRR). This paper aims to explore and review the existing framework for big data used in disaster management and to provide an insightful view of how cloud-based big data platform toward DRR is applied.

Design/methodology/approach

A systematic mapping study is conducted to answer four research questions with papers related to Big Data Analytics, cloud computing and disaster management ranging from the year 2013 to 2019. A total of 26 papers were finalised after going through five steps of systematic mapping.

Findings

Findings are based on each research question.

Research limitations/implications

A specific study on big data platforms on the application of disaster management, in general is still limited. The lack of study in this field is opened for further research sources.

Practical implications

In terms of technology, research in DRR leverage on existing big data platform is still lacking. In terms of data, many disaster data are available, but scientists still struggle to learn and listen to the data and take more proactive disaster preparedness.

Originality/value

This study shows that a very famous platform selected by researchers is central processing unit based processing, namely, Apache Hadoop. Apache Spark which uses memory processing requires a big capacity of memory, therefore this is less preferred in the world of research.

Details

Journal of Science and Technology Policy Management, vol. 14 no. 6
Type: Research Article
ISSN: 2053-4620

Keywords

Article
Publication date: 19 February 2021

C. Lakshmi and K. Usha Rani

Resilient distributed processing technique (RDPT), in which mapper and reducer are simplified with the Spark contexts and support distributed parallel query processing.

Abstract

Purpose

Resilient distributed processing technique (RDPT), in which mapper and reducer are simplified with the Spark contexts and support distributed parallel query processing.

Design/methodology/approach

The proposed work is implemented with Pig Latin with Spark contexts to develop query processing in a distributed environment.

Findings

Query processing in Hadoop influences the distributed processing with the MapReduce model. MapReduce caters to the works on different nodes with the implementation of complex mappers and reducers. Its results are valid for some extent size of the data.

Originality/value

Pig supports the required parallel processing framework with the following constructs during the processing of queries: FOREACH; FLATTEN; COGROUP.

Details

International Journal of Intelligent Computing and Cybernetics, vol. 14 no. 2
Type: Research Article
ISSN: 1756-378X

Keywords

Book part
Publication date: 18 July 2022

Manish Bhardwaj and Shivani Agarwal

Introduction: In recent years, fresh big data ideas and concepts have emerged to address the massive increase in data volumes in several commercial areas. Meanwhile, the…

Abstract

Introduction: In recent years, fresh big data ideas and concepts have emerged to address the massive increase in data volumes in several commercial areas. Meanwhile, the phenomenal development of internet use and social media has not only added to the enormous volumes of data available but has also posed new hurdles to traditional data processing methods. For example, the insurance industry is known for being data-driven, as it generates massive volumes of accumulated material, both structured and unstructured, that typical data processing techniques can’t handle.

Purpose: In this study, the authors compare the benefits of big data technologies to the needs for insurance data processing and decision-making. There is also a case study evaluation concentrating on the primary use cases of big data in the insurance business.

Methodology: This chapter examines the essential big data technologies and tools from the insurance industry’s perspective. The study also included an analytical analysis that supported several gains made by insurance companies, such as more efficient processing of large, heterogeneous data sets or better decision-making support. In addition, the study examines in depth the top seven use cases of big data in insurance and justifying their use and adding value. Finally, it also reviewed contemporary big data technologies and tools, concentrating on their key concepts and recommended applications in the insurance business through examples.

Findings: The study has demonstrated the value of implementing big data technologies and tools, which enable the development of powerful new business models, allowing insurance to advance from ‘understand and protect’ to ‘predict and prevent’.

Details

Big Data Analytics in the Insurance Market
Type: Book
ISBN: 978-1-80262-638-4

Keywords

1 – 10 of over 102000