Search results

1 – 10 of 22
Book part
Publication date: 11 September 2020

D. K. Malhotra, Kunal Malhotra and Rashmi Malhotra

Traditionally, loan officers use different credit scoring models to complement judgmental methods to classify consumer loan applications. This study explores the use of decision…

Abstract

Traditionally, loan officers use different credit scoring models to complement judgmental methods to classify consumer loan applications. This study explores the use of decision trees, AdaBoost, and support vector machines (SVMs) to identify potential bad loans. Our results show that AdaBoost does provide an improvement over simple decision trees as well as SVM models in predicting good credit clients and bad credit clients. To cross-validate our results, we use k-fold classification methodology.

Book part
Publication date: 18 July 2022

Yakub Kayode Saheed, Usman Ahmad Baba and Mustafa Ayobami Raji

Purpose: This chapter aims to examine machine learning (ML) models for predicting credit card fraud (CCF).Need for the study: With the advance of technology, the world is…

Abstract

Purpose: This chapter aims to examine machine learning (ML) models for predicting credit card fraud (CCF).

Need for the study: With the advance of technology, the world is increasingly relying on credit cards rather than cash in daily life. This creates a slew of new opportunities for fraudulent individuals to abuse these cards. As of December 2020, global card losses reached $28.65billion, up 2.9% from $27.85 billion in 2018, according to the Nilson 2019 research. To safeguard the safety of credit card users, the credit card issuer should include a service that protects customers from potential risks. CCF has become a severe threat as internet buying has grown. To this goal, various studies in the field of automatic and real-time fraud detection are required. Due to their advantageous properties, the most recent ones employ a variety of ML algorithms and techniques to construct a well-fitting model to detect fraudulent transactions. When it comes to recognising credit card risk is huge and high-dimensional data, feature selection (FS) is critical for improving classification accuracy and fraud detection.

Methodology/design/approach: The objectives of this chapter are to construct a new model for credit card fraud detection (CCFD) based on principal component analysis (PCA) for FS and using supervised ML techniques such as K-nearest neighbour (KNN), ridge classifier, gradient boosting, quadratic discriminant analysis, AdaBoost, and random forest for classification of fraudulent and legitimate transactions. When compared to earlier experiments, the suggested approach demonstrates a high capacity for detecting fraudulent transactions. To be more precise, our model’s resilience is constructed by integrating the power of PCA for determining the most useful predictive features. The experimental analysis was performed on German credit card and Taiwan credit card data sets.

Findings: The experimental findings revealed that the KNN achieved an accuracy of 96.29%, recall of 100%, and precision of 96.29%, which is the best performing model on the German data set. While the ridge classifier was the best performing model on Taiwan Credit data with an accuracy of 81.75%, recall of 34.89, and precision of 66.61%.

Practical implications: The poor performance of the models on the Taiwan data revealed that it is an imbalanced credit card data set. The comparison of our proposed models with state-of-the-art credit card ML models showed that our results were competitive.

Book part
Publication date: 6 September 2019

Son Nguyen, Gao Niu, John Quinn, Alan Olinsky, Jonathan Ormsbee, Richard M. Smith and James Bishop

In recent years, the problem of classification with imbalanced data has been growing in popularity in the data-mining and machine-learning communities due to the emergence of an…

Abstract

In recent years, the problem of classification with imbalanced data has been growing in popularity in the data-mining and machine-learning communities due to the emergence of an abundance of imbalanced data in many fields. In this chapter, we compare the performance of six classification methods on an imbalanced dataset under the influence of four resampling techniques. These classification methods are the random forest, the support vector machine, logistic regression, k-nearest neighbor (KNN), the decision tree, and AdaBoost. Our study has shown that all of the classification methods have difficulty when working with the imbalanced data, with the KNN performing the worst, detecting only 27.4% of the minority class. However, with the help of resampling techniques, all of the classification methods experience improvement on overall performances. In particular, the Random Forest, in combination with the random over-sampling technique, performs the best, achieving 82.8% balanced accuracy (the average of the true-positive rate and true-negative rate).

We then propose a new procedure to resample the data. Our method is based on the idea of eliminating “easy” majority observations before under-sampling them. It has further improved the balanced accuracy of the Random Forest to 83.7%, making it the best approach for the imbalanced data.

Details

Advances in Business and Management Forecasting
Type: Book
ISBN: 978-1-78754-290-7

Keywords

Book part
Publication date: 24 March 2006

Valeriy V. Gavrishchaka

Increasing availability of the financial data has opened new opportunities for quantitative modeling. It has also exposed limitations of the existing frameworks, such as low…

Abstract

Increasing availability of the financial data has opened new opportunities for quantitative modeling. It has also exposed limitations of the existing frameworks, such as low accuracy of the simplified analytical models and insufficient interpretability and stability of the adaptive data-driven algorithms. I make the case that boosting (a novel, ensemble learning technique) can serve as a simple and robust framework for combining the best features of the analytical and data-driven models. Boosting-based frameworks for typical financial and econometric applications are outlined. The implementation of a standard boosting procedure is illustrated in the context of the problem of symbolic volatility forecasting for IBM stock time series. It is shown that the boosted collection of the generalized autoregressive conditional heteroskedastic (GARCH)-type models is systematically more accurate than both the best single model in the collection and the widely used GARCH(1,1) model.

Details

Econometric Analysis of Financial and Economic Time Series
Type: Book
ISBN: 978-1-84950-388-4

Book part
Publication date: 15 March 2021

Jochen Hartmann

Across disciplines, researchers and practitioners employ decision tree ensembles such as random forests and XGBoost with great success. What explains their popularity? This…

Abstract

Across disciplines, researchers and practitioners employ decision tree ensembles such as random forests and XGBoost with great success. What explains their popularity? This chapter showcases how marketing scholars and decision-makers can harness the power of decision tree ensembles for academic and practical applications. The author discusses the origin of decision tree ensembles, explains their theoretical underpinnings, and illustrates them empirically using a real-world telemarketing case, with the objective of predicting customer conversions. Readers unfamiliar with decision tree ensembles will learn to appreciate them for their versatility, competitive accuracy, ease of application, and computational efficiency and will gain a comprehensive understanding why decision tree ensembles contribute to every data scientist's methodological toolbox.

Details

The Machine Age of Customer Insight
Type: Book
ISBN: 978-1-83909-697-6

Keywords

Book part
Publication date: 10 February 2023

Saurabh Sharma and Romica Bhat

Need of the Study: Artificial intelligence (AI) can be regarded as a big leap in the case of technological advancement. Developments in AI have profound implications for economic…

Abstract

Need of the Study: Artificial intelligence (AI) can be regarded as a big leap in the case of technological advancement. Developments in AI have profound implications for economic sectors and on the societal level. In contemporary times, AI is applied widely in assisting organisations in informing managerial decisions, organisational goals, and business strategies. One can very well witness the interest of human resource (HR) professionals in the implementation of AI for the formulation of HR policies and future frameworks. In the past few years, various research works have been carried out on how these two critical branches can be combined for bringing out the best in human resource management (HRM). The fundamental explanation for this is found in every organisation’s most important management aim is employee retention and elevation.

Purpose: In this direction, this chapter will try to analyse the probability of employees leaving the company, the key drivers behind it, recommendations or strategies that can be implemented in improving employee retention, elevation predictions with the help of different features of machine learning, and the possibility of some other techniques other than key performance indicators (KPI), and rating and training score in this field.

Methodology: The goal will be achieved with the help of implementing machine learning-based classification tools and an ensemble learning approach to the data set of the corporate sector.

Findings: Machine learning techniques can be utilised to develop reliable models to find different factors for elevation and employee attrition.

Details

The Adoption and Effect of Artificial Intelligence on Human Resources Management, Part A
Type: Book
ISBN: 978-1-80382-027-9

Keywords

Book part
Publication date: 1 September 2021

Alicia T. Lamere, Son Nguyen, Gao Niu, Alan Olinsky and John Quinn

Predicting a patient's length of stay (LOS) in a hospital setting has been widely researched. Accurately predicting an individual's LOS can have a significant impact on a…

Abstract

Predicting a patient's length of stay (LOS) in a hospital setting has been widely researched. Accurately predicting an individual's LOS can have a significant impact on a healthcare provider's ability to care for individuals by allowing them to properly prepare and manage resources. A hospital's productivity requires a delicate balance of maintaining enough staffing and resources without being overly equipped or wasteful. This has become even more important in light of the current COVID-19 pandemic, during which emergency departments around the globe have been inundated with patients and are struggling to manage their resources.

In this study, the authors focus on the prediction of LOS at the time of admission in emergency departments at Rhode Island hospitals through discharge data obtained from the Rhode Island Department of Health over the time period of 2012 and 2013. This work also explores the distribution of discharge dispositions in an effort to better characterize the resources patients require upon leaving the emergency department.

Book part
Publication date: 18 July 2022

Kamal Gulati and Pallavi Seth

Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency…

Abstract

Introduction: With many new technologies requiring real-time data processing, cloud computing has become challenging to implement due to high bandwidth and high latency requirements.

Purpose: To overcome this issue, edge computing is used to process data at the network’s edge. Edge computing is a distributed computing paradigm that brings computation and data storage closer to the location where it is needed. It is used to process time-sensitive data.

Methodology: The authors implemented the model using Linux Foundation’s open-source platform EdgeX Foundry to create an edge-computing device. The model involved getting data from an on-board sensor (on-board diagnostics (OBD-II)) and the GPS sensor of a car. The data are then observed and computed to the EdgeX server. The single server will send data to serve three real-life internet of things (IoT) use cases: auto insurance, supporting a smart city, and building a personal driving record.

Findings: The main aim of this model is to illustrate how edge computing can improve both latency and bandwidth usage needed for real-world IoT applications.

Details

Big Data Analytics in the Insurance Market
Type: Book
ISBN: 978-1-80262-638-4

Keywords

Content available
Book part
Publication date: 11 September 2020

Abstract

Details

Applications of Management Science
Type: Book
ISBN: 978-1-83867-001-6

Content available
Book part
Publication date: 6 September 2019

Abstract

Details

Advances in Business and Management Forecasting
Type: Book
ISBN: 978-1-78754-290-7

Access

Year

Content type

Book part (22)
1 – 10 of 22