Search results

1 – 10 of 321
Open Access
Article
Publication date: 17 October 2019

Mahmoud ELsayed and Amr Soliman

The purpose of this study is to estimate the linear regression parameters using two alternative techniques. First technique is to apply the generalized linear model (GLM) and the…

3175

Abstract

Purpose

The purpose of this study is to estimate the linear regression parameters using two alternative techniques. First technique is to apply the generalized linear model (GLM) and the second technique is the Markov Chain Monte Carlo (MCMC) method.

Design/methodology/approach

In this paper, the authors adopted the incurred claims of Egyptian non-life insurance market as a dependent variable during a 10-year period. MCMC uses Gibbs sampling to generate a sample from a posterior distribution of a linear regression to estimate the parameters of interest. However, the authors used the R package to estimate the parameters of the linear regression using the above techniques.

Findings

These procedures will guide the decision-maker for estimating the reserve and set proper investment strategy.

Originality/value

In this paper, the authors will estimate the parameters of a linear regression model using MCMC method via R package. Furthermore, MCMC uses Gibbs sampling to generate a sample from a posterior distribution of a linear regression to estimate parameters to predict future claims. In the same line, these procedures will guide the decision-maker for estimating the reserve and set proper investment strategy.

Details

Journal of Humanities and Applied Social Sciences, vol. 2 no. 1
Type: Research Article
ISSN: 2632-279X

Keywords

Article
Publication date: 24 August 2020

Youngkeun Choi and Jae Won Choi

Job involvement can be linked with important work outcomes. One way for organizations to increase job involvement is to use machine learning technology to predict employees’ job…

Abstract

Purpose

Job involvement can be linked with important work outcomes. One way for organizations to increase job involvement is to use machine learning technology to predict employees’ job involvement, so that their leaders of human resource (HR) management can take proactive measures or plan succession for preservation. This paper aims to develop a reliable job involvement prediction model using machine learning technique.

Design/methodology/approach

This study used the data set, which is available at International Business Machines (IBM) Watson Analytics in IBM community and applied a generalized linear model (GLM) including linear regression and binomial classification. This study essentially had two primary approaches. First, this paper intends to understand the role of variables in job involvement prediction modeling better. Second, the study seeks to evaluate the predictive performance of GLM including linear regression and binomial classification.

Findings

In these results, first, employees’ job involvement with a lot of individual factors can be predicted. Second, for each model, this model showed the outstanding predictive performance.

Practical implications

The pre-access and modeling methodology used in this paper can be viewed as a roadmap for the reader to follow the steps taken in this study and to apply procedures to identify the causes of many other HR management problems.

Originality/value

This paper is the first one to attempt to come up with the best-performing model for predicting job involvement based on a limited set of features including employees’ demographics using machine learning technique.

Details

International Journal of Organizational Analysis, vol. 29 no. 3
Type: Research Article
ISSN: 1934-8835

Keywords

Book part
Publication date: 12 December 2003

James W. Hardin

This article examines the history, development, and application of the sandwich estimate of variance. In describing this estimator, we pay attention to applications that have…

Abstract

This article examines the history, development, and application of the sandwich estimate of variance. In describing this estimator, we pay attention to applications that have appeared in the literature and examine the nature of the problems for which this estimator is used. We describe various adjustments to the estimate for use with small samples, and illustrate the estimator’s construction for a variety of models. Finally, we discuss interpretation of results.

Details

Maximum Likelihood Estimation of Misspecified Models: Twenty Years Later
Type: Book
ISBN: 978-1-84950-253-5

Book part
Publication date: 30 August 2019

Zhe Yu, Raquel Prado, Steve C. Cramer, Erin B. Quinlan and Hernando Ombao

We develop a Bayesian approach for modeling brain activation and connectivity from functional magnetic resonance image (fMRI) data. Our approach simultaneously estimates local…

Abstract

We develop a Bayesian approach for modeling brain activation and connectivity from functional magnetic resonance image (fMRI) data. Our approach simultaneously estimates local hemodynamic response functions (HRFs) and activation parameters, as well as global effective and functional connectivity parameters. Existing methods assume identical HRFs across brain regions, which may lead to erroneous conclusions in inferring activation and connectivity patterns. Our approach addresses this limitation by estimating region-specific HRFs. Additionally, it enables neuroscientists to compare effective connectivity networks for different experimental conditions. Furthermore, the use of spike and slab priors on the connectivity parameters allows us to directly select significant effective connectivities in a given network.

We include a simulation study that demonstrates that, compared to the standard generalized linear model (GLM) approach, our model generally has higher power and lower type I error and bias than the GLM approach, and it also has the ability to capture condition-specific connectivities. We applied our approach to a dataset from a stroke study and found different effective connectivity patterns for task and rest conditions in certain brain regions of interest (ROIs).

Details

Topics in Identification, Limited Dependent Variables, Partial Observability, Experimentation, and Flexible Modeling: Part A
Type: Book
ISBN: 978-1-78973-241-2

Keywords

Book part
Publication date: 18 April 2018

Mohammed Quddus

Purpose – Time-series regression models are applied to analyse transport safety data for three purposes: (1) to develop a relationship between transport accidents (or incidents…

Abstract

Purpose – Time-series regression models are applied to analyse transport safety data for three purposes: (1) to develop a relationship between transport accidents (or incidents) and various time-varying factors, with the aim of identifying the most important factors; (2) to develop a time-series accident model in forecasting future accidents for the given values of future time-varying factors and (3) to evaluate the impact of a system-wide policy, education or engineering intervention on accident counts. Regression models for analysing transport safety data are well established, especially in analysing cross-sectional and panel datasets. There is, however, a dearth of research relating to time-series regression models in the transport safety literature. The purpose of this chapter is to examine existing literature with the aim of identifying time-series regression models that have been employed in safety analysis in relation to wider applications. The aim is to identify time-series regression models that are applicable in analysing disaggregated accident counts.

Methodology/Approach – There are two main issues in modelling time-series accident counts: (1) a flexible approach in addressing serial autocorrelation inherent in time-series processes of accident counts and (2) the fact that the conditional distribution (conditioned on past observations and covariates) of accident counts follow a Poisson-type distribution. Various time-series regression models are explored to identify the models most suitable for analysing disaggregated time-series accident datasets. A recently developed time-series regression model – the generalised linear autoregressive and moving average (GLARMA) – has been identified as the best model to analyse safety data.

Findings – The GLARMA model was applied to a time-series dataset of airproxes (aircraft proximity) that indicate airspace safety in the United Kingdom. The aim was to evaluate the impact of an airspace intervention (i.e., the introduction of reduced vertical separation minima, RVSM) on airspace safety while controlling for other factors, such as air transport movements (ATMs) and seasonality. The results indicate that the GLARMA model is more appropriate than a generalised linear model (e.g., Poisson or Poisson-Gamma), and it has been found that the introduction of RVSM has reduced the airprox events by 15%. In addition, it was found that a 1% increase in ATMs within UK airspace would lead to a 1.83% increase in monthly airproxes in UK airspace.

Practical applications – The methodology developed in this chapter is applicable to many time-series processes of accident counts. The models recommended in this chapter could be used to identify different time-varying factors and to evaluate the effectiveness of various policy and engineering interventions on transport safety or similar data (e.g., crimes).

Originality/value of paper – The GLARMA model has not been properly explored in modelling time-series safety data. This new class of model has been applied to a dataset in evaluating the effectiveness of an intervention. The model recommended in this chapter would greatly benefit researchers and analysts working with time-series data.

Details

Safe Mobility: Challenges, Methodology and Solutions
Type: Book
ISBN: 978-1-78635-223-1

Keywords

Article
Publication date: 12 January 2015

Ângelo Márcio Oliveira Sant'Anna

The purpose of this paper is to propose a framework of decision making to aid practitioners in modeling and optimization experimental data for improvement quality of industrial…

Abstract

Purpose

The purpose of this paper is to propose a framework of decision making to aid practitioners in modeling and optimization experimental data for improvement quality of industrial processes, reinforcing idea that planning and conducting data modeling are as important as formal analysis.

Design/methodology/approach

The paper presents an application was carried out about the modeling of experimental data at mining company, with support at Catholic University from partnership projects. The literature seems to be more focussed on the data analysis than on providing a sequence of operational steps or decision support which would lead to the best regression model given for the problem that researcher is confronted with. The authors use the concept of statistical regression technique called generalized linear models.

Findings

The authors analyze the relevant case study in mining company, based on best statistical regression models. Starting from this analysis, the results of the industrial case study illustrates the strong relationship of the improvement process with the presented framework approach into practice. Moreover, the case study consolidating a fundamental advantage of regression models: modeling guided provides more knowledge about products, processes and technologies, even in unsuccessful case studies.

Research limitations/implications

The study advances in regression model for data modeling are applicable in several types of industrial processes and phenomena random. It is possible to find unsuccessful data modeling due to lack of knowledge of statistical technique.

Originality/value

An essential point is that the study is based on the feedback from practitioners and industrial managers, which makes the analyses and conclusions from practical points of view, without relevant theoretical knowledge of relationship among the process variables. Regression model has its own characteristics related to response variable and factors, and misspecification of the regression model or their components can yield inappropriate inferences and erroneous experimental results.

Article
Publication date: 30 November 2021

Julián Martínez-Vargas, Pedro Carmona and Pol Torrelles

The purpose of this paper is to study the influence of different quantitative (traditionally used) and qualitative variables, such as the possible negative effect in determined…

Abstract

Purpose

The purpose of this paper is to study the influence of different quantitative (traditionally used) and qualitative variables, such as the possible negative effect in determined periods of certain socio-political factors on share price formation.

Design/methodology/approach

We first analyse descriptively the evolution of the Ibex-35 in recent years and compare it with other international benchmark indices. Bellow, two techniques have been compared: a classic linear regression statistical model (GLM) and a method based on machine learning techniques called Extreme Gradient Boosting (XGBoost).

Findings

XGBoost yields a very accurate market value prediction model that clearly outperforms the other, with a coefficient of determination close to 90%, calculated on validation sets.

Practical implications

According to our analysis, individual accounts are equally or more important than consolidated information in predicting the behaviour of share prices. This would justify Spain maintaining the obligation to present individual interim financial statements, which does not happen in other European Union countries because IAS 34 only stipulates consolidated interim financial statements.

Social implications

The descriptive analysis allows us to see how the Ibex-35 has moved away from international trends, especially in periods in which some relevant socio-political events occurred, such as the independence referendum in Catalonia, the double elections of 2019 or the early handling of the Covid-19 pandemic in 2020.

Originality/value

Compared to other variables, the XGBoost model assigns little importance to socio-political factors when it comes to share price formation; however, this model explains 89.33% of its variance.

Propósito

El propósito de este artículo es estudiar la influencia de diferentes variables cuantitativas (tradicionalmente usadas) y cualitativas, como la posible influencia negativa en determinados períodos de ciertos factores sociopolíticos, sobre la formación del precio de.

Diseño/metodología/enfoque

Primero analizamos de forma descriptiva la evolución del Ibex-35 en los últimos años y la comparamos con la de otros índices internacionales de referencia. A continuación, se han contrastado dos técnicas: un modelo estadístico clásico de regresión lineal (GLM) y un método basado en el aprendizaje automático denominado Extreme Gradient Boosting (XGBoost).

Resultados

XGBoost nos permite obtener un modelo de predicción del valor de mercado muy preciso y claramente superior al otro, con un coeficiente de determinación cercano al 90%, calculado sobre las muestras de validación.

Implicaciones prácticas

De acuerdo con nuestro análisis, la información contable individual es igual o más importante que la consolidada para predecir el comportamiento del precio de las acciones. Esto justificaría que España mantenga la obligación de presentar estados financieros intermedios individuales, lo que no ocurre en otros países de la Unión Europea porque la NIC 34 solo obliga a realizar estados financieros intermedios consolidados.

Implicaciones sociales

El análisis descriptivo permite ver cómo el Ibex-35 se ha alejado de las tendencias internacionales, especialmente en periodos en los que se produjo algún hecho sociopolítico relevante, como el referéndum de autodeterminación de Cataluña, el doble proceso electoral de 2019 o la gestión inicial de la pandemia generada por el Covid-19.

Originalidad/valor

En comparación con otras variables, el modelo XGBoost asigna poca importancia a los factores sociopolíticos cuando se trata de la formación del precio de las acciones; sin embargo, este modelo explica el 89.33% de su varianza.

Details

Academia Revista Latinoamericana de Administración, vol. 35 no. 1
Type: Research Article
ISSN: 1012-8255

Keywords

Article
Publication date: 24 July 2018

Marcelo Cajias

This paper aims to explore the in-sample explanatory and out-of-sample forecasting accuracy of the generalized additive model for location, scale and shape (GAMLSS) model in…

Abstract

Purpose

This paper aims to explore the in-sample explanatory and out-of-sample forecasting accuracy of the generalized additive model for location, scale and shape (GAMLSS) model in contrast to the GAM method in Munich’s residential market.

Design/methodology/approach

The paper explores the in-sample explanatory results via comparison of coefficients and a graphical analysis of non-linear effects. The out-of-sample forecasting accuracy focusses on 50 loops of three models excluding 10 per cent of the observations randomly. Afterwards, it obtains the predicted functional forms and predicts the remaining 10 per cent. The forecasting performance is measured via error variance, root mean squared error, mean absolute error and the mean percentage error.

Findings

The results show that the complexity of asking rents in Munich is more accurately captured by the GAMLSS approach than the GAM as shown by an outperformance in the in-sample explanatory accuracy. The results further show that the theoretical and empirical complexities do pay off in view of the increased out-of-sample forecasting power of the GAMLSS approach.

Research limitations/implications

The computational requirements necessary to estimate GAMLSS models in terms of number of cores and RAM are high and might constitute one of the limiting factors for (institutional) researchers. Moreover, large and detailed knowledge on statistical inference and programming is necessary.

Practical implications

The usage of the GAMLSS approach would lead policymakers to better understand the local factors affecting rents. Institutional researchers, instead, would clearly aim at calibrating the forecasting accuracy of the model to better forecast rents in investment strategies. Finally, future researchers are encouraged to exploit the large potential of the GAMLSS framework and its modelling flexibility.

Originality/value

The GAMLSS approach is widely recognised and used by international institutions such as the World Health Organisation, the International Monetary Fund and the European Commission. This is the first study to the best of the author’s knowledge to assess the properties of the GAMLSS approach in applied real estate research from a statistical asymptotic perspective by using a unique data basis with more than 38,000 observations.

Details

Journal of European Real Estate Research, vol. 11 no. 2
Type: Research Article
ISSN: 1753-9269

Keywords

Article
Publication date: 13 March 2024

Carla Ramos, Adriana Bruscato Bortoluzzo and Danny P. Claro

This study aims to capture how the association between a multichannel relational communication strategy (MRCS) and customer performance is contingent upon such customer…

Abstract

Purpose

This study aims to capture how the association between a multichannel relational communication strategy (MRCS) and customer performance is contingent upon such customer performance (low- versus high-performance customers) and to reconcile past contradictory results in this marketing-related topic. To this end, the authors propose and validate the method of quantile regression as an unconventional, yet effective, means to proceed to that reconciliation.

Design/methodology/approach

This study collected data from 4,934 customers of a private pension fund firm and accounted for both firm- and customer-initiated relational communication channels (RCCs) and for customer lifetime value (CLV). This study estimated a generalized linear model and then a quantile regression model was used to account for customer performance heterogeneity.

Findings

This study finds that specific RCCs present different levels of association with performance for low- versus high-performance customers, where outcome customer performance is the dependent variable. For example, the relation between firm-initiated communication (FIC) and performance is stronger for low-CLV customers, whereas the relation between customer-initiated communication (CIC) and performance is increasingly stronger for high-CLV customers but not for low-CLV ones. This study also finds that combining different forms of FIC can result in a negative association with customer performance, especially for low-CLV customers.

Research limitations/implications

The authors tested the conceptual model in one single firm in the specific context of financial services and with cross-sectional data, so there should be caution when extrapolating this study’s findings.

Practical implications

This study offers nuanced and precise managerial insights on recommended resource allocation along with relational communication efforts, showing how managers can benefit from adopting a differentiated-customer performance approach when designing their MRCS.

Originality/value

This study provides an overview of the state of the art of MRCS, proposes a contingency analysis of the relationship between MRCS and performance based on customer performance heterogeneity and suggests the quantile method to perform such analysis and help reconcile past contradictory findings. This study shows how the association between RCCs and CLV varies across the conditional quantiles of the distribution of customer performance. This study also addresses a recent call for a more holistic perspective on the relationships between independent and dependent variables.

Article
Publication date: 8 April 2021

Lamia Jamel, Hanadi Eid Albogami, Mazen Abduljalil Abdulaal and Nuha Ahmed Aljohani

The purpose of this paper is to examine the impact of agency conflicts between managers and shareholders on corporate risk management and financial performance of Saudi firms…

Abstract

Purpose

The purpose of this paper is to examine the impact of agency conflicts between managers and shareholders on corporate risk management and financial performance of Saudi firms listed in the Saudi Stock Exchange Tadawul.

Design/methodology/approach

To investigate the effect of agency conflicts between managers and shareholders on corporate risk management and financial performance, we use a sample of 180 Saudi firms listed in the Saudi Stock Exchange Tadawul during the period from 2009 to 2018. Econometrically, we employ Vector Autoregressive (VAR) and General Linear Model (GLM) techniques as an appropriate methodology.

Findings

Our findings show that the risk level of the last year increase the corporate risk management and the performance of Saudi firm. We remark that the separation amongst control and ownership generates agency conflicts amongst managers and shareholders which can affect their behavior in decision-making and performance of the Saudi firms. Thus, the conflicts of interest arise from the differences among the work horizon, the risk assumed, the performance of enterprises, and the level of remuneration desired by the managers and shareholders in the case of Saudi firms.

Originality/value

The main contributions of our paper prove that the deepen the study of agency costs linked to a shareholding structure through the analysis of monitoring, obligation, and opportunity costs in the Saudi firms.

Details

Journal of Investment Compliance, vol. 22 no. 1
Type: Research Article
ISSN: 1528-5812

Keywords

1 – 10 of 321