Search results

1 – 10 of over 8000
Book part
Publication date: 23 October 2023

Nathaniel T. Wilcox

The author presents new estimates of the probability weighting functions found in rank-dependent theories of choice under risk. These estimates are unusual in two senses. First…

Abstract

The author presents new estimates of the probability weighting functions found in rank-dependent theories of choice under risk. These estimates are unusual in two senses. First, they are free of functional form assumptions about both utility and weighting functions, and they are entirely based on binary discrete choices and not on matching or valuation tasks, though they depend on assumptions concerning the nature of probabilistic choice under risk. Second, estimated weighting functions contradict widely held priors of an inverse-s shape with fixed point well in the interior of the (0,1) interval: Instead the author usually finds populations dominated by “optimists” who uniformly overweight best outcomes in risky options. The choice pairs used here mostly do not provoke similarity-based simplifications. In a third experiment, the author shows that the presence of choice pairs that provoke similarity-based computational shortcuts does indeed flatten estimated probability weighting functions.

Details

Models of Risk Preferences: Descriptive and Normative Challenges
Type: Book
ISBN: 978-1-83797-269-2

Keywords

Book part
Publication date: 23 October 2023

Brian Albert Monroe

Risk preferences play a critical role in almost every facet of economic activity. Experimental economists have sought to infer the risk preferences of subjects from choice…

Abstract

Risk preferences play a critical role in almost every facet of economic activity. Experimental economists have sought to infer the risk preferences of subjects from choice behavior over lotteries. To help mitigate the influence of observable, and unobservable, heterogeneity in their samples, risk preferences have been estimated at the level of the individual subject. Recent work has detailed the lack of statistical power in descriptively classifying individual subjects as conforming to Expected Utility Theory (EUT) or Rank Dependent Utility (RDU). I discuss the normative consequences of this lack of power and provide some suggestions to improve the accuracy of normative inferences about individual-level choice behavior.

Details

Models of Risk Preferences: Descriptive and Normative Challenges
Type: Book
ISBN: 978-1-83797-269-2

Keywords

Book part
Publication date: 5 April 2024

Zhichao Wang and Valentin Zelenyuk

Estimation of (in)efficiency became a popular practice that witnessed applications in virtually any sector of the economy over the last few decades. Many different models were…

Abstract

Estimation of (in)efficiency became a popular practice that witnessed applications in virtually any sector of the economy over the last few decades. Many different models were deployed for such endeavors, with Stochastic Frontier Analysis (SFA) models dominating the econometric literature. Among the most popular variants of SFA are Aigner, Lovell, and Schmidt (1977), which launched the literature, and Kumbhakar, Ghosh, and McGuckin (1991), which pioneered the branch taking account of the (in)efficiency term via the so-called environmental variables or determinants of inefficiency. Focusing on these two prominent approaches in SFA, the goal of this chapter is to try to understand the production inefficiency of public hospitals in Queensland. While doing so, a recognized yet often overlooked phenomenon emerges where possible dramatic differences (and consequently very different policy implications) can be derived from different models, even within one paradigm of SFA models. This emphasizes the importance of exploring many alternative models, and scrutinizing their assumptions, before drawing policy implications, especially when such implications may substantially affect people’s lives, as is the case in the hospital sector.

Book part
Publication date: 23 October 2023

Glenn W. Harrison and J. Todd Swarthout

We take Cumulative Prospect Theory (CPT) seriously by rigorously estimating structural models using the full set of CPT parameters. Much of the literature only estimates a subset…

Abstract

We take Cumulative Prospect Theory (CPT) seriously by rigorously estimating structural models using the full set of CPT parameters. Much of the literature only estimates a subset of CPT parameters, or more simply assumes CPT parameter values from prior studies. Our data are from laboratory experiments with undergraduate students and MBA students facing substantial real incentives and losses. We also estimate structural models from Expected Utility Theory (EUT), Dual Theory (DT), Rank-Dependent Utility (RDU), and Disappointment Aversion (DA) for comparison. Our major finding is that a majority of individuals in our sample locally asset integrate. That is, they see a loss frame for what it is, a frame, and behave as if they evaluate the net payment rather than the gross loss when one is presented to them. This finding is devastating to the direct application of CPT to these data for those subjects. Support for CPT is greater when losses are covered out of an earned endowment rather than house money, but RDU is still the best single characterization of individual and pooled choices. Defenders of the CPT model claim, correctly, that the CPT model exists “because the data says it should.” In other words, the CPT model was borne from a wide range of stylized facts culled from parts of the cognitive psychology literature. If one is to take the CPT model seriously and rigorously then it needs to do a much better job of explaining the data than we see here.

Details

Models of Risk Preferences: Descriptive and Normative Challenges
Type: Book
ISBN: 978-1-83797-269-2

Keywords

Book part
Publication date: 5 April 2024

Taining Wang and Daniel J. Henderson

A semiparametric stochastic frontier model is proposed for panel data, incorporating several flexible features. First, a constant elasticity of substitution (CES) production…

Abstract

A semiparametric stochastic frontier model is proposed for panel data, incorporating several flexible features. First, a constant elasticity of substitution (CES) production frontier is considered without log-transformation to prevent induced non-negligible estimation bias. Second, the model flexibility is improved via semiparameterization, where the technology is an unknown function of a set of environment variables. The technology function accounts for latent heterogeneity across individual units, which can be freely correlated with inputs, environment variables, and/or inefficiency determinants. Furthermore, the technology function incorporates a single-index structure to circumvent the curse of dimensionality. Third, distributional assumptions are eschewed on both stochastic noise and inefficiency for model identification. Instead, only the conditional mean of the inefficiency is assumed, which depends on related determinants with a wide range of choice, via a positive parametric function. As a result, technical efficiency is constructed without relying on an assumed distribution on composite error. The model provides flexible structures on both the production frontier and inefficiency, thereby alleviating the risk of model misspecification in production and efficiency analysis. The estimator involves a series based nonlinear least squares estimation for the unknown parameters and a kernel based local estimation for the technology function. Promising finite-sample performance is demonstrated through simulations, and the model is applied to investigate productive efficiency among OECD countries from 1970–2019.

Article
Publication date: 24 August 2023

Banumathy Sundararaman and Neelakandan Ramalingam

This study was carried out to analyze the importance of consumer preference data in forecasting demand in apparel retailing.

Abstract

Purpose

This study was carried out to analyze the importance of consumer preference data in forecasting demand in apparel retailing.

Methodology

To collect preference data, 729 hypothetical stock keeping units (SKU) were derived using a full factorial design, from a combination of six attributes and three levels each. From the hypothetical SKU's, 63 practical SKU's were selected for further analysis. Two hundred two responses were collected from a store intercept survey. Respondents' utility scores for all 63 SKUs were calculated using conjoint analysis. In estimating aggregate demand, to allow for consumer substitution and to make the SKU available when a consumer wishes to buy more than one item in the same SKU, top three highly preferred SKU's utility scores of each individual were selected and classified using a decision tree and was aggregated. A choice rule was modeled to include substitution; by applying this choice rule, aggregate demand was estimated.

Findings

The respondents' utility scores were calculated. The value of Kendall's tau is 0.88, the value of Pearson's R is 0.98 and internal predictive validity using Kendall's tau is 1.00, and this shows the high quality of data obtained. The proposed model was used to estimate the demand for 63 SKUs. The demand was estimated at 6.04 per cent for the SKU cotton, regular style, half sleeve, medium priced, private label. The proposed model for estimating demand using consumer preference data gave better estimates close to actual sales than expert opinion data. The Spearman's rank correlation between actual sales and consumer preference data is 0.338 and is significant at 5 per cent level. The Spearman's rank correlation between actual sales and expert opinion is −0.059, and there is no significant relation between expert opinion data and actual sales. Thus, consumer preference model proves to be better in estimating demand than expert opinion data.

Research implications

There has been a considerable amount of work done in choice-based models. There is a lot of scope in working in deterministic models.

Practical implication

The proposed consumer preference-based demand estimation model can be beneficial to the apparel retailers in increasing their profit by reducing stock-out and overstocking situations. Though conjoint analysis is used in demand estimation in other industries, it is not used in apparel for demand estimations and can be greater use in its simplest form.

Originality/value

This research is the first one to model consumer preferences-based data to estimate demand in apparel. This research was practically tested in an apparel retail store. It is original.

Details

Journal of Fashion Marketing and Management: An International Journal, vol. 28 no. 2
Type: Research Article
ISSN: 1361-2026

Keywords

Article
Publication date: 26 December 2023

Savannah (Yuanyuan) Guo, Beilei Mei, Yanchao Rao and Jianfang Ye

This study investigates the implementation challenges and economic consequences of the International Financial Reporting Standards 9 (IFRS 9) Financial Instruments.

Abstract

Purpose

This study investigates the implementation challenges and economic consequences of the International Financial Reporting Standards 9 (IFRS 9) Financial Instruments.

Design/methodology/approach

Descriptive evidence on equity asset reclassifications and estimated impairment using the new expected credit loss (ECL) model are presented. Multivariate analyses on the disposal of available-for-sale (AFS) and fund investment post-announcement and the value relevance of impairments to financial assets post-implementation are performed.

Findings

Over 60% of sample firms report inconsistent equity asset reclassifications and do not change estimated impairment using the new expected credit loss model. Firms also switch from AFS to equity fund investments post-announcement. Lastly, impairments to financial assets increase in value relevance to investors’ post-implementation, but only in financial institutions and firms with Big 4 auditors.

Originality/value

This study's findings suggest that IFRS 9 presents implementation challenges and changes equity investment strategies. They also indicate cross-sectional differences in firms' ability to effectively apply the new standards. This study is valuable for policymakers, business leaders, investors and academics.

Details

Journal of Accounting Literature, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 0737-4607

Keywords

Article
Publication date: 27 December 2022

Bright Awuku, Eric Asa, Edmund Baffoe-Twum and Adikie Essegbey

Challenges associated with ensuring the accuracy and reliability of cost estimation of highway construction bid items are of significant interest to state highway transportation…

Abstract

Purpose

Challenges associated with ensuring the accuracy and reliability of cost estimation of highway construction bid items are of significant interest to state highway transportation agencies. Even with the existing research undertaken on the subject, the problem of inaccurate estimation of highway bid items still exists. This paper aims to assess the accuracy of the cost estimation methods employed in the selected studies to provide insights into how well they perform empirically. Additionally, this research seeks to identify, synthesize and assess the impact of the factors affecting highway unit prices because they affect the total cost of highway construction costs.

Design/methodology/approach

This paper systematically searched, selected and reviewed 105 papers from Scopus, Google Scholar, American Society of Civil Engineers (ASCE), Transportation Research Board (TRB) and Science Direct (SD) on conceptual cost estimation of highway bid items. This study used content and nonparametric statistical analyses to determine research trends, identify, categorize the factors influencing highway unit prices and assess the combined performance of conceptual cost prediction models.

Findings

Findings from the trend analysis showed that between 1983 and 2019 North America, Asia, Europe and the Middle East contributed the most to improving highway cost estimation research. Aggregating the quantitative results and weighting the findings using each study's sample size revealed that the average error between the actual and the estimated project costs of Monte-Carlo simulation models (5.49%) performed better compared to the Bayesian model (5.95%), support vector machines (6.03%), case-based reasoning (11.69%), artificial neural networks (12.62%) and regression models (13.96%). This paper identified 41 factors and was grouped into three categories, namely: (1) factors relating to project characteristics; (2) organizational factors and (3) estimate factors based on the common classification used in the selected papers. The mean ranking analysis showed that most of the selected papers used project-specific factors more when estimating highway construction bid items than the other factors.

Originality/value

This paper contributes to the body of knowledge by analyzing and comparing the performance of highway cost estimation models, identifying and categorizing a comprehensive list of cost drivers to stimulate future studies in improving highway construction cost estimates.

Details

Engineering, Construction and Architectural Management, vol. 31 no. 3
Type: Research Article
ISSN: 0969-9988

Keywords

Article
Publication date: 25 September 2023

Anchal Patil, Vipulesh Shardeo, Jitender Madaan, Ashish Dwivedi and Sanjoy Kumar Paul

This study aims to evaluate the dynamics between healthcare resource capacity expansion and disease spread. Further, the study estimates the resources required to respond to a…

Abstract

Purpose

This study aims to evaluate the dynamics between healthcare resource capacity expansion and disease spread. Further, the study estimates the resources required to respond to a pandemic appropriately.

Design/methodology/approach

This study adopts a system dynamics simulation and scenario analysis to experiment with the modification of the susceptible exposed infected and recovered (SEIR) model. The experiments evaluate diagnostic capacity expansion to identify suitable expansion plans and timelines. Afterwards, two popularly used forecasting tools, artificial neural network (ANN) and auto-regressive integrated moving average (ARIMA), are used to estimate the requirement of beds for a period when infection data became available.

Findings

The results from the study reflect that aggressive testing with isolation and integration of quarantine can be effective strategies to prevent disease outbreaks. The findings demonstrate that decision-makers must rapidly expand the diagnostic capacity during the first two weeks of the outbreak to support aggressive testing and isolation. Further, results confirm a healthcare resource deficit of at least two months for Delhi in the absence of these strategies. Also, the study findings highlight the importance of capacity expansion timelines by simulating a range of contact rates and disease infectivity in the early phase of the outbreak when various parameters are unknown. Further, it has been reflected that forecasting tools can effectively estimate healthcare resource requirements when pandemic data is available.

Practical implications

The models developed in the present study can be utilised by policymakers to suitably design the response plan. The decisions regarding how much diagnostics capacity is needed and when to expand capacity to minimise infection spread have been demonstrated for Delhi city. Also, the study proposed a decision support system (DSS) to assist the decision-maker in short- and long-term planning during the disease outbreak.

Originality/value

The study estimated the resources required for adopting an aggressive testing strategy. Several experiments were performed to successfully validate the robustness of the simulation model. The modification of SEIR model with diagnostic capacity increment, quarantine and testing block has been attempted to provide a distinct perspective on the testing strategy. The prevention of outbreaks has been addressed systematically.

Details

International Journal of Physical Distribution & Logistics Management, vol. 53 no. 10
Type: Research Article
ISSN: 0960-0035

Keywords

Article
Publication date: 26 June 2023

Corrado Andini and José Eusébio Santos

The aim is to study the impact of schooling on between-groups wage inequality beyond the lens of the standard approach in the literature.

Abstract

Purpose

The aim is to study the impact of schooling on between-groups wage inequality beyond the lens of the standard approach in the literature.

Design/methodology/approach

Simple econometric theory is used to make the main point of the paper. Supporting empirical evidence is also presented.

Findings

Disregarding the persistence of current earnings implies a bias in the estimation of the wage return to schooling both at labour-market entry and in the rest of the working life.

Research limitations/implications

The use of current earnings as a dependent variable in wage-schooling models may be problematic and requires specific handling.

Social implications

The impact of schooling on the between-groups dimension of wage inequality may be different than previously thought.

Originality/value

The paper is the first to show that, when current earnings are used as a dependent variable, the identification of a wage-schooling model with the standard (time-invariant external instrument-variable) approach may lead to misleading conclusions.

Details

Journal of Economic Studies, vol. 51 no. 2
Type: Research Article
ISSN: 0144-3585

Keywords

1 – 10 of over 8000