Search results
1 – 10 of 206Christine Amsler, Robert James, Artem Prokhorov and Peter Schmidt
The traditional predictor of technical inefficiency proposed by Jondrow, Lovell, Materov, and Schmidt (1982) is a conditional expectation. This chapter explores whether, and by…
Abstract
The traditional predictor of technical inefficiency proposed by Jondrow, Lovell, Materov, and Schmidt (1982) is a conditional expectation. This chapter explores whether, and by how much, the predictor can be improved by using auxiliary information in the conditioning set. It considers two types of stochastic frontier models. The first type is a panel data model where composed errors from past and future time periods contain information about contemporaneous technical inefficiency. The second type is when the stochastic frontier model is augmented by input ratio equations in which allocative inefficiency is correlated with technical inefficiency. Compared to the standard kernel-smoothing estimator, a newer estimator based on a local linear random forest helps mitigate the curse of dimensionality when the conditioning set is large. Besides numerous simulations, there is an illustrative empirical example.
Details
Keywords
The main goal of this paper is to investigate whether there is long-memory behavior in the CBOE Brazil ETF volatility index (named here VIXBR). As structural breaks may create a…
Abstract
Purpose
The main goal of this paper is to investigate whether there is long-memory behavior in the CBOE Brazil ETF volatility index (named here VIXBR). As structural breaks may create a spurious long-range dependence, the presence of structural breaks is also gauged.
Design/methodology/approach
The study considers the period from October 2011 to March 2021, using daily data. To test the long-memory behavior, three empirical approaches are adopted: GPH, ELW and robust GPH (RGPH) estimator. To estimate the structural break points adopted to date the subsamples, the ICSS algorithm is used.
Findings
Results considering the total period (TP) and subsamples show that the breaks did not create a spurious long-memory behavior and together with the rolling estimation, reveal strong evidence of the long-range dependence in the CBOE Brazil ETF volatility index. The higher degree of persistent of the VIXBR series suggests an extended period of increased uncertainty that agents need consider when making their investment decision.
Research limitations/implications
As possible extension of this study is to investigate the behavior of long memory and structural breaks for different frequencies (weekly, monthly, among others).
Practical implications
The presence of long-range dependence in the CBOE Brazil ETF volatility index reveals that the past information is important for the predictability of risks, and therefore, can help to protect against market risks, which has important implications regarding the future decisions of economic agents (for example, policy makers and investors).
Originality/value
Brazil is an emerging capital market (ECM) that has attracted a great deal of attention from investors and investment funds seeking to diversify its assets. This paper contributes to the empirical financial literature, by studying the long-memory behavior of the CBOE Brazil ETF volatility index, considering possible structural breaks. To the best of knowledge, this has not been done so far.
Details
Keywords
The author develops a bilateral Nash bargaining model under value uncertainty and private/asymmetric information, combining ideas from axiomatic and strategic bargaining theory…
Abstract
The author develops a bilateral Nash bargaining model under value uncertainty and private/asymmetric information, combining ideas from axiomatic and strategic bargaining theory. The solution to the model leads organically to a two-tier stochastic frontier (2TSF) setup with intra-error dependence. The author presents two different statistical specifications to estimate the model, one that accounts for regressor endogeneity using copulas, the other able to identify separately the bargaining power from the private information effects at the individual level. An empirical application using a matched employer–employee data set (MEEDS) from Zambia and a second using another one from Ghana showcase the applied potential of the approach.
Details
Keywords
This paper investigates income convergence using different convergence concepts and methodologies for 72 countries over the period between 1960 and 2010.
Abstract
Purpose
This paper investigates income convergence using different convergence concepts and methodologies for 72 countries over the period between 1960 and 2010.
Design/methodology/approach
This study applies beta (β), sigma (s), stochastic and club convergence approaches. For β-convergence analysis, it derives the cross-country growth regressions of the Solow growth model under the basic and augmented Cobb–Douglass (CD) production functions and estimates them using cross-section and panel data estimators. While it employs both the widely used coefficient of variation and recently developed weak s-convergence approaches for s-convergence, it applies three different unit root tests for stochastic convergence. To test club convergence, it estimates the log-t regression.
Findings
The results reveal that (1) there exists conditional β-convergence, meaning that poorer countries grow faster than richer countries; (2) income per worker is not (weakly) s-converging, and cross-sectional variation does not tend to fall over the years; (3) stochastic convergence is not found and (4) countries in the sample do not converge to the unique equilibrium, and there exist five distinctive convergence clubs.
Research limitations/implications
The results clearly show that heavily relying on one of the convergence techniques might lead researchers to obtain misleading results regarding the existence of convergence. Therefore, to draw reliable inferences, the results should be checked using different convergence concepts and methodologies.
Originality/value
Contrary to the previous literature, which is generally restricted to testing the existence of absolute and conditional β-convergence between countries, to the best of the author’s knowledge, this is the first study to consider and compare all originally and recently developed fundamental concepts of convergence altogether. Besides, it uses the Penn World Table (PWT) 9.1 and extends the period to 2010. From this point of view, this study is believed to provide the most up-to-date empirical evidence.
Details
Keywords
Classification of remote sensing images (RSI) is a challenging task in computer vision. Recently, researchers have proposed a variety of creative methods for automatic recognition…
Abstract
Purpose
Classification of remote sensing images (RSI) is a challenging task in computer vision. Recently, researchers have proposed a variety of creative methods for automatic recognition of RSI, and feature fusion is a research hotspot for its great potential to boost performance. However, RSI has a unique imaging condition and cluttered scenes with complicated backgrounds. This larger difference from nature images has made the previous feature fusion methods present insignificant performance improvements.
Design/methodology/approach
This work proposed a two-convolutional neural network (CNN) fusion method named main and branch CNN fusion network (MBC-Net) as an improved solution for classifying RSI. In detail, the MBC-Net employs an EfficientNet-B3 as its main CNN stream and an EfficientNet-B0 as a branch, named MC-B3 and BC-B0, respectively. In particular, MBC-Net includes a long-range derivation (LRD) module, which is specially designed to learn the dependence of different features. Meanwhile, MBC-Net also uses some unique ideas to tackle the problems coming from the two-CNN fusion and the inherent nature of RSI.
Findings
Extensive experiments on three RSI sets prove that MBC-Net outperforms the other 38 state-of-the-art (STOA) methods published from 2020 to 2023, with a noticeable increase in overall accuracy (OA) values. MBC-Net not only presents a 0.7% increased OA value on the most confusing NWPU set but also has 62% fewer parameters compared to the leading approach that ranks first in the literature.
Originality/value
MBC-Net is a more effective and efficient feature fusion approach compared to other STOA methods in the literature. Given the visualizations of grad class activation mapping (Grad-CAM), it reveals that MBC-Net can learn the long-range dependence of features that a single CNN cannot. Based on the tendency stochastic neighbor embedding (t-SNE) results, it demonstrates that the feature representation of MBC-Net is more effective than other methods. In addition, the ablation tests indicate that MBC-Net is effective and efficient for fusing features from two CNNs.
Details
Keywords
Taining Wang and Daniel J. Henderson
A semiparametric stochastic frontier model is proposed for panel data, incorporating several flexible features. First, a constant elasticity of substitution (CES) production…
Abstract
A semiparametric stochastic frontier model is proposed for panel data, incorporating several flexible features. First, a constant elasticity of substitution (CES) production frontier is considered without log-transformation to prevent induced non-negligible estimation bias. Second, the model flexibility is improved via semiparameterization, where the technology is an unknown function of a set of environment variables. The technology function accounts for latent heterogeneity across individual units, which can be freely correlated with inputs, environment variables, and/or inefficiency determinants. Furthermore, the technology function incorporates a single-index structure to circumvent the curse of dimensionality. Third, distributional assumptions are eschewed on both stochastic noise and inefficiency for model identification. Instead, only the conditional mean of the inefficiency is assumed, which depends on related determinants with a wide range of choice, via a positive parametric function. As a result, technical efficiency is constructed without relying on an assumed distribution on composite error. The model provides flexible structures on both the production frontier and inefficiency, thereby alleviating the risk of model misspecification in production and efficiency analysis. The estimator involves a series based nonlinear least squares estimation for the unknown parameters and a kernel based local estimation for the technology function. Promising finite-sample performance is demonstrated through simulations, and the model is applied to investigate productive efficiency among OECD countries from 1970–2019.
Details
Keywords
Hrishikesh Vinod, Kurt Jetta and Minaya Eric Rengifo
This study aims to highlight potential savings in advertising budgets.
Abstract
Purpose
This study aims to highlight potential savings in advertising budgets.
Design/methodology/approach
This study uses modern computer-based tools including stochastic dominance to check if advertising expenses are increasing sales by using modern causality assessment tools which allow for nonlinearities and use sophisticated assessment of causal impact of ads on sales.
Findings
This study identifies specific media spots where ad budget savings are possible. The marketing managers can take the next step to make small-scale local experiments to reassess this study’s findings.
Research limitations/implications
This study is a statistical observational assessment not based on controlled experiments.
Practical implications
The authors have tools to identify ineffective advertising which can produce huge savings for the organization. The over-the-counter cold remedies have become important due to the pandemic. The tools have wider applicability in marketing research.
Social implications
Less wasteful expenses always benefit the society.
Originality/value
To the best of the authors’ knowledge, this may be the first such attempt to use sophisticated causal identification tools. Remedies for the common cold sold by seven major US retailers help identify specific retailers and specific media with negative returns on investment.
Details
Keywords
Saroj Kumar Pani and Madhusmita Tripathy
This paper explains why some firms manage to capture disproportionate value from their network of relationships, leading to superior performance. The paper examines how a firm's…
Abstract
Purpose
This paper explains why some firms manage to capture disproportionate value from their network of relationships, leading to superior performance. The paper examines how a firm's dependencies affect its value appropriation potential (VAP) in economic networks.
Design/methodology/approach
The paper follows the axiomatic method and the embeddedness perspective of firms to develop an index called nodal power, which captures the power that accrues to a firm in exchange-based economic networks. Thereafter, using the formal method and simulation, it shows nodal power reflects a firm's VAP in economic networks.
Findings
The study analysis and findings prove that a firm's dyadic level exchange relations and the embedded network structure determine its VAP by affecting the nodal power. A firm with lesser nodal power is likely to appropriate less value from its relations even if it equally contributes to the value creation. This finding explains how the structural and relational characteristics of a firm's network enable disproportionate value appropriation.
Practical implications
Nodal power furthers the scope of analyzing firms' economic relationships and changing power equations in dynamic networks. It can help firms build optimal strategic networks and manage the portfolio of relationships by predicting the impact of changing relations on firms' VAP.
Originality/value
The paper's original contribution is to explain, through formal analysis, why and how the structure and nature of relations of firms affect their VAP. The paper also formalizes the power-dependence principle through a dependency-based index called nodal power and uses it to show how interfirm dependencies are key to value appropriation.
Details
Keywords
The purpose of this study is to reveal the dynamics of house prices and sales in spatial and temporal dimensions across British regions.
Abstract
Purpose
The purpose of this study is to reveal the dynamics of house prices and sales in spatial and temporal dimensions across British regions.
Design/methodology/approach
This paper incorporates two empirical approaches to describe the behaviour of property prices across British regions. The models are applied to two different data sets. The first empirical approach is to apply the price diffusion model proposed by Holly et al. (2011) to the UK house price index data set. The second empirical approach is to apply a bivariate global vector autoregression model without a time trend to house prices and transaction volumes retrieved from the nationwide building society.
Findings
Identifying shocks to London house prices in the GVAR model, based on the generalized impulse response functions framework, I find some heterogeneity in responses to house price changes; for example, South East England responds stronger than the remaining provincial regions. The main pattern detected in responses and characteristic for each region is the fairly rapid fading of the shock. The spatial-temporal diffusion model demonstrates the presence of a ripple effect: a shock emanating from London is dispersed contemporaneously and spatially to other regions, affecting prices in nondominant regions with a delay.
Originality/value
The main contribution of this work is the betterment in understanding how house price changes move across regions and time within a UK context.
Details