Search results
1 – 10 of over 2000Yao 'Henry' Jin, Brent D. Williams, Matthew A. Waller and Adriana Rossiter Hofer
The accurate measurement of demand variability amplification across different nodes in the supply chain, or “bullwhip effect,” is critical for firms to achieve more efficient…
Abstract
Purpose
The accurate measurement of demand variability amplification across different nodes in the supply chain, or “bullwhip effect,” is critical for firms to achieve more efficient inventory, production, and ordering planning processes. Building on recent analytical research that suggests that data aggregation tends to mask the bullwhip effect in the retail industry, the purpose of this paper is to empirically investigate whether different patterns of data aggregation influence its measurement.
Design/methodology/approach
Utilizing weekly, product-level order and sales data from three product categories of a consumer packaged goods manufacturer, the study uses hierarchical linear modeling to empirically test the effects of data aggregation on different measures of bullwhip.
Findings
The authors findings lend strong support to the masking effect of aggregating sales and order data along product-location and temporal dimensions, as well as the dampening effect of seasonality on the measurement of the bullwhip effect.
Research limitations/implications
These findings indicate that inconsistencies found in the literature may be due to measurement aggregation and statistical techniques, both of which should be applied with care by academics and practitioners in order to preserve the fidelity of their analyses.
Originality/value
Using product-weekly level data that cover both seasonal and non-seasonal demand, this study is the first, to the author’s knowledge, to systematically aggregate data up to category and monthly levels to empirically examine the impact of data aggregation and seasonality on bullwhip measurement.
Details
Keywords
Anthony Owusu-Ansah, William Mark Adolwine and Eric Yeboah
The purpose of this paper is to test whether temporal aggregation matters when constructing hedonic house price indices for developing markets using Ghana as a case study.
Abstract
Purpose
The purpose of this paper is to test whether temporal aggregation matters when constructing hedonic house price indices for developing markets using Ghana as a case study.
Design/methodology/approach
Monthly, quarterly, semi-yearly and yearly hedonic price indices are constructed and six null hypotheses are tested using the F-ratios to examine the temporal aggregation effect.
Findings
The results show that temporal aggregation may not be a serious issue when constructing hedonic house price indices for developing markets as a result of the smaller sample size which these markets normally have. At even 10 per cent significance level, none of the F-ratios estimated is statistically significant. Analysis of the mean returns and volatilities reveal that indices constructed at the lower level of temporal aggregation are very volatile, suggesting that the volume of transactions can affect the level of temporal aggregation, and so, the temporal aggregation level should not be generalised, as is currently observed in the literature.
Originality/value
The diversification importance of real estate and the introduction of real estate derivatives and home equity insurance as financial products call for the construction of robust and accurate real estate indices in all markets. While almost all empirical research recommends real estate price indices to be conducted at the lower level of temporal aggregation, these studies are largely conducted in developed markets where transactions take place frequently and large transaction databases exist. Unfortunately, little is known about the importance of temporal aggregation effect when constructing indices for developing real estate markets. This paper contributes to fill these gaps.
Details
Keywords
The purpose of this paper is to examine if temporal aggregation matters in the construction of house price indices and to test the accuracy of alternative index construction…
Abstract
Purpose
The purpose of this paper is to examine if temporal aggregation matters in the construction of house price indices and to test the accuracy of alternative index construction methods.
Design/methodology/approach
Five index construction models based on the hedonic, repeat‐sales and hybrid methods are examined. The accuracy of the alternative index construction methods are examined using the mean squared error and out‐of‐sample technique. Monthly, quarterly, semi‐yearly and yearly indices are constructed for each of the methods and six null hypotheses are tested to examine the temporal aggregation effect.
Findings
Overall, the hedonic is the best method to use. While running separate regressions to estimate the index is best at the broader level of time aggregation like the annual, pooling data together and including time dummies to estimate the index is the best at the lower level of time aggregation. The repeat‐sales method is the least preferred method. The results also show that it is important to limit time to the lowest level of temporal aggregation when construction property price indices.
Practical implications
This paper provides alternative method, the mean squared error method based on an out‐of‐sample technique to evaluate the accuracy of alternative index construction methods.
Originality/value
The introduction of financial products like the property derivatives and home equity insurances to the financial market calls for accurate and robust property price indices. However, the index method and level of temporal aggregation to use still remain unresolved in the index construction literature. This paper contributes to fill these gaps.
Details
Keywords
Thomas B. Götz, Alain Hecq and Jean-Pierre Urbain
This article proposes a new approach to detecting the presence of common cyclical features when several time series are sampled at different frequencies. We generalize the…
Abstract
This article proposes a new approach to detecting the presence of common cyclical features when several time series are sampled at different frequencies. We generalize the common-frequency approach introduced by Engle and Kozicki (1993) and Vahid and Engle (1993). We start with the mixed-frequency VAR representation investigated in Ghysels (2012) for stationary time series. For non-stationary time series in levels, we show that one has to account for the presence of two sets of long-run relationships. The first set is implied by identities stemming from the fact that the differences of the high-frequency
Details
Keywords
Dag Einar Sommervoll and Gavin Wood
This paper aims to study to what extent an insurance based on a house price index provides equity protection for homeowners.
Abstract
Purpose
This paper aims to study to what extent an insurance based on a house price index provides equity protection for homeowners.
Design/methodology/approach
The paper uses a novel dataset of all housing market transactions in the metropolitan area of Melbourne 1990‐2006, to construct repeated sales indices of various temporal spatial aggregation. These indices are used to discuss the efficiency of index‐based insurance schemes. The paper also considers efficiency under different specifications of legitimate claims.
Findings
It is found that the payout efficiency is surprisingly stable (around 50 percent) for all temporal spatial aggregations. A neighborhood index outperforms the metropolitan index with respect to target efficiency (the probability of payout given a loss). The introduction of maturity times, say legitimate claim five years after purchase, does improve efficiency somewhat. However, the idiosyncratic component of housing market transactions remains high, and the insurance probably unattractive from a homeowner perspective.
Originality/value
To the authors' knowledge, this is the first time an index‐based insurance scheme is analyzed using real‐market transactions.
Details
Keywords
In this paper, an emerging state-of-the-art machine intelligence technique called the Hierarchical Temporal Memory (HTM) is applied to the task of short-term load forecasting…
Abstract
In this paper, an emerging state-of-the-art machine intelligence technique called the Hierarchical Temporal Memory (HTM) is applied to the task of short-term load forecasting (STLF). A HTM Spatial Pooler (HTM-SP) stage is used to continually form sparse distributed representations (SDRs) from a univariate load time series data, a temporal aggregator is used to transform the SDRs into a sequential bivariate representation space and an overlap classifier makes temporal classifications from the bivariate SDRs through time. The comparative performance of HTM on several daily electrical load time series data including the Eunite competition dataset and the Polish power system dataset from 2002 to 2004 are presented. The robustness performance of HTM is also further validated using hourly load data from three more recent electricity markets. The results obtained from experimenting with the Eunite and Polish dataset indicated that HTM will perform better than the existing techniques reported in the literature. In general, the robustness test also shows that the error distribution performance of the proposed HTM technique is positively skewed for most of the years considered and with kurtosis values mostly lower than a base value of 3 indicating a reasonable level of outlier rejections.
Details
Keywords
Eric Ghysels and J. Isaac Miller
We analyze the sizes of standard cointegration tests applied to data subject to linear interpolation, discovering evidence of substantial size distortions induced by the…
Abstract
We analyze the sizes of standard cointegration tests applied to data subject to linear interpolation, discovering evidence of substantial size distortions induced by the interpolation. We propose modifications to these tests to effectively eliminate size distortions from such tests conducted on data interpolated from end-of-period sampled low-frequency series. Our results generally do not support linear interpolation when alternatives such as aggregation or mixed-frequency-modified tests are possible.
Details
Keywords
The research community currently employs four very different versions of the social network concept: A social network is seen as a set of socially constructed role relations…
Abstract
Purpose
The research community currently employs four very different versions of the social network concept: A social network is seen as a set of socially constructed role relations (e.g., friends, business partners), a set of interpersonal sentiments (e.g., liking, trust), a pattern of behavioral social interaction (e.g., conversations, citations), or an opportunity structure for exchange. Researchers conventionally assume these conceptualizations are interchangeable as social ties, and some employ composite measures that aim to capture more than one dimension. Even so, important discrepancies often appear for non-ties (as dyads where a specific role relation or sentiment is not reported, a specific form of interaction is not observed, or exchange is not possible).
Methodology/Approach
Investigating the interplay across the four definitions is a step toward developing scope conditions for generalization and application of theory across these domains.
Research Implications
This step is timely because emerging tools of computational social science – wearable sensors, logs of telecommunication, online exchange, or other interaction – now allow us to observe the fine-grained dynamics of interaction over time. Combined with cutting-edge methods for analysis, these lenses allow us to move beyond reified notions of social ties (and non-ties) and instead directly observe and analyze the dynamic and structural interdependencies of social interaction behavior.
Originality/Value of the Paper
This unprecedented opportunity invites us to refashion dynamic structural theories of exchange that advance “beyond networks” to unify previously disjoint research streams on relationships, interaction, and opportunity structures.
Details
Keywords
Orland Hoeber, Larena Hoeber, Maha El Meseery, Kenneth Odoh and Radhika Gopi
Due to the size and velocity at which user generated content is created on social media services such as Twitter, analysts are often limited by the need to pre-determine the…
Abstract
Purpose
Due to the size and velocity at which user generated content is created on social media services such as Twitter, analysts are often limited by the need to pre-determine the specific topics and themes they wish to follow. Visual analytics software may be used to support the interactive discovery of emergent themes. The paper aims to discuss these issues.
Design/methodology/approach
Tweets collected from the live Twitter stream matching a user’s query are stored in a database, and classified based on their sentiment. The temporally changing sentiment is visualized, along with sparklines showing the distribution of the top terms, hashtags, user mentions, and authors in each of the positive, neutral, and negative classes. Interactive tools are provided to support sub-querying and the examination of emergent themes.
Findings
A case study of using Vista to analyze sport fan engagement within a mega-sport event (2013 Le Tour de France) is provided. The authors illustrate how emergent themes can be identified and isolated from the large collection of data, without the need to identify these a priori.
Originality/value
Vista provides mechanisms that support the interactive exploration among Twitter data. By combining automatic data processing and machine learning methods with interactive visualization software, researchers are relieved of tedious data processing tasks, and can focus on the analysis of high-level features of the data. In particular, patterns of Twitter use can be identified, emergent themes can be isolated, and purposeful samples of the data can be selected by the researcher for further analysis.
Details