Search results
1 – 10 of over 3000Paritosh Pramanik, Rabin K. Jana and Indranil Ghosh
New business density (NBD) is the ratio of the number of newly registered liability corporations to the working-age population per year. NBD is critical to assessing a country's…
Abstract
Purpose
New business density (NBD) is the ratio of the number of newly registered liability corporations to the working-age population per year. NBD is critical to assessing a country's business environment. The present work endeavors to discover and gauge the contribution of 28 potential socio-economic enablers of NBD for 2006–2021 across developed and developing economies separately and to make a comparative assessment between those two regions.
Design/methodology/approach
Using World Bank data, the study first performs exploratory data analysis (EDA). Then, it deploys a deep learning (DL)-based regression framework by utilizing a deep neural network (DNN) to perform predictive modeling of NBD for developed and developing nations. Subsequently, we use two explainable artificial intelligence (XAI) techniques, Shapley values and a partial dependence plot, to unveil the influence patterns of chosen enablers. Finally, the results from the DL method are validated with the explainable boosting machine (EBM) method.
Findings
This research analyzes the role of 28 potential socio-economic enablers of NBD in developed and developing countries. This research finds that the NBD in developed countries is predominantly governed by the contribution of manufacturing and service sectors to GDP. In contrast, the propensity for research and development and ease of doing business control the NBD of developing nations. The research findings also indicate four common enablers – business disclosure, ease of doing business, employment in industry and startup procedures for developed and developing countries.
Practical implications
NBD is directly linked to any nation's economic affairs. Therefore, assessing the NBD enablers is of paramount significance for channelizing capital for new business formation. It will guide investment firms and entrepreneurs in discovering the factors that significantly impact the NBD dynamics across different regions of the globe. Entrepreneurs fraught with inevitable market uncertainties while developing a new idea into a successful new business can momentously benefit from the awareness of crucial NBD enablers, which can serve as a basis for business risk assessment.
Originality/value
DL-based regression framework simultaneously caters to successful predictive modeling and model explanation for practical insights about NBD at the global level. It overcomes the limitations in the present literature that assume the NBD is country- and industry-specific, and factors of the NBD cannot be generalized globally. With DL-based regression and XAI methods, we prove our research hypothesis that NBD can be effectively assessed and compared with the help of global macro-level indicators. This research justifies the robustness of the findings by using the socio-economic data from the renowned data repository of the World Bank and by implementing the DL modeling with validation through the EBM method.
Details
Keywords
Xiao Yao, Dongxiao Wu, Zhiyong Li and Haoxiang Xu
Since stock return and volatility matters to investors, this study proposes to incorporate the textual sentiment of annual reports in stock price crash risk prediction.
Abstract
Purpose
Since stock return and volatility matters to investors, this study proposes to incorporate the textual sentiment of annual reports in stock price crash risk prediction.
Design/methodology/approach
Specific sentences gathered from management discussions and their subsequent analyses are tokenized and transformed into numeric vectors using textual mining techniques, and then the Naïve Bayes method is applied to score the sentiment, which is used as an input variable for crash risk prediction. The results are compared between a collection of predictive models, including linear regression (LR) and machine learning techniques.
Findings
The experimental results find that those predictive models that incorporate textual sentiment significantly outperform the baseline models with only accounting and market variables included. These conclusions hold when crash risk is proxied by either the negative skewness of the return distribution or down-to-up volatility (DUVOL).
Research limitations/implications
It should be noted that the authors' study focuses on examining the predictive power of textual sentiment in crash risk prediction, while other dimensions of textual features such as readability and thematic contents are not considered. More analysis is needed to explore the predictive power of textual features from various dimensions, with the most recent sample data included in future studies.
Originality/value
The authors' study provides implications for the information value of textual data in financial analysis and risk management. It suggests that the soft information contained within annual reports may prove informative in crash risk prediction, and the incorporation of textual sentiment provides an incremental improvement in overall predictive performance.
Details
Keywords
V. Chowdary Boppana and Fahraz Ali
This paper presents an experimental investigation in establishing the relationship between FDM process parameters and tensile strength of polycarbonate (PC) samples using the…
Abstract
Purpose
This paper presents an experimental investigation in establishing the relationship between FDM process parameters and tensile strength of polycarbonate (PC) samples using the I-Optimal design.
Design/methodology/approach
I-optimal design methodology is used to plan the experiments by means of Minitab-17.1 software. Samples are manufactured using Stratsys FDM 400mc and tested as per ISO standards. Additionally, an artificial neural network model was developed and compared to the regression model in order to select an appropriate model for optimisation. Finally, the genetic algorithm (GA) solver is executed for improvement of tensile strength of FDM built PC components.
Findings
This study demonstrates that the selected process parameters (raster angle, raster to raster air gap, build orientation about Y axis and the number of contours) had significant effect on tensile strength with raster angle being the most influential factor. Increasing the build orientation about Y axis produced specimens with compact structures that resulted in improved fracture resistance.
Research limitations/implications
The fitted regression model has a p-value less than 0.05 which suggests that the model terms significantly represent the tensile strength of PC samples. Further, from the normal probability plot it was found that the residuals follow a straight line, thus the developed model provides adequate predictions. Furthermore, from the validation runs, a close agreement between the predicted and actual values was seen along the reference line which further supports satisfactory model predictions.
Practical implications
This study successfully investigated the effects of the selected process parameters - raster angle, raster to raster air gap, build orientation about Y axis and the number of contours - on tensile strength of PC samples utilising the I-optimal design and ANOVA. In addition, for prediction of the part strength, regression and ANN models were developed. The selected ANN model was optimised using the GA-solver for determination of optimal parameter settings.
Originality/value
The proposed ANN-GA approach is more appropriate to establish the non-linear relationship between the selected process parameters and tensile strength. Further, the proposed ANN-GA methodology can assist in manufacture of various industrial products with Nylon, polyethylene terephthalate glycol (PETG) and PET as new 3DP materials.
Details
Keywords
Christian Nnaemeka Egwim, Hafiz Alaka, Youlu Pan, Habeeb Balogun, Saheed Ajayi, Abdul Hye and Oluwapelumi Oluwaseun Egunjobi
The study aims to develop a multilayer high-effective ensemble of ensembles predictive model (stacking ensemble) using several hyperparameter optimized ensemble machine learning…
Abstract
Purpose
The study aims to develop a multilayer high-effective ensemble of ensembles predictive model (stacking ensemble) using several hyperparameter optimized ensemble machine learning (ML) methods (bagging and boosting ensembles) trained with high-volume data points retrieved from Internet of Things (IoT) emission sensors, time-corresponding meteorology and traffic data.
Design/methodology/approach
For a start, the study experimented big data hypothesis theory by developing sample ensemble predictive models on different data sample sizes and compared their results. Second, it developed a standalone model and several bagging and boosting ensemble models and compared their results. Finally, it used the best performing bagging and boosting predictive models as input estimators to develop a novel multilayer high-effective stacking ensemble predictive model.
Findings
Results proved data size to be one of the main determinants to ensemble ML predictive power. Second, it proved that, as compared to using a single algorithm, the cumulative result from ensemble ML algorithms is usually always better in terms of predicted accuracy. Finally, it proved stacking ensemble to be a better model for predicting PM2.5 concentration level than bagging and boosting ensemble models.
Research limitations/implications
A limitation of this study is the trade-off between performance of this novel model and the computational time required to train it. Whether this gap can be closed remains an open research question. As a result, future research should attempt to close this gap. Also, future studies can integrate this novel model to a personal air quality messaging system to inform public of pollution levels and improve public access to air quality forecast.
Practical implications
The outcome of this study will aid the public to proactively identify highly polluted areas thus potentially reducing pollution-associated/ triggered COVID-19 (and other lung diseases) deaths/ complications/ transmission by encouraging avoidance behavior and support informed decision to lock down by government bodies when integrated into an air pollution monitoring system
Originality/value
This study fills a gap in literature by providing a justification for selecting appropriate ensemble ML algorithms for PM2.5 concentration level predictive modeling. Second, it contributes to the big data hypothesis theory, which suggests that data size is one of the most important factors of ML predictive capability. Third, it supports the premise that when using ensemble ML algorithms, the cumulative output is usually always better in terms of predicted accuracy than using a single algorithm. Finally developing a novel multilayer high-performant hyperparameter optimized ensemble of ensembles predictive model that can accurately predict PM2.5 concentration levels with improved model interpretability and enhanced generalizability, as well as the provision of a novel databank of historic pollution data from IoT emission sensors that can be purchased for research, consultancy and policymaking.
Details
Keywords
Bikram Jit Singh, Rippin Sehgal, Ayon Chakraborty and Rakesh Kumar Phanden
The use of technology in 4th industrial revolution is at its peak. Industries are trying to reduce the consumption of resources by effectively utilizing information and technology…
Abstract
Purpose
The use of technology in 4th industrial revolution is at its peak. Industries are trying to reduce the consumption of resources by effectively utilizing information and technology to connect different functioning agents of the manufacturing industry. Without digitization “Industry 4.0” will be a virtual reality. The present survey-based study explores the factual status of digital manufacturing in the Northern India.
Design/methodology/approach
After an extensive literature review, a questionnaire was designed to gather different viewpoints of Indian industrial practitioners. The first half contains questions related to north Indian demographic factors which may affect digitalization of India. The latter half includes the queries concerned with various operational factors (or drivers) driving the digital revolution without ignoring Indian constraints.
Findings
The focus of this survey was to understand the current level of digital revolution under the ongoing push by the Indian government focused upon digital movement. The analysis included non-parametric testing of the various demographic and functional factors impacting the digital echoes, specifically in Northern India. Findings such as technological upgradations were independent of type of industry, the turnover or the location. About 10 key operational factors were thoughtfully grouped into three major categories—internal Research and Development (R&D), the capability of the supply chain and the capacity to adapt to the market. These factors were then examined to understand how they contribute to digital manufacturing, utilizing an appropriate ordinal logistic regression. The resulting predictive analysis provides seldom-seen insights and valuable suggestions for the most effective deployment of digitalization in Indian industries.
Research limitations/implications
The country-specific Industry 4.0 literature is quite limited. The survey mainly focuses on the National Capital Region. The number of demographic and functional factors can further be incorporated. Moreover, an addition of factors related to ecology, environment and society can make the study more insightful.
Practical implications
The present work provides valuable insights about the current status of digitization and expects to facilitate public or private policymakers to implement digital technologies in India with less efforts and the least resistance. It empowers India towards Industry 4.0 based tools and techniques and creates new socio-economic dimensions for the sustainable development.
Originality/value
The quantitative nature of the study and its statistical predictions (data-based) are novel. The clubbing of similar success factors to avoid inter-collinearity and complexity is seldom seen. The predictive analytics provided in this study is quite elusive as it provides directions with logic. It will help the Indian Government and industrial strategists to plan and perform their interventions accordingly.
Details
Keywords
Mohammadreza Tavakoli Baghdadabad
We propose a risk factor for idiosyncratic entropy and explore the relationship between this factor and expected stock returns.
Abstract
Purpose
We propose a risk factor for idiosyncratic entropy and explore the relationship between this factor and expected stock returns.
Design/methodology/approach
We estimate a cross-sectional model of expected entropy that uses several common risk factors to predict idiosyncratic entropy.
Findings
We find a negative relationship between expected idiosyncratic entropy and returns. Specifically, the Carhart alpha of a low expected entropy portfolio exceeds the alpha of a high expected entropy portfolio by −2.37% per month. We also find a negative and significant price of expected idiosyncratic entropy risk using the Fama-MacBeth cross-sectional regressions. Interestingly, expected entropy helps us explain the idiosyncratic volatility puzzle that stocks with high idiosyncratic volatility earn low expected returns.
Originality/value
We propose a risk factor of idiosyncratic entropy and explore the relationship between this factor and expected stock returns. Interestingly, expected entropy helps us explain the idiosyncratic volatility puzzle that stocks with high idiosyncratic volatility earn low expected returns.
Details
Keywords
Lucinda Brabbins, Nima Moghaddam and David Dawson
Background: Quality of life is a core concern for cancer patients, which can be negatively affected by illness-related death anxiety; yet understanding of how to appropriately…
Abstract
Background: Quality of life is a core concern for cancer patients, which can be negatively affected by illness-related death anxiety; yet understanding of how to appropriately target psycho-oncological interventions remains lacking. We aimed to explore experiential acceptance in cancer patients, and whether acceptance – as an alternative to avoidant coping – was related to and predictive of better quality of life and death anxiety outcomes.
Methods: We used a longitudinal, quantitative design with a follow-up after three months. Seventy-two participants completed a questionnaire-battery measuring illness appraisals, acceptance and non-acceptance coping-styles, quality of life, and death anxiety; 31 participants repeated the battery after three months.
Results: Acceptance was an independent explanatory and predictive variable for quality of life and death anxiety, in the direction of psychological health. Acceptance had greater explanatory power for outcomes than either cancer appraisals or avoidant response styles. Avoidant response styles were associated with greater death anxiety and poorer quality of life.
Conclusions: The findings support the role of an accepting response-style in favourable psychological outcomes, identifying a possible target for future psychological intervention. Response styles that might be encouraged in other therapies, such as active coping, planning, and positive reframing, were not associated with beneficial outcomes.
Details
Keywords
Ikhlaas Gurrib, Firuz Kamalov, Olga Starkova, Elgilani Eltahir Elshareif and Davide Contu
This paper aims to investigate the role of price-based information from major cryptocurrencies, foreign exchange, equity markets and key commodities in predicting the next-minute…
Abstract
Purpose
This paper aims to investigate the role of price-based information from major cryptocurrencies, foreign exchange, equity markets and key commodities in predicting the next-minute Bitcoin (BTC) price. This study answers the following research questions: What is the best sparse regression model to predict the next-minute price of BTC? What are the key drivers of the BTC price in high-frequency trading?
Design/methodology/approach
Least absolute shrinkage and selection operator and Ridge regressions are adopted using minute-based open-high-low-close prices, volume and trade count for eight major cryptos, global stock market indices, foreign currency pairs, crude oil and gold price information for February 2020–March 2021. This study also examines whether there was any significant break and how the accuracy of the selected models was impacted.
Findings
Findings suggest that Ridge regression is the most effective model for predicting next-minute BTC prices based on BTC-related covariates such as BTC-open, BTC-high and BTC-low, with a moderate amount of regularization. While BTC-based covariates BTC-open and BTC-low were most significant in predicting BTC closing prices during stable periods, BTC-open and BTC-high were most important during volatile periods. Overall findings suggest that BTC’s price information is the most helpful to predict its next-minute closing price after considering various other asset classes’ price information.
Originality/value
To the best of the authors’ knowledge, this is the first paper to identify the covariates of major cryptocurrencies and predict the next-minute BTC crypto price, with a focus on both crypto-asset and cross-market information.
Details
Keywords
Once regional financial risks erupt, they not only affect the stability and security of the financial system in the region, but also trigger a comprehensive financial crisis…
Abstract
Purpose
Once regional financial risks erupt, they not only affect the stability and security of the financial system in the region, but also trigger a comprehensive financial crisis, damage the national economy, and affect social stability. Therefore, it is necessary to regulate regional financial risks through artificial intelligence methods.
Design/methodology/approach
In this manuscript, we scrutinize the loan data pertaining to aggregated regional financial risks and proffer an ARIMA-SVR loan data regression model, amalgamating traditional statistical regression methods with a machine learning framework. This model initially employs the ARIMA model to accomplish historical data fitting and subsequently utilizes the resultant error as input for SVR to refine the non-linear error. Building upon this, it integrates with the original data to derive optimized prediction results.
Findings
The experimental findings reveal that the ARIMA-SVR (Autoregress Integrated Moving Average Model-Support Vector Regression) method advanced in this discourse surpasses individual methods in terms of RMSE (Root Mean Square Error) and MAE (Mean Absolute Error) indices, exhibiting superiority to the deep learning LSTM method.
Originality/value
An ARIMA-SVR framework for the financial risk recognition is proposed. This presentation furnishes a benchmark for future financial risk prediction and the forecasting of associated time series data.
Details
Keywords
Vicente-Segundo Ruiz-Jacinto, Karina-Silvana Gutiérrez-Valverde, Abrahan-Pablo Aslla-Quispe, José-Manuel Burga-Falla, Aldo Alarcón-Sucasaca and Yersi-Luis Huamán-Romaní
This paper aims to present the novel stacked machine learning approach (SMLA) to estimate low-cycle fatigue (LCF) life of SAC305 solder across structural parts. Using the finite…
Abstract
Purpose
This paper aims to present the novel stacked machine learning approach (SMLA) to estimate low-cycle fatigue (LCF) life of SAC305 solder across structural parts. Using the finite element simulation (FEM) and continuous damage mechanics (CDM) model, a fatigue life database is built. The stacked machine learning (ML) model's iterative optimization during training enables precise fatigue predictions (2.41% root mean square error [RMSE], R2 = 0.975) for diverse structural components. Outliers are found in regression analysis, indicating potential overestimation for thickness transition specimens with extended lifetimes and underestimation for open-hole specimens. Correlations between fatigue life, stress factors, nominal stress and temperature are unveiled, enriching comprehension of LCF, thus enhancing solder behavior predictions.
Design/methodology/approach
This paper introduces stacked ML as a novel approach for estimating LCF life of SAC305 solder in various structural parts. It builds a fatigue life database using FEM and CDM model. The stacked ML model iteratively optimizes its structure, yielding accurate fatigue predictions (2.41% RMSE, R2 = 0.975). Outliers are observed: overestimation for thickness transition specimens and underestimation for open-hole ones. Correlations between fatigue life, stress factors, nominal stress and temperature enhance predictions, deepening understanding of solder behavior.
Findings
The findings of this paper highlight the successful application of the SMLA in accurately estimating the LCF life of SAC305 solder across diverse structural components. The stacked ML model, trained iteratively, demonstrates its effectiveness by producing precise fatigue lifetime predictions with a RMSE of 2.41% and an “R2” value of 0.975. The study also identifies distinct outlier behaviors associated with different structural parts: overestimations for thickness transition specimens with extended fatigue lifetimes and underestimations for open-hole specimens. The research further establishes correlations between fatigue life, stress concentration factors, nominal stress and temperature, enriching the understanding of solder behavior prediction.
Originality/value
The authors confirm the originality of this paper.
Details