Search results
1 – 10 of 267Ghoulemallah Boukhalfa, Sebti Belkacem, Abdesselem Chikhi and Said Benaggoune
This paper presents the particle swarm optimization (PSO) algorithm in conjuction with the fuzzy logic method in order to achieve an optimized tuning of a proportional integral…
Abstract
This paper presents the particle swarm optimization (PSO) algorithm in conjuction with the fuzzy logic method in order to achieve an optimized tuning of a proportional integral derivative controller (PID) in the DTC control loops of dual star induction motor (DSIM). The fuzzy controller is insensitive to parametric variations, however, with the PSO-based optimization approach we obtain a judicious choice of the gains to make the system more robust. According to Matlab simulation, the results demonstrate that the hybrid DTC of DSIM improves the speed loop response, ensures the system stability, reduces the steady state error and enhances the rising time. Moreover, with this controller, the disturbances do not affect the motor performances.
Details
Keywords
Carlos Alberto Rojas Trejos, Jose D. Meisel and Wilson Adarme Jaimes
The purpose of this paper is to review the relevant literature in order to identify trends and suggest some possible directions for future research in the framework of…
Abstract
Purpose
The purpose of this paper is to review the relevant literature in order to identify trends and suggest some possible directions for future research in the framework of humanitarian aid distribution logistics with accessibility constraints.
Design/methodology/approach
The authors developed a systematic literature review to study the state of the art on distribution logistics considering accessibility constraints. The electronic databases used were Web of science, Scopus, Science Direct, Jstor, Emerald, EBSCO, Scielo and Redalyc. As a result, 49 articles were reviewed in detail.
Findings
This study identified some gaps, as well as some research opportunities. The main conclusions are the need for further studies on the interrelationships and hierarchies of multiple actors, explore intermodality, transshipment options and redistribution relief goods to avoid severe shortages in some nodes and excess inventory in others, studies of the vulnerability of transport networks, correlational analysis of road failures and other future lines.
Research limitations/implications
The bibliography is limited to peer-reviewed academic journals due to their academic relevance, accessibility and ease of searching. Most of the studies included in the review were conducted in high-income countries, which may limit the generalizability of the results to low-income countries. However, the authors focused on databases covering important journals on humanitarian logistics.
Originality/value
This paper contextualises and synthesises research into humanitarian aid distribution logistics with accessibility constrains, highlights key themes and suggests areas for further research.
Details
Keywords
Edoardo Ramalli and Barbara Pernici
Experiments are the backbone of the development process of data-driven predictive models for scientific applications. The quality of the experiments directly impacts the model…
Abstract
Purpose
Experiments are the backbone of the development process of data-driven predictive models for scientific applications. The quality of the experiments directly impacts the model performance. Uncertainty inherently affects experiment measurements and is often missing in the available data sets due to its estimation cost. For similar reasons, experiments are very few compared to other data sources. Discarding experiments based on the missing uncertainty values would preclude the development of predictive models. Data profiling techniques are fundamental to assess data quality, but some data quality dimensions are challenging to evaluate without knowing the uncertainty. In this context, this paper aims to predict the missing uncertainty of the experiments.
Design/methodology/approach
This work presents a methodology to forecast the experiments’ missing uncertainty, given a data set and its ontological description. The approach is based on knowledge graph embeddings and leverages the task of link prediction over a knowledge graph representation of the experiments database. The validity of the methodology is first tested in multiple conditions using synthetic data and then applied to a large data set of experiments in the chemical kinetic domain as a case study.
Findings
The analysis results of different test case scenarios suggest that knowledge graph embedding can be used to predict the missing uncertainty of the experiments when there is a hidden relationship between the experiment metadata and the uncertainty values. The link prediction task is also resilient to random noise in the relationship. The knowledge graph embedding outperforms the baseline results if the uncertainty depends upon multiple metadata.
Originality/value
The employment of knowledge graph embedding to predict the missing experimental uncertainty is a novel alternative to the current and more costly techniques in the literature. Such contribution permits a better data quality profiling of scientific repositories and improves the development process of data-driven models based on scientific experiments.
Details
Keywords
Rogelio Ladrón de Guevara Cortés, Leticia Eva Tolosa and María Paula Rojo
This paper aims to provide empirical evidence for using the prospect theory (PT) basic assumptions in the Argentine context. Mainly, this study analysed the financial…
Abstract
Purpose
This paper aims to provide empirical evidence for using the prospect theory (PT) basic assumptions in the Argentine context. Mainly, this study analysed the financial decision-making process in students of the economic-administrative academic area of two universities, one public and one private, in Córdoba.
Design/methodology/approach
The analysis methodology included (1) the descriptive statistical analysis to identify the presence of the certainty, reflection and isolation effects; (2) the construction of a set of indicators on the application of the PT; (3) the chi-squared independence test, to determine if the decisions made are independent of the degree course taken; (4) the non-parametric Kruskal–Wallis test, to determine if the decisions made by individuals vary according to the semesters taken or students' levels of progress; and (5) the non-parametric Mann–Whitney test, to determine if there are differences between the decisions made by men and women.
Findings
The empirical results provided evidence on the effects of certainty, reflection and isolation in both universities, concluding that the study participants make financial decisions in situations of uncertainty based more on PT than on expected utility theory.
Originality/value
This study contributes to the empirical evidence in a different Latin-American context, confirming that individuals make financial decisions based on the PT independently of their degree course, semester, level of advance, gender or the kind of university where they belong (public or private).
Details
Keywords
When the probability of each model is known, a natural idea is to select the most probable model. However, in many practical situations, the exact values of these probabilities…
Abstract
Purpose
When the probability of each model is known, a natural idea is to select the most probable model. However, in many practical situations, the exact values of these probabilities are not known; only the intervals that contain these values are known. In such situations, a natural idea is to select some probabilities from these intervals and to select a model with the largest selected probabilities. The purpose of this study is to decide how to most adequately select these probabilities.
Design/methodology/approach
It is desirable to have a probability-selection method that preserves independence. If, according to the probability intervals, the two events were independent, then the selection of probabilities within the intervals should preserve this independence.
Findings
The paper describes all techniques for decision making under interval uncertainty about probabilities that are consistent with independence. It is proved that these techniques form a 1-parametric family, a family that has already been successfully used in such decision problems.
Originality/value
This study provides a theoretical explanation of an empirically successful technique for decision-making under interval uncertainty about probabilities. This explanation is based on the natural idea that the method for selecting probabilities from the corresponding intervals should preserve independence.
Details
Keywords
Olga Kosheleva, Vladik Kreinovich and Uyen Pham
In many real-life situations, we do not know the exact values of the expected gain corresponding to different possible actions, we only have lower and upper bounds on these gains…
Abstract
Purpose
In many real-life situations, we do not know the exact values of the expected gain corresponding to different possible actions, we only have lower and upper bounds on these gains – i.e., in effect, intervals of possible gain values. The purpose of this study is to describe all possible ways to make decisions under such interval uncertainty.
Design/methodology/approach
The authors used both natural invariance and additivity requirements.
Findings
The authors demonstrated that natural requirements – invariance or additivity – led to a two-parametric family of possible decision-making strategies.
Originality/value
This is a first description of all reasonable strategies for decision-making under interval uncertainty – strategies that satisfy natural requirements of invariance or additivity.
Details
Keywords
Munazza Jabeen and Saba Kausar
This paper aims to examine the performance of Islamic and conventional stocks listed at the Pakistan Stock Exchange by using both parametric and non-parametric approaches. The…
Abstract
Purpose
This paper aims to examine the performance of Islamic and conventional stocks listed at the Pakistan Stock Exchange by using both parametric and non-parametric approaches. The motivation is to do risk-return analysis of Islamic stock prices and conventional stock prices.
Design/methodology/approach
It uses various measures of performance, e.g. Sharpe ratio, Treynor ratio, Jensen's alpha, beta, generalized auto-regressive conditional heteroskedasticity and stochastic dominance. Using the Karachi Meezan Index-30 (KMI-30) and the Karachi Stock Exchange Index-30 (KSE-30) as proxies for Islamic and conventional stock prices, respectively, it examines the performance of Islamic and conventional stocks. The daily data of KMI-30 and KSE-30, covering period from June 9, 2009 to June 20, 2020 are used.
Findings
The results show that the overall KMI-30 outperforms the KSE-30. The returns of the KMI-30 are greater than the KSE-30. However, the risk and volatility of the KMI-30 and KSE-30 are similar. Further, the KMI-30 has higher excess returns per unit of total risk than the KSE-30. But both indexes have similar excess returns per unit of systematic risk. Moreover, the KMI-30 returns have stochastically dominance over the KSE-30 returns. These results reveal that the Islamic index performs better than the conventional index.
Practical implications
The findings provide several practical implications in financial and investment decisions making by investors, managers and policymakers such as strategies for asset allocation and investment. Further, in risk management, it provides guidance for allocating portfolios and managing risk. The investment in Islamic stocks may mitigate potential risk within asset portfolios.
Originality/value
This research is unique in its approach to the analysis of the performance comparison of conventional and Islamic stock by using comprehensive parametric and non-parametric estimation techniques. Such research has not been undertaken in the Pakistan's equity market since.
Details
Keywords
Warisa Thangjai and Sa-Aat Niwitpong
Confidence intervals play a crucial role in economics and finance, providing a credible range of values for an unknown parameter along with a corresponding level of certainty…
Abstract
Purpose
Confidence intervals play a crucial role in economics and finance, providing a credible range of values for an unknown parameter along with a corresponding level of certainty. Their applications encompass economic forecasting, market research, financial forecasting, econometric analysis, policy analysis, financial reporting, investment decision-making, credit risk assessment and consumer confidence surveys. Signal-to-noise ratio (SNR) finds applications in economics and finance across various domains such as economic forecasting, financial modeling, market analysis and risk assessment. A high SNR indicates a robust and dependable signal, simplifying the process of making well-informed decisions. On the other hand, a low SNR indicates a weak signal that could be obscured by noise, so decision-making procedures need to take this into serious consideration. This research focuses on the development of confidence intervals for functions derived from the SNR and explores their application in the fields of economics and finance.
Design/methodology/approach
The construction of the confidence intervals involved the application of various methodologies. For the SNR, confidence intervals were formed using the generalized confidence interval (GCI), large sample and Bayesian approaches. The difference between SNRs was estimated through the GCI, large sample, method of variance estimates recovery (MOVER), parametric bootstrap and Bayesian approaches. Additionally, confidence intervals for the common SNR were constructed using the GCI, adjusted MOVER, computational and Bayesian approaches. The performance of these confidence intervals was assessed using coverage probability and average length, evaluated through Monte Carlo simulation.
Findings
The GCI approach demonstrated superior performance over other approaches in terms of both coverage probability and average length for the SNR and the difference between SNRs. Hence, employing the GCI approach is advised for constructing confidence intervals for these parameters. As for the common SNR, the Bayesian approach exhibited the shortest average length. Consequently, the Bayesian approach is recommended for constructing confidence intervals for the common SNR.
Originality/value
This research presents confidence intervals for functions of the SNR to assess SNR estimation in the fields of economics and finance.
Details
Keywords
Philipp Geiberger, Zhendong Liu, Mats Berg and Christoph Domay
For billing purposes, heavy-haul locomotives in Sweden are equipped with on-board energy meters, which can record several parameters, e.g., used energy, regenerated energy, speed…
Abstract
Purpose
For billing purposes, heavy-haul locomotives in Sweden are equipped with on-board energy meters, which can record several parameters, e.g., used energy, regenerated energy, speed and position. Since there is a strong demand for improving energy efficiency in Sweden, data from the energy meters can be used to obtain a better understanding of the detailed energy usage of heavy-haul trains and identify potential for future improvements.
Design/methodology/approach
To monitor energy efficiency, the present study, therefore, develops key performance indicators (KPIs), which can be calculated with energy meter data to reflect the energy efficiency of heavy-haul trains in operation. Energy meter data of IORE class locomotives, hauling highly uniform 30-tonne axle load trains with 68 wagons, together with additional data sources, are analysed to identify significant parameters for describing driver influence on energy usage.
Findings
Results show that driver behaviour varies significantly and has the single largest influence on energy usage. Furthermore, parametric studies are performed with help of simulation to identify the influence of different operational and rolling stock conditions, e.g., axle loads and number of wagons, on energy usage.
Originality/value
Based on the parametric studies, some operational parameters which have significant impact on energy efficiency are found and then the KPIs are derived. In the end, some possible measures for improving energy performance in heavy-haul operations are given.
Details
Keywords
Wassim Ben Ayed and Rim Ben Hassen
This research aims to evaluate the accuracy of several Value-at-Risk (VaR) approaches for determining the Minimum Capital Requirement (MCR) for Islamic stock markets during the…
Abstract
Purpose
This research aims to evaluate the accuracy of several Value-at-Risk (VaR) approaches for determining the Minimum Capital Requirement (MCR) for Islamic stock markets during the pandemic health crisis.
Design/methodology/approach
This research evaluates the performance of numerous VaR models for computing the MCR for market risk in compliance with the Basel II and Basel II.5 guidelines for ten Islamic indices. Five models were applied—namely the RiskMetrics, Generalized Autoregressive Conditional Heteroskedasticity, denoted (GARCH), fractional integrated GARCH, denoted (FIGARCH), and SPLINE-GARCH approaches—under three innovations (normal (N), Student (St) and skewed-Student (Sk-t) and the extreme value theory (EVT).
Findings
The main findings of this empirical study reveal that (1) extreme value theory performs better for most indices during the market crisis and (2) VaR models under a normal distribution provide quite poor performance than models with fat-tailed innovations in terms of risk estimation.
Research limitations/implications
Since the world is now undergoing the third wave of the COVID-19 pandemic, this study will not be able to assess performance of VaR models during the fourth wave of COVID-19.
Practical implications
The results suggest that the Islamic Financial Services Board (IFSB) should enhance market discipline mechanisms, while central banks and national authorities should harmonize their regulatory frameworks in line with Basel/IFSB reform agenda.
Originality/value
Previous studies focused on evaluating market risk models using non-Islamic indexes. However, this research uses the Islamic indexes to analyze the VaR forecasting models. Besides, they tested the accuracy of VaR models based on traditional GARCH models, whereas the authors introduce the Spline GARCH developed by Engle and Rangel (2008). Finally, most studies have focus on the period of 2007–2008 financial crisis, while the authors investigate the issue of market risk quantification for several Islamic market equity during the sanitary crisis of COVID-19.
Details