Search results

1 – 10 of 67

Abstract

Details

Understanding Financial Risk Management, Third Edition
Type: Book
ISBN: 978-1-83753-253-7

Article
Publication date: 19 April 2024

Jitendra Gaur, Kumkum Bharti and Rahul Bajaj

Allocation of the marketing budget has become increasingly challenging due to the diverse channel exposure to customers. This study aims to enhance global marketing knowledge by…

Abstract

Purpose

Allocation of the marketing budget has become increasingly challenging due to the diverse channel exposure to customers. This study aims to enhance global marketing knowledge by introducing an ensemble attribution model to optimize marketing budget allocation for online marketing channels. As empirical research, this study demonstrates the supremacy of the ensemble model over standalone models.

Design/methodology/approach

The transactional data set for car insurance from an Indian insurance aggregator is used in this empirical study. The data set contains information from more than three million platform visitors. A robust ensemble model is created by combining results from two probabilistic models, namely, the Markov chain model and the Shapley value. These results are compared and validated with heuristic models. Also, the performances of online marketing channels and attribution models are evaluated based on the devices used (i.e. desktop vs mobile).

Findings

Channel importance charts for desktop and mobile devices are analyzed to understand the top contributing online marketing channels. Customer relationship management-emailers and Google cost per click a paid advertising is identified as the top two marketing channels for desktop and mobile channels. The research reveals that ensemble model accuracy is better than the standalone model, that is, the Markov chain model and the Shapley value.

Originality/value

To the best of the authors’ knowledge, the current research is the first of its kind to introduce ensemble modeling for solving attribution problems in online marketing. A comparison with heuristic models using different devices (desktop and mobile) offers insights into the results with heuristic models.

Details

Global Knowledge, Memory and Communication, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2514-9342

Keywords

Article
Publication date: 15 December 2023

Mondher Bouattour and Anthony Miloudi

The purpose of this paper is to bridge the gap between the existing theoretical and empirical studies by examining the asymmetric return–volume relationship. Indeed, the authors…

Abstract

Purpose

The purpose of this paper is to bridge the gap between the existing theoretical and empirical studies by examining the asymmetric return–volume relationship. Indeed, the authors aim to shed light on the return–volume linkages for French-listed small and medium-sized enterprises (SMEs) compared to blue chips across different market regimes.

Design/methodology/approach

This study includes both large capitalizations included in the CAC 40 index and listed SMEs included in the Euronext Growth All Share index. The Markov-switching (MS) approach is applied to understand the asymmetric relationship between trading volume and stock returns. The study investigates also the causal impact between stock returns and trading volume using regime-dependent Granger causality tests.

Findings

Asymmetric contemporaneous and lagged relationships between stock returns and trading volume are found for both large capitalizations and listed SMEs. However, the causality investigation reveals some differences between large capitalizations and SMEs. Indeed, causal relationships depend on market conditions and the size of the market.

Research limitations/implications

This paper explains the asymmetric return–volume relationship for both large capitalizations and listed SMEs by incorporating several psychological biases, such as the disposition effect, investor overconfidence and self-attribution bias. Future research needs to deepen the analysis especially for SMEs as most of the literature focuses on large capitalizations.

Practical implications

This empirical study has fundamental implications for portfolio management. The findings provide a deeper understanding of how trading activity impact current returns and vice versa. The authors’ results constitute an important input to build and control trading strategies.

Originality/value

This paper fills the literature gap on the asymmetric return–volume relationship across different regimes. To the best of the authors’ knowledge, the present study is the first empirical attempt to test the asymmetric return–volume relationship for listed SMEs by using an accurate MS framework.

Details

Review of Accounting and Finance, vol. 23 no. 2
Type: Research Article
ISSN: 1475-7702

Keywords

Abstract

Details

Understanding Financial Risk Management, Third Edition
Type: Book
ISBN: 978-1-83753-253-7

Article
Publication date: 19 March 2024

Mingke Gao, Zhenyu Zhang, Jinyuan Zhang, Shihao Tang, Han Zhang and Tao Pang

Because of the various advantages of reinforcement learning (RL) mentioned above, this study uses RL to train unmanned aerial vehicles to perform two tasks: target search and…

Abstract

Purpose

Because of the various advantages of reinforcement learning (RL) mentioned above, this study uses RL to train unmanned aerial vehicles to perform two tasks: target search and cooperative obstacle avoidance.

Design/methodology/approach

This study draws inspiration from the recurrent state-space model and recurrent models (RPM) to propose a simpler yet highly effective model called the unmanned aerial vehicles prediction model (UAVPM). The main objective is to assist in training the UAV representation model with a recurrent neural network, using the soft actor-critic algorithm.

Findings

This study proposes a generalized actor-critic framework consisting of three modules: representation, policy and value. This architecture serves as the foundation for training UAVPM. This study proposes the UAVPM, which is designed to aid in training the recurrent representation using the transition model, reward recovery model and observation recovery model. Unlike traditional approaches reliant solely on reward signals, RPM incorporates temporal information. In addition, it allows the inclusion of extra knowledge or information from virtual training environments. This study designs UAV target search and UAV cooperative obstacle avoidance tasks. The algorithm outperforms baselines in these two environments.

Originality/value

It is important to note that UAVPM does not play a role in the inference phase. This means that the representation model and policy remain independent of UAVPM. Consequently, this study can introduce additional “cheating” information from virtual training environments to guide the UAV representation without concerns about its real-world existence. By leveraging historical information more effectively, this study enhances UAVs’ decision-making abilities, thus improving the performance of both tasks at hand.

Details

International Journal of Web Information Systems, vol. 20 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Content available
Book part
Publication date: 27 May 2024

Angelo Corelli

Abstract

Details

Understanding Financial Risk Management, Third Edition
Type: Book
ISBN: 978-1-83753-253-7

Article
Publication date: 8 May 2024

Hongze Wang

Many practical control problems require achieving multiple objectives, and these objectives often conflict with each other. The existing multi-objective evolutionary reinforcement…

Abstract

Purpose

Many practical control problems require achieving multiple objectives, and these objectives often conflict with each other. The existing multi-objective evolutionary reinforcement learning algorithms cannot achieve good search results when solving such problems. It is necessary to design a new multi-objective evolutionary reinforcement learning algorithm with a stronger searchability.

Design/methodology/approach

The multi-objective reinforcement learning algorithm proposed in this paper is based on the evolutionary computation framework. In each generation, this study uses the long-short-term selection method to select parent policies. The long-term selection is based on the improvement of policy along the predefined optimization direction in the previous generation. The short-term selection uses a prediction model to predict the optimization direction that may have the greatest improvement on overall population performance. In the evolutionary stage, the penalty-based nonlinear scalarization method is used to scalarize the multi-dimensional advantage functions, and the nonlinear multi-objective policy gradient is designed to optimize the parent policies along the predefined directions.

Findings

The penalty-based nonlinear scalarization method can force policies to improve along the predefined optimization directions. The long-short-term optimization method can alleviate the exploration-exploitation problem, enabling the algorithm to explore unknown regions while ensuring that potential policies are fully optimized. The combination of these designs can effectively improve the performance of the final population.

Originality/value

A multi-objective evolutionary reinforcement learning algorithm with stronger searchability has been proposed. This algorithm can find a Pareto policy set with better convergence, diversity and density.

Details

Robotic Intelligence and Automation, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 2754-6969

Keywords

Open Access
Article
Publication date: 19 April 2024

Bong-Gyu Jang and Hyeng Keun Koo

We present an approach for pricing American put options with a regime-switching volatility. Our method reveals that the option price can be expressed as the sum of two components…

Abstract

We present an approach for pricing American put options with a regime-switching volatility. Our method reveals that the option price can be expressed as the sum of two components: the price of a European put option and the premium associated with the early exercise privilege. Our analysis demonstrates that, under these conditions, the perpetual put option consistently commands a higher price during periods of high volatility compared to those of low volatility. Moreover, we establish that the optimal exercise boundary is lower in high-volatility regimes than in low-volatility regimes. Additionally, we develop an analytical framework to describe American puts with an Erlang-distributed random-time horizon, which allows us to propose a numerical technique for approximating the value of American puts with finite expiry. We also show that a combined approach involving randomization and Richardson extrapolation can be a robust numerical algorithm for estimating American put prices with finite expiry.

Details

Journal of Derivatives and Quantitative Studies: 선물연구, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1229-988X

Keywords

Article
Publication date: 1 April 2024

Tao Pang, Wenwen Xiao, Yilin Liu, Tao Wang, Jie Liu and Mingke Gao

This paper aims to study the agent learning from expert demonstration data while incorporating reinforcement learning (RL), which enables the agent to break through the…

Abstract

Purpose

This paper aims to study the agent learning from expert demonstration data while incorporating reinforcement learning (RL), which enables the agent to break through the limitations of expert demonstration data and reduces the dimensionality of the agent’s exploration space to speed up the training convergence rate.

Design/methodology/approach

Firstly, the decay weight function is set in the objective function of the agent’s training to combine both types of methods, and both RL and imitation learning (IL) are considered to guide the agent's behavior when updating the policy. Second, this study designs a coupling utilization method between the demonstration trajectory and the training experience, so that samples from both aspects can be combined during the agent’s learning process, and the utilization rate of the data and the agent’s learning speed can be improved.

Findings

The method is superior to other algorithms in terms of convergence speed and decision stability, avoiding training from scratch for reward values, and breaking through the restrictions brought by demonstration data.

Originality/value

The agent can adapt to dynamic scenes through exploration and trial-and-error mechanisms based on the experience of demonstrating trajectories. The demonstration data set used in IL and the experience samples obtained in the process of RL are coupled and used to improve the data utilization efficiency and the generalization ability of the agent.

Details

International Journal of Web Information Systems, vol. 20 no. 3
Type: Research Article
ISSN: 1744-0084

Keywords

Abstract

Details

Understanding Financial Risk Management, Third Edition
Type: Book
ISBN: 978-1-83753-253-7

1 – 10 of 67