Search results

1 – 10 of over 4000
Article
Publication date: 1 February 2016

Manoj Manuja and Deepak Garg

Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those…

Abstract

Purpose

Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those scenarios where similarity among documents is computed considering semantic relationships among their terms. Kernel functions have received major attention because of the unprecedented popularity of SVMs in the field of TC. Most of the kernel functions exploit syntactic structures of the text, but quite a few also use a priori semantic information for knowledge extraction. The purpose of this paper is to investigate semantic kernel functions in the context of TC.

Design/methodology/approach

This work presents performance and accuracy analysis of seven semantic kernel functions (Semantic Smoothing Kernel, Latent Semantic Kernel, Semantic WordNet-based Kernel, Semantic Smoothing Kernel having Implicit Superconcept Expansions, Compactness-based Disambiguation Kernel Function, Omiotis-based S-VSM semantic kernel function and Top-k S-VSM semantic kernel) being implemented with SVM as kernel method. All seven semantic kernels are implemented in SVM-Light tool.

Findings

Performance and accuracy parameters of seven semantic kernel functions have been evaluated and compared. The experimental results show that Top-k S-VSM semantic kernel has the highest performance and accuracy among all the evaluated kernel functions which make it a preferred building block for kernel methods for TC and retrieval.

Research limitations/implications

A combination of semantic kernel function with syntactic kernel function needs to be investigated as there is a scope of further improvement in terms of accuracy and performance in all the seven semantic kernel functions.

Practical implications

This research provides an insight into TC using a priori semantic knowledge. Three commonly used data sets are being exploited. It will be quite interesting to explore these kernel functions on live web data which may test their actual utility in real business scenarios.

Originality/value

Comparison of performance and accuracy parameters is the novel point of this research paper. To the best of the authors’ knowledge, this type of comparison has not been done previously.

Details

Program, vol. 50 no. 1
Type: Research Article
ISSN: 0033-0337

Keywords

Article
Publication date: 14 June 2019

Pingping Xiong, Zhiqing He, Shiting Chen and Mao Peng

In recent years, domestic smog has become increasingly frequent and the adverse effects of smog have increasingly become the focus of public attention. It is a way to analyze such…

Abstract

Purpose

In recent years, domestic smog has become increasingly frequent and the adverse effects of smog have increasingly become the focus of public attention. It is a way to analyze such problems and provide solutions by mathematical methods.

Design/methodology/approach

This paper establishes a new gray model (GM) (1,N) prediction model based on the new kernel and degree of grayness sequences under the case that the interval gray number distribution information is known. First, the new kernel and degree of grayness sequences of the interval gray number sequence are calculated using the reconstruction definition of the kernel and degree of grayness. Then, the GM(1,N) model is formed based on the above new sequences to simulate and predict the kernel and degree of the grayness of the interval gray number sequence. Finally, the upper and lower bounds of the interval gray number are deduced based on the calculation formulas of the kernel and degree of grayness.

Findings

To verify further the practical significance of the model proposed in this paper, the authors apply the model to the simulation and prediction of smog. Compared with the traditional GM(1,N) model, the new GM(1,N) prediction model established in this paper has better prediction effect and accuracy.

Originality/value

This paper improves the traditional GM(1,N) prediction model and establishes a new GM(1,N) prediction model in the case of the known distribution information of the interval gray number of the smog pollutants concentrations data.

Details

Kybernetes, vol. 49 no. 3
Type: Research Article
ISSN: 0368-492X

Keywords

Book part
Publication date: 16 December 2009

Jeffrey S. Racine

The R environment for statistical computing and graphics (R Development Core Team, 2008) offers practitioners a rich set of statistical methods ranging from random number…

Abstract

The R environment for statistical computing and graphics (R Development Core Team, 2008) offers practitioners a rich set of statistical methods ranging from random number generation and optimization methods through regression, panel data, and time series methods, by way of illustration. The standard R distribution (base R) comes preloaded with a rich variety of functionality useful for applied econometricians. This functionality is enhanced by user-supplied packages made available via R servers that are mirrored around the world. Of interest in this chapter are methods for estimating nonparametric and semiparametric models. We summarize many of the facilities in R and consider some tools that might be of interest to those wishing to work with nonparametric methods who want to avoid resorting to programming in C or Fortran but need the speed of compiled code as opposed to interpreted code such as Gauss or Matlab by way of example. We encourage those working in the field to strongly consider implementing their methods in the R environment thereby making their work accessible to the widest possible audience via an open collaborative forum.

Details

Nonparametric Econometric Methods
Type: Book
ISBN: 978-1-84950-624-3

Article
Publication date: 23 January 2019

Dongliang Qi, Dongdong Wang, Like Deng, Xiaolan Xu and Cheng-Tang Wu

Although high-order smooth reproducing kernel mesh-free approximation enables the analysis of structural vibrations in an efficient collocation formulation, there is still a lack…

Abstract

Purpose

Although high-order smooth reproducing kernel mesh-free approximation enables the analysis of structural vibrations in an efficient collocation formulation, there is still a lack of systematic theoretical accuracy assessment for such approach. The purpose of this paper is to present a detailed accuracy analysis for the reproducing kernel mesh-free collocation method regarding structural vibrations.

Design/methodology/approach

Both second-order problems such as one-dimensional (1D) rod and two-dimensional (2D) membrane and fourth-order problems such as Euler–Bernoulli beam and Kirchhoff plate are considered. Staring from a generic equation of motion deduced from the reproducing kernel mesh-free collocation method, a frequency error measure is rationally attained through properly introducing the consistency conditions of reproducing kernel mesh-free shape functions.

Findings

This paper reveals that for the second-order structural vibration problems, the frequency accuracy orders are p and (p − 1) for even and odd degree basis functions; for the fourth-order structural vibration problems, the frequency accuracy orders are (p − 2) and (p − 3) for even and odd degree basis functions, respectively, where p denotes the degree of the basis function used in mesh-free approximation.

Originality/value

A frequency accuracy estimation is achieved for the reproducing kernel mesh-free collocation analysis of structural vibrations, which can effectively underpin the practical applications of this method.

Details

Engineering Computations, vol. 36 no. 3
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 5 March 2018

Tuan Minh Nguyen, Abdelraheem M. Aly and Sang-Wook Lee

The purpose of this paper is to improve the 2D incompressible smoothed particle hydrodynamics (ISPH) method by working on the wall boundary conditions in ISPH method. Here, two…

Abstract

Purpose

The purpose of this paper is to improve the 2D incompressible smoothed particle hydrodynamics (ISPH) method by working on the wall boundary conditions in ISPH method. Here, two different wall boundary conditions in ISPH method including dummy wall particles and analytical kernel renormalization wall boundary conditions have been discussed in details.

Design/methodology/approach

The ISPH algorithm based on the projection method with a divergence velocity condition with improved boundary conditions has been adapted.

Findings

The authors tested the current ISPH method with the improved boundary conditions by a lid-driven cavity for different Reynolds number 100 ≤ Re ≤ 1,000. The results are well validated with the benchmark problems.

Originality/value

In the case of dummy wall boundary particles, the homogeneous Newman boundary condition was applied in solving the linear systems of pressure Poisson equation. In the case of renormalization wall boundary conditions, the authors analytically computed the renormalization factor and its gradient based on a quintic kernel function.

Details

International Journal of Numerical Methods for Heat & Fluid Flow, vol. 28 no. 3
Type: Research Article
ISSN: 0961-5539

Keywords

Article
Publication date: 10 June 2022

Yasser Alharbi

This strategy significantly reduces the computational overhead and storage overhead required when using the kernel density estimation method to calculate the abnormal evaluation…

Abstract

Purpose

This strategy significantly reduces the computational overhead and storage overhead required when using the kernel density estimation method to calculate the abnormal evaluation value of the test sample.

Design/methodology/approach

To effectively deal with the security threats of botnets to the home and personal Internet of Things (IoT), especially for the objective problem of insufficient resources for anomaly detection in the home environment, a novel kernel density estimation-based federated learning-based lightweight Internet of Things anomaly traffic detection based on nuclear density estimation (KDE-LIATD) method. First, the KDE-LIATD method uses Gaussian kernel density estimation method to estimate every normal sample in the training set. The eigenvalue probability density function of the dimensional feature and the corresponding probability density; then, a feature selection algorithm based on kernel density estimation, obtained features that make outstanding contributions to anomaly detection, thereby reducing the feature dimension while improving the accuracy of anomaly detection; finally, the anomaly evaluation value of the test sample is calculated by the cubic spine interpolation method and anomaly detection is performed.

Findings

The simulation experiment results show that the proposed KDE-LIATD method is relatively strong in the detection of abnormal traffic for heterogeneous IoT devices.

Originality/value

With its robustness and compatibility, it can effectively detect abnormal traffic of household and personal IoT botnets.

Details

International Journal of Pervasive Computing and Communications, vol. ahead-of-print no. ahead-of-print
Type: Research Article
ISSN: 1742-7371

Keywords

Article
Publication date: 19 July 2021

Hassan Abdolrezaei, Hassan Siahkali and Javad Olamaei

This paper aims to present a hybrid model to mid-term forecast the load of transmission substations based on the knowledge of expert site and multi-objective posterior framework…

Abstract

Purpose

This paper aims to present a hybrid model to mid-term forecast the load of transmission substations based on the knowledge of expert site and multi-objective posterior framework. The main important challenges in load forecasting are the different behavior of load in specific days. Regular days, holidays and special holidays, days after a holidays and days of load shifting are characterized by abnormal load profiles. The knowledge of these days is verified by expert operators in regional dispatching centers.

Design/methodology/approach

In this paper, a hybrid model for power prediction of transmission substations based on the combination of similar day selection and multi-objective posterior technique has been proposed. In the first step, the important data for prediction is provided. Posterior method is used in the second step for prediction that it is based on kernel functions. A multi-objective optimization has been formulated with three type of output accuracy measurement function that it is solved by non-dominated sorting genetic technique II (NSGT-II) method. TOPSIS way is used to find the best point of Pareto.

Findings

The presented method has been tested in four scenarios for three different transmission stations, and the test results have been compared. The presented results indicate that the presentation method has better results and is robust to different load characteristics, which can be used for better forecasting of different stations for better planning of repairs and network operation.

Originality/value

The main contributions of this paper can be categorized as follows: A hybrid model based on similar days selection and multi-objective framework posterior is presented. Similar day selection is done by expert site that the day type and days with scheduled repair are considered. Hyperparameters of posterior process are found by NSGT-II based on TOPSIS method.

Details

Journal of Engineering, Design and Technology , vol. 19 no. 6
Type: Research Article
ISSN: 1726-0531

Keywords

Book part
Publication date: 16 December 2009

Daniel J. Henderson and Christopher F. Parmeter

Economic conditions such as convexity, homogeneity, homotheticity, and monotonicity are all important assumptions or consequences of assumptions of economic functionals to be…

Abstract

Economic conditions such as convexity, homogeneity, homotheticity, and monotonicity are all important assumptions or consequences of assumptions of economic functionals to be estimated. Recent research has seen a renewed interest in imposing constraints in nonparametric regression. We survey the available methods in the literature, discuss the challenges that present themselves when empirically implementing these methods, and extend an existing method to handle general nonlinear constraints. A heuristic discussion on the empirical implementation for methods that use sequential quadratic programming is provided for the reader, and simulated and empirical evidence on the distinction between constrained and unconstrained nonparametric regression surfaces is covered.

Details

Nonparametric Econometric Methods
Type: Book
ISBN: 978-1-84950-624-3

Article
Publication date: 3 August 2010

Abdoul G. Sam

While the extant literature is replete with theoretical and empirical studies of value at risk (VaR) methods, only a few papers have applied the concept of VaR to quantify market…

1083

Abstract

Purpose

While the extant literature is replete with theoretical and empirical studies of value at risk (VaR) methods, only a few papers have applied the concept of VaR to quantify market risk in the context of agricultural finance. Furthermore, papers that have done so have largely relied on parametric methods to recover estimates of the VaR. The purpose of this paper is to assess extreme market risk on investment in three actively traded agricultural commodity futures.

Design/methodology/approach

A nonparametric Kernel method was implemented which accommodates fat tails and asymmetry of the portfolio return density as well as serial correlation of the data, to estimate market risk for investments in three actively traded agricultural futures contracts: corn, soybeans, and wheat. As a futures contract is a zero‐sum game, the VaR for both short and long sides of the market was computed.

Findings

It was found that wheat futures are riskier than either corn or soybeans futures over both periods considered in the study (2000‐2008 and 2006‐2008) and that all three commodities have experienced a sharp increase in market risk over the 2006‐2008 period, with VaR estimates 10‐43 percent higher than the long‐run estimates.

Research limitations/implications

Research is based on cross‐sectional data and does not allow for dynamic assessment of expenditure elasticities.

Originality/value

This paper differs methodologically from previous applications of VaR in agricultural finance in that a nonparametric Kernel estimator was implemented which is exempt of misspecification risk, in the context of risk management of investment in agricultural futures contracts. The application is particularly relevant to grain elevator businesses which purchase grain from farmers on a forward contract basis and then turn to the futures markets to insure against falling prices.

Details

Agricultural Finance Review, vol. 70 no. 2
Type: Research Article
ISSN: 0002-1466

Keywords

Article
Publication date: 5 April 2021

Zhixin Wang, Peng Xu, Bohan Liu, Yankun Cao, Zhi Liu and Zhaojun Liu

This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some…

Abstract

Purpose

This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some challenges in this field are discussed.

Design/methodology/approach

First, the paper summarized the current research status of the hyperspectral techniques. Then, the paper demonstrated the development of underwater hyperspectral techniques from three major aspects, which are UHI preprocess, unmixing and applications. Finally, the paper presents a conclusion of applications of hyperspectral imaging and future research directions.

Findings

Various methods and scenarios for underwater object detection with hyperspectral imaging are compared, which include preprocessing, unmixing and classification. A summary is made to demonstrate the application scope and results of different methods, which may play an important role in the application of underwater hyperspectral object detection in the future.

Originality/value

This paper introduced several methods of hyperspectral image process, give out the conclusion of the advantages and disadvantages of each method, then demonstrated the challenges we face and the possible way to deal with them.

Details

Sensor Review, vol. 41 no. 2
Type: Research Article
ISSN: 0260-2288

Keywords

1 – 10 of over 4000