Search results
1 – 10 of over 4000Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those…
Abstract
Purpose
Syntax-based text classification (TC) mechanisms have been overtly replaced by semantic-based systems in recent years. Semantic-based TC systems are particularly useful in those scenarios where similarity among documents is computed considering semantic relationships among their terms. Kernel functions have received major attention because of the unprecedented popularity of SVMs in the field of TC. Most of the kernel functions exploit syntactic structures of the text, but quite a few also use a priori semantic information for knowledge extraction. The purpose of this paper is to investigate semantic kernel functions in the context of TC.
Design/methodology/approach
This work presents performance and accuracy analysis of seven semantic kernel functions (Semantic Smoothing Kernel, Latent Semantic Kernel, Semantic WordNet-based Kernel, Semantic Smoothing Kernel having Implicit Superconcept Expansions, Compactness-based Disambiguation Kernel Function, Omiotis-based S-VSM semantic kernel function and Top-k S-VSM semantic kernel) being implemented with SVM as kernel method. All seven semantic kernels are implemented in SVM-Light tool.
Findings
Performance and accuracy parameters of seven semantic kernel functions have been evaluated and compared. The experimental results show that Top-k S-VSM semantic kernel has the highest performance and accuracy among all the evaluated kernel functions which make it a preferred building block for kernel methods for TC and retrieval.
Research limitations/implications
A combination of semantic kernel function with syntactic kernel function needs to be investigated as there is a scope of further improvement in terms of accuracy and performance in all the seven semantic kernel functions.
Practical implications
This research provides an insight into TC using a priori semantic knowledge. Three commonly used data sets are being exploited. It will be quite interesting to explore these kernel functions on live web data which may test their actual utility in real business scenarios.
Originality/value
Comparison of performance and accuracy parameters is the novel point of this research paper. To the best of the authors’ knowledge, this type of comparison has not been done previously.
Details
Keywords
Pingping Xiong, Zhiqing He, Shiting Chen and Mao Peng
In recent years, domestic smog has become increasingly frequent and the adverse effects of smog have increasingly become the focus of public attention. It is a way to analyze such…
Abstract
Purpose
In recent years, domestic smog has become increasingly frequent and the adverse effects of smog have increasingly become the focus of public attention. It is a way to analyze such problems and provide solutions by mathematical methods.
Design/methodology/approach
This paper establishes a new gray model (GM) (1,N) prediction model based on the new kernel and degree of grayness sequences under the case that the interval gray number distribution information is known. First, the new kernel and degree of grayness sequences of the interval gray number sequence are calculated using the reconstruction definition of the kernel and degree of grayness. Then, the GM(1,N) model is formed based on the above new sequences to simulate and predict the kernel and degree of the grayness of the interval gray number sequence. Finally, the upper and lower bounds of the interval gray number are deduced based on the calculation formulas of the kernel and degree of grayness.
Findings
To verify further the practical significance of the model proposed in this paper, the authors apply the model to the simulation and prediction of smog. Compared with the traditional GM(1,N) model, the new GM(1,N) prediction model established in this paper has better prediction effect and accuracy.
Originality/value
This paper improves the traditional GM(1,N) prediction model and establishes a new GM(1,N) prediction model in the case of the known distribution information of the interval gray number of the smog pollutants concentrations data.
Details
Keywords
The R environment for statistical computing and graphics (R Development Core Team, 2008) offers practitioners a rich set of statistical methods ranging from random number…
Abstract
The R environment for statistical computing and graphics (R Development Core Team, 2008) offers practitioners a rich set of statistical methods ranging from random number generation and optimization methods through regression, panel data, and time series methods, by way of illustration. The standard R distribution (base R) comes preloaded with a rich variety of functionality useful for applied econometricians. This functionality is enhanced by user-supplied packages made available via R servers that are mirrored around the world. Of interest in this chapter are methods for estimating nonparametric and semiparametric models. We summarize many of the facilities in R and consider some tools that might be of interest to those wishing to work with nonparametric methods who want to avoid resorting to programming in C or Fortran but need the speed of compiled code as opposed to interpreted code such as Gauss or Matlab by way of example. We encourage those working in the field to strongly consider implementing their methods in the R environment thereby making their work accessible to the widest possible audience via an open collaborative forum.
Dongliang Qi, Dongdong Wang, Like Deng, Xiaolan Xu and Cheng-Tang Wu
Although high-order smooth reproducing kernel mesh-free approximation enables the analysis of structural vibrations in an efficient collocation formulation, there is still a lack…
Abstract
Purpose
Although high-order smooth reproducing kernel mesh-free approximation enables the analysis of structural vibrations in an efficient collocation formulation, there is still a lack of systematic theoretical accuracy assessment for such approach. The purpose of this paper is to present a detailed accuracy analysis for the reproducing kernel mesh-free collocation method regarding structural vibrations.
Design/methodology/approach
Both second-order problems such as one-dimensional (1D) rod and two-dimensional (2D) membrane and fourth-order problems such as Euler–Bernoulli beam and Kirchhoff plate are considered. Staring from a generic equation of motion deduced from the reproducing kernel mesh-free collocation method, a frequency error measure is rationally attained through properly introducing the consistency conditions of reproducing kernel mesh-free shape functions.
Findings
This paper reveals that for the second-order structural vibration problems, the frequency accuracy orders are p and (p − 1) for even and odd degree basis functions; for the fourth-order structural vibration problems, the frequency accuracy orders are (p − 2) and (p − 3) for even and odd degree basis functions, respectively, where p denotes the degree of the basis function used in mesh-free approximation.
Originality/value
A frequency accuracy estimation is achieved for the reproducing kernel mesh-free collocation analysis of structural vibrations, which can effectively underpin the practical applications of this method.
Details
Keywords
Tuan Minh Nguyen, Abdelraheem M. Aly and Sang-Wook Lee
The purpose of this paper is to improve the 2D incompressible smoothed particle hydrodynamics (ISPH) method by working on the wall boundary conditions in ISPH method. Here, two…
Abstract
Purpose
The purpose of this paper is to improve the 2D incompressible smoothed particle hydrodynamics (ISPH) method by working on the wall boundary conditions in ISPH method. Here, two different wall boundary conditions in ISPH method including dummy wall particles and analytical kernel renormalization wall boundary conditions have been discussed in details.
Design/methodology/approach
The ISPH algorithm based on the projection method with a divergence velocity condition with improved boundary conditions has been adapted.
Findings
The authors tested the current ISPH method with the improved boundary conditions by a lid-driven cavity for different Reynolds number 100 ≤ Re ≤ 1,000. The results are well validated with the benchmark problems.
Originality/value
In the case of dummy wall boundary particles, the homogeneous Newman boundary condition was applied in solving the linear systems of pressure Poisson equation. In the case of renormalization wall boundary conditions, the authors analytically computed the renormalization factor and its gradient based on a quintic kernel function.
Details
Keywords
This strategy significantly reduces the computational overhead and storage overhead required when using the kernel density estimation method to calculate the abnormal evaluation…
Abstract
Purpose
This strategy significantly reduces the computational overhead and storage overhead required when using the kernel density estimation method to calculate the abnormal evaluation value of the test sample.
Design/methodology/approach
To effectively deal with the security threats of botnets to the home and personal Internet of Things (IoT), especially for the objective problem of insufficient resources for anomaly detection in the home environment, a novel kernel density estimation-based federated learning-based lightweight Internet of Things anomaly traffic detection based on nuclear density estimation (KDE-LIATD) method. First, the KDE-LIATD method uses Gaussian kernel density estimation method to estimate every normal sample in the training set. The eigenvalue probability density function of the dimensional feature and the corresponding probability density; then, a feature selection algorithm based on kernel density estimation, obtained features that make outstanding contributions to anomaly detection, thereby reducing the feature dimension while improving the accuracy of anomaly detection; finally, the anomaly evaluation value of the test sample is calculated by the cubic spine interpolation method and anomaly detection is performed.
Findings
The simulation experiment results show that the proposed KDE-LIATD method is relatively strong in the detection of abnormal traffic for heterogeneous IoT devices.
Originality/value
With its robustness and compatibility, it can effectively detect abnormal traffic of household and personal IoT botnets.
Details
Keywords
Hassan Abdolrezaei, Hassan Siahkali and Javad Olamaei
This paper aims to present a hybrid model to mid-term forecast the load of transmission substations based on the knowledge of expert site and multi-objective posterior framework…
Abstract
Purpose
This paper aims to present a hybrid model to mid-term forecast the load of transmission substations based on the knowledge of expert site and multi-objective posterior framework. The main important challenges in load forecasting are the different behavior of load in specific days. Regular days, holidays and special holidays, days after a holidays and days of load shifting are characterized by abnormal load profiles. The knowledge of these days is verified by expert operators in regional dispatching centers.
Design/methodology/approach
In this paper, a hybrid model for power prediction of transmission substations based on the combination of similar day selection and multi-objective posterior technique has been proposed. In the first step, the important data for prediction is provided. Posterior method is used in the second step for prediction that it is based on kernel functions. A multi-objective optimization has been formulated with three type of output accuracy measurement function that it is solved by non-dominated sorting genetic technique II (NSGT-II) method. TOPSIS way is used to find the best point of Pareto.
Findings
The presented method has been tested in four scenarios for three different transmission stations, and the test results have been compared. The presented results indicate that the presentation method has better results and is robust to different load characteristics, which can be used for better forecasting of different stations for better planning of repairs and network operation.
Originality/value
The main contributions of this paper can be categorized as follows: A hybrid model based on similar days selection and multi-objective framework posterior is presented. Similar day selection is done by expert site that the day type and days with scheduled repair are considered. Hyperparameters of posterior process are found by NSGT-II based on TOPSIS method.
Details
Keywords
Daniel J. Henderson and Christopher F. Parmeter
Economic conditions such as convexity, homogeneity, homotheticity, and monotonicity are all important assumptions or consequences of assumptions of economic functionals to be…
Abstract
Economic conditions such as convexity, homogeneity, homotheticity, and monotonicity are all important assumptions or consequences of assumptions of economic functionals to be estimated. Recent research has seen a renewed interest in imposing constraints in nonparametric regression. We survey the available methods in the literature, discuss the challenges that present themselves when empirically implementing these methods, and extend an existing method to handle general nonlinear constraints. A heuristic discussion on the empirical implementation for methods that use sequential quadratic programming is provided for the reader, and simulated and empirical evidence on the distinction between constrained and unconstrained nonparametric regression surfaces is covered.
While the extant literature is replete with theoretical and empirical studies of value at risk (VaR) methods, only a few papers have applied the concept of VaR to quantify market…
Abstract
Purpose
While the extant literature is replete with theoretical and empirical studies of value at risk (VaR) methods, only a few papers have applied the concept of VaR to quantify market risk in the context of agricultural finance. Furthermore, papers that have done so have largely relied on parametric methods to recover estimates of the VaR. The purpose of this paper is to assess extreme market risk on investment in three actively traded agricultural commodity futures.
Design/methodology/approach
A nonparametric Kernel method was implemented which accommodates fat tails and asymmetry of the portfolio return density as well as serial correlation of the data, to estimate market risk for investments in three actively traded agricultural futures contracts: corn, soybeans, and wheat. As a futures contract is a zero‐sum game, the VaR for both short and long sides of the market was computed.
Findings
It was found that wheat futures are riskier than either corn or soybeans futures over both periods considered in the study (2000‐2008 and 2006‐2008) and that all three commodities have experienced a sharp increase in market risk over the 2006‐2008 period, with VaR estimates 10‐43 percent higher than the long‐run estimates.
Research limitations/implications
Research is based on cross‐sectional data and does not allow for dynamic assessment of expenditure elasticities.
Originality/value
This paper differs methodologically from previous applications of VaR in agricultural finance in that a nonparametric Kernel estimator was implemented which is exempt of misspecification risk, in the context of risk management of investment in agricultural futures contracts. The application is particularly relevant to grain elevator businesses which purchase grain from farmers on a forward contract basis and then turn to the futures markets to insure against falling prices.
Details
Keywords
Zhixin Wang, Peng Xu, Bohan Liu, Yankun Cao, Zhi Liu and Zhaojun Liu
This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some…
Abstract
Purpose
This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some challenges in this field are discussed.
Design/methodology/approach
First, the paper summarized the current research status of the hyperspectral techniques. Then, the paper demonstrated the development of underwater hyperspectral techniques from three major aspects, which are UHI preprocess, unmixing and applications. Finally, the paper presents a conclusion of applications of hyperspectral imaging and future research directions.
Findings
Various methods and scenarios for underwater object detection with hyperspectral imaging are compared, which include preprocessing, unmixing and classification. A summary is made to demonstrate the application scope and results of different methods, which may play an important role in the application of underwater hyperspectral object detection in the future.
Originality/value
This paper introduced several methods of hyperspectral image process, give out the conclusion of the advantages and disadvantages of each method, then demonstrated the challenges we face and the possible way to deal with them.
Details