Search results
1 – 10 of over 47000Product configurator is a sales and production‐planning tool that helps to transform customer requirements into bills‐of‐materials, lists of features and cost estimations. The…
Abstract
Purpose
Product configurator is a sales and production‐planning tool that helps to transform customer requirements into bills‐of‐materials, lists of features and cost estimations. The purpose of this paper is to introduce a method of how to analyse sales configuration models by using a design structure matrix (DSM) tool. By applying the DSM techniques, the sales configuration managers may sequence the product configuration questions and organize the connection to production.
Design/methodology/approach
First, the paper explains a sales configuration system structure from published academic and non‐academic works. These sources employ both theoretical and practical views on the topic of computer‐based sales expert systems. Second, the paper demonstrates an application of using DSM for configuration modelling.
Findings
The current sales configuration approaches include constraint‐based, rules‐based, and object‐oriented approaches. Product description methods vary, but the general problem remains the same: the configuration process should be designed in such a way that customer selections do not affect the previous selections. From the user point of view, answering the questions should be smooth and fast. In turn this will lead to the growing importance of building more effective product configuration models. DSM offers a systematic way to organise customer interface in sales configuration systems.
Research limitations/implications
This paper analyses how DSM could help in planning product configuration modelling. Comparison of different sequences is presented. The examples used are hypothetical, but illustrate the suitability of DSM analysis. Companies are trying to establish easily configured product models, which are fast, flexible and cost‐effective for adjustments and modifications. Use of DSM may help in the roll‐out of sales configuration projects. DSM may also be used as a quick view to represent the complexity of product configurability. The future needs for configuration tools will be focused towards product model management from the technical limitations of different data storage approaches.
Practical implications
Configurator software creates product variants, which are logical descriptions of physical products. Variants have parameters which describe the customer‐made selections. The parameter selections may have interconnections between the choices. Some selections may affect further selections and some combinations may not be allowed for incompatibility, cost or safety reasons. There are several commercial software packages available for creating product configurations. Product description methods vary, but the general problem remains the same: the configuration process should be designed in such a way that customer selections do not affect the previous selections. Answering the questions should be smooth and fast. Configuration of complex products, for instance, airplanes, may include several sub‐systems and have various loops within the quotation process. The use of DSM may help in the roll‐out of sales configuration projects. DSM may also be used as a quick view to represent the complexity of product configurability.
Originality/value
The paper helps both researchers and practitioners to obtain a clearer view on the development of sales configuration systems and the potential of systematic DSM‐based product model analysis.
Details
Keywords
The purpose of this paper is to enhance the performance of spammer identification problem in online social networks. Hyperparameter tuning has been performed by researchers in the…
Abstract
Purpose
The purpose of this paper is to enhance the performance of spammer identification problem in online social networks. Hyperparameter tuning has been performed by researchers in the past to enhance the performance of classifiers. The AdaBoost algorithm belongs to a class of ensemble classifiers and is widely applied in binary classification problems. A single algorithm may not yield accurate results. However, an ensemble of classifiers built from multiple models has been successfully applied to solve many classification tasks. The search space to find an optimal set of parametric values is vast and so enumerating all possible combinations is not feasible. Hence, a hybrid modified whale optimization algorithm for spam profile detection (MWOA-SPD) model is proposed to find optimal values for these parameters.
Design/methodology/approach
In this work, the hyperparameters of AdaBoost are fine-tuned to find its application to identify spammers in social networks. AdaBoost algorithm linearly combines several weak classifiers to produce a stronger one. The proposed MWOA-SPD model hybridizes the whale optimization algorithm and salp swarm algorithm.
Findings
The technique is applied to a manually constructed Twitter data set. It is compared with the existing optimization and hyperparameter tuning methods. The results indicate that the proposed method outperforms the existing techniques in terms of accuracy and computational efficiency.
Originality/value
The proposed method reduces the server load by excluding complex features retaining only the lightweight features. It aids in identifying the spammers at an earlier stage thereby offering users a propitious environment.
Details
Keywords
Kin Fun Li, Yali Wang and Wei Yu
Purpose — To develop methodologies to evaluate search engines according to an individual's preference in an easy and reliable manner, and to formulate user-oriented metrics to…
Abstract
Purpose — To develop methodologies to evaluate search engines according to an individual's preference in an easy and reliable manner, and to formulate user-oriented metrics to compare freshness and duplication in search results.
Design/methodology/approach — A personalised evaluation model for comparing search engines is designed as a hierarchy of weighted parameters. These commonly found search engine features and performance measures are given quantitative and qualitative ratings by an individual user. Furthermore, three performance measurement metrics are formulated and presented as histograms for visual inspection. A methodology is introduced to quantitatively compare and recognise the different histogram patterns within the context of search engine performance.
Findings — Precision and recall are the fundamental measures used in many search engine evaluations due to their simplicity, fairness and reliability. Most recent evaluation models are user oriented and focus on relevance issues. Identifiable statistical patterns are found in performance measures of search engines.
Research limitations/implications — The specific parameters used in the evaluation model could be further refined. A larger scale user study would confirm the validity and usefulness of the model. The three performance measures presented give a reasonably informative overview of the characteristics of a search engine. However, additional performance parameters and their resulting statistical patterns would make the methodology more valuable to the users.
Practical implications — The easy-to-use personalised search engine evaluation model can be tailored to an individual's preference and needs simply by changing the weights and modifying the features considered. A user is able to get an idea of the characteristics of a search engine quickly using the quantitative measure of histogram patterns that represent the search performance metrics introduced.
Originality/value — The presented work is considered original as one of the first search engine evaluation models that can be personalised. This enables a Web searcher to choose an appropriate search engine for his/her needs and hence finding the right information in the shortest time with the least effort.
Details
Keywords
Wenda Wei, Chengxia Liu and Jianing Wang
Nowadays, most methods of illusion garment evaluation are based on the subjective evaluation of experienced practitioners, which consumes time and the results are too subjective…
Abstract
Purpose
Nowadays, most methods of illusion garment evaluation are based on the subjective evaluation of experienced practitioners, which consumes time and the results are too subjective to be accurate enough. It is necessary to explore a method that can quantify professional experience into objective indicators to evaluate the sensory comfort of the optical illusion skirt quickly and accurately. The purpose of this paper is to propose a method to objectively evaluate the sensory comfort of optical illusion skirt patterns by combining texture feature extraction and prediction model construction.
Design/methodology/approach
Firstly, 10 optical illusion sample skirts are produced, and 10 experimental images are collected for each sample skirt. Then a Likert five-level evaluation scale is designed to obtain the sensory comfort level of each skirt through the questionnaire survey. Synchronously, the coarseness, contrast, directionality, line-likeness, regularity and roughness of the sample image are calculated based on Tamura texture feature algorithm, and the mean, contrast and entropy are extracted of the image transformed by Gabor wavelet. Both are set as objective parameters. Two final indicators T1 and T2 are refined from the objective parameters previously obtained to construct the predictive model of the subjective comfort of the visual illusion skirt. The linear regression model and the MLP neural network model are constructed.
Findings
Results show that the accuracy of the linear regression model is 92%, and prediction accuracy of the MLP neural network model is 97.9%. It is feasible to use Tamura texture features, Gabor wavelet transform and MLP neural network methods to objectively predict the sensory comfort of visual illusion skirt images.
Originality/value
Compared with the existing uncertain and non-reproducible subjective evaluation of optical illusion clothing based on experienced experts. The main advantage of the authors' method is that this method can objectively obtain evaluation parameters, quickly and accurately obtain evaluation grades without repeated evaluation by experienced experts. It is a method of objectively quantifying the experience of experts.
Details
Keywords
Ning Zhang, Ruru Pan, Lei Wang, Shanshan Wang, Jun Xiang and Weidong Gao
The purpose of this paper is to propose a novel method using support vector machine (SVM) classifiers for objective seam pucker evaluation. Features are extracted using wavelet…
Abstract
Purpose
The purpose of this paper is to propose a novel method using support vector machine (SVM) classifiers for objective seam pucker evaluation. Features are extracted using wavelet analysis and gray-level co-occurrence matrix (GLCM), and the samples are evaluated using SVM classifiers. The study aims to solve the problem of inappropriate parameters and large required samples in objective seam pucker evaluation.
Design/methodology/approach
Initially, seam pucker image was captured, and Edge detection and Hough transform were utilized to normalize the seam position and orientation. After cropping the image, the intensity was adjusted to the same identical level through histogram specification. Then, the standard deviations of the horizontal image and diagonal image, reconstructed using wavelet decomposition and reconstruction, were calculated based on parameter optimization. Meanwhile, GLCM was extracted from the restructured horizontal detail image, then the contrast and correlation of GLCM were calculated. Finally, these four features were imported to SVM classifiers based on genetic algorithm for evaluation.
Findings
The four extracted features reflected linear relationships among five grades. The experimental results showed that the classification accuracy was 96 percent, which catches up to the performance of human vision, and resolves ambiguity and subjective of the manual evaluation.
Originality/value
There are large required samples in current research. This paper provides a novel method using finite samples, and the parameters of the methods were discussed for parameter optimization. The evaluation results can provide references for analyzing the reason of wrinkles during garment manufacturing.
Details
Keywords
Hu Qiao, Rong Mo and Ying Xiang
The purpose of this paper is to establish an adaptive assembly, to realize the adaptive changing of the models and to improve the flexibility and reliability of assembly change…
Abstract
Purpose
The purpose of this paper is to establish an adaptive assembly, to realize the adaptive changing of the models and to improve the flexibility and reliability of assembly change. For a three-dimensional (3D) computer-aided design (CAD) assembly in a changing process, there are two practical problems. One is delivering parameters’ information not smoothly. The other one is to easily destroy an assembly structure.
Design/methodology/approach
The paper establishes associated parameters design structure matrix of related parts, and predicts possible propagation paths of the parameters. Based on the predicted path, structured storage is made for the affected parameters, tolerance range and the calculation relations. The study combines structured path information and all constrained assemblies to build the adaptive assembly, proposes an adaptive change algorithm for assembly changing and discusses the extendibility of the adaptive assembly.
Findings
The approach would improve the flexibility and reliability of assembly change and be applied to different CAD platform.
Practical implications
The examples illustrate the construction and adaptive behavior of the assembly and verify the feasibility and reasonability of the adaptive assembly in practical application.
Originality/value
The adaptive assembly model proposed in the paper is an original method to assembly change. And compared with other methods, good results have been obtained.
Details
Keywords
Sharanabasappa and Suvarna Nandyal
In order to prevent accidents during driving, driver drowsiness detection systems have become a hot topic for researchers. There are various types of features that can be used to…
Abstract
Purpose
In order to prevent accidents during driving, driver drowsiness detection systems have become a hot topic for researchers. There are various types of features that can be used to detect drowsiness. Detection can be done by utilizing behavioral data, physiological measurements and vehicle-based data. The existing deep convolutional neural network (CNN) models-based ensemble approach analyzed the behavioral data comprises eye or face or head movement captured by using a camera images or videos. However, the developed model suffered from the limitation of high computational cost because of the application of approximately 140 million parameters.
Design/methodology/approach
The proposed model uses significant feature parameters from the feature extraction process such as ReliefF, Infinite, Correlation, Term Variance are used for feature selection. The features that are selected are undergone for classification using ensemble classifier.
Findings
The output of these models is classified into non-drowsiness or drowsiness categories.
Research limitations/implications
In this research work higher end camera are required to collect videos as it is cost-effective. Therefore, researches are encouraged to use the existing datasets.
Practical implications
This paper overcomes the earlier approach. The developed model used complex deep learning models on small dataset which would also extract additional features, thereby provided a more satisfying result.
Originality/value
Drowsiness can be detected at the earliest using ensemble model which restricts the number of accidents.
Details
Keywords
Chuyu Tang, Hao Wang, Genliang Chen and Shaoqiu Xu
This paper aims to propose a robust method for non-rigid point set registration, using the Gaussian mixture model and accommodating non-rigid transformations. The posterior…
Abstract
Purpose
This paper aims to propose a robust method for non-rigid point set registration, using the Gaussian mixture model and accommodating non-rigid transformations. The posterior probabilities of the mixture model are determined through the proposed integrated feature divergence.
Design/methodology/approach
The method involves an alternating two-step framework, comprising correspondence estimation and subsequent transformation updating. For correspondence estimation, integrated feature divergences including both global and local features, are coupled with deterministic annealing to address the non-convexity problem of registration. For transformation updating, the expectation-maximization iteration scheme is introduced to iteratively refine correspondence and transformation estimation until convergence.
Findings
The experiments confirm that the proposed registration approach exhibits remarkable robustness on deformation, noise, outliers and occlusion for both 2D and 3D point clouds. Furthermore, the proposed method outperforms existing analogous algorithms in terms of time complexity. Application of stabilizing and securing intermodal containers loaded on ships is performed. The results demonstrate that the proposed registration framework exhibits excellent adaptability for real-scan point clouds, and achieves comparatively superior alignments in a shorter time.
Originality/value
The integrated feature divergence, involving both global and local information of points, is proven to be an effective indicator for measuring the reliability of point correspondences. This inclusion prevents premature convergence, resulting in more robust registration results for our proposed method. Simultaneously, the total operating time is reduced due to a lower number of iterations.
Details
Keywords
Jia Yan, Shukai Duan, Tingwen Huang and Lidan Wang
The purpose of this paper is to improve the performance of E-nose in the detection of wound infection. Feature extraction and selection methods have a strong impact on the…
Abstract
Purpose
The purpose of this paper is to improve the performance of E-nose in the detection of wound infection. Feature extraction and selection methods have a strong impact on the performance of pattern classification of electronic nose (E-nose). A new hybrid feature matrix construction method and multi-objective binary quantum-behaved particle swarm optimization (BQPSO) have been proposed for feature extraction and selection of sensor array.
Design/methodology/approach
A hybrid feature matrix constructed by maximum value and wavelet coefficients is proposed to realize feature extraction. Multi-objective BQPSO whose fitness function contains classification accuracy and a number of selected sensors is used for feature selection. Quantum-behaved particle swarm optimization (QPSO) is used for synchronization optimization of selected features and parameter of classifier. Radical basis function (RBF) network is used for classification.
Findings
E-nose obtains the highest classification accuracy when the maximum value and db 5 wavelet coefficients are extracted as the hybrid features and only six sensors are selected for classification. All results make it clear that the proposed method is an ideal feature extraction and selection method of E-nose in the detection of wound infection.
Originality/value
The innovative concept improves the performance of E-nose in wound monitoring, and is beneficial for realizing the clinical application of E-nose.
Details
Keywords
Armin Mahmoodi, Leila Hashemi and Milad Jasemi
In this study, the central objective is to foresee stock market signals with the use of a proper structure to achieve the highest accuracy possible. For this purpose, three hybrid…
Abstract
Purpose
In this study, the central objective is to foresee stock market signals with the use of a proper structure to achieve the highest accuracy possible. For this purpose, three hybrid models have been developed for the stock markets which are a combination of support vector machine (SVM) with meta-heuristic algorithms of particle swarm optimization (PSO), imperialist competition algorithm (ICA) and genetic algorithm (GA).All the analyses are technical and are based on the Japanese candlestick model.
Design/methodology/approach
Further as per the results achieved, the most suitable algorithm is chosen to anticipate sell and buy signals. Moreover, the authors have compared the results of the designed model validations in this study with basic models in three articles conducted in the past years. Therefore, SVM is examined by PSO. It is used as a classification agent to search the problem-solving space precisely and at a faster pace. With regards to the second model, SVM and ICA are tested to stock market timing, in a way that ICA is used as an optimization agent for the SVM parameters. At last, in the third model, SVM and GA are studied, where GA acts as an optimizer and feature selection agent.
Findings
As per the results, it is observed that all new models can predict accurately for only 6 days; however, in comparison with the confusion matrix results, it is observed that the SVM-GA and SVM-ICA models have correctly predicted more sell signals, and the SCM-PSO model has correctly predicted more buy signals. However, SVM-ICA has shown better performance than other models considering executing the implemented models.
Research limitations/implications
In this study, the data for stock market of the years 2013–2021 were analyzed; the long length of timeframe makes the input data analysis challenging as they must be moderated with respect to the conditions where they have been changed.
Originality/value
In this study, two methods have been developed in a candlestick model; they are raw-based and signal-based approaches in which the hit rate is determined by the percentage of correct evaluations of the stock market for a 16-day period.
Details