Search results
1 – 10 of over 1000Chengjun Chen, Zhongke Tian, Dongnian Li, Lieyong Pang, Tiannuo Wang and Jun Hong
This study aims to monitor and guide the assembly process. The operators need to change the assembly process according to the products’ specifications during manual assembly of…
Abstract
Purpose
This study aims to monitor and guide the assembly process. The operators need to change the assembly process according to the products’ specifications during manual assembly of mass customized production. Traditional information inquiry and display methods, such as manual lookup of assembly drawings or electronic manuals, are inefficient and error-prone.
Design/methodology/approach
This paper proposes a projection-based augmented reality system (PBARS) for assembly guidance and monitoring. The system includes a projection method based on viewpoint tracking, in which the position of the operator’s head is tracked and the projection images are changed correspondingly. The assembly monitoring phase applies a method for parts recognition. First, the pixel local binary pattern (PX-LBP) operator is achieved by merging the classical LBP operator with the pixel classification process. Afterward, the PX-LBP features of the depth images are extracted and the randomized decision forests classifier is used to get the pixel classification prediction image (PCPI). Parts recognition and assembly monitoring is performed by PCPI analysis.
Findings
The projection image changes with the viewpoint of the human body, hence the operators always perceive the three-dimensional guiding scene from different viewpoints, improving the human-computer interaction. Part recognition and assembly monitoring were achieved by comparing the PCPIs, in which missing and erroneous assembly can be detected online.
Originality/value
This paper designed the PBARS to monitor and guide the assembly process simultaneously, with potential applications in mass customized production. The parts recognition and assembly monitoring based on pixels classification provides a novel method for assembly monitoring.
Details
Keywords
Mohamed Hammami, Youssef Chahir and Liming Chen
Along with the ever growingWeb is the proliferation of objectionable content, such as sex, violence, racism, etc. We need efficient tools for classifying and filtering undesirable…
Abstract
Along with the ever growingWeb is the proliferation of objectionable content, such as sex, violence, racism, etc. We need efficient tools for classifying and filtering undesirable web content. In this paper, we investigate this problem through WebGuard, our automatic machine learning based pornographic website classification and filtering system. Facing the Internet more and more visual and multimedia as exemplified by pornographic websites, we focus here our attention on the use of skin color related visual content based analysis along with textual and structural content based analysis for improving pornographic website filtering. While the most commercial filtering products on the marketplace are mainly based on textual content‐based analysis such as indicative keywords detection or manually collected black list checking, the originality of our work resides on the addition of structural and visual content‐based analysis to the classical textual content‐based analysis along with several major‐data mining techniques for learning and classifying. Experimented on a testbed of 400 websites including 200 adult sites and 200 non pornographic ones, WebGuard, our Web filtering engine scored a 96.1% classification accuracy rate when only textual and structural content based analysis are used, and 97.4% classification accuracy rate when skin color related visual content based analysis is driven in addition. Further experiments on a black list of 12 311 adult websites manually collected and classified by the French Ministry of Education showed that WebGuard scored 87.82% classification accuracy rate when using only textual and structural content‐based analysis, and 95.62% classification accuracy rate when the visual content‐based analysis is driven in addition. The basic framework of WebGuard can apply to other categorization problems of websites which combine, as most of them do today, textual and visual content.
Details
Keywords
Jin Young Jung, Seonkoo Chee and InHwan Sul
Increasingly 3D printing is used for parts of garments or for making whole garments due to their flexibility and comfort and for functionalizing or enhancing the aesthetics of the…
Abstract
Purpose
Increasingly 3D printing is used for parts of garments or for making whole garments due to their flexibility and comfort and for functionalizing or enhancing the aesthetics of the final garment and hence adding value. Many of these applications rely on complex programming of the 3D printer and are usually provided by the vendor company. This paper introduces a simpler, easier platform for designing 3D-printed textiles, garments and other artifacts, by predicting the optimal orientation of the target objects to minimize the use of plastic filaments.
Design/methodology/approach
The main idea is based on the shadow-casting analogy, which assumes that the volume of the support structure is similar to that of the shadow from virtual sunlight. The triangular elements of the target object are converted into 3D pixels with integer-based normal vectors and real-numbered coordinates via vertically sparse voxelization. The pixels are classified into several groups and their noise is suppressed using a specially designed noise-filtering algorithm called slot pairing. The final support structure volume information was rendered as a two-dimensional (2D) figure, similar to a medical X-ray image. Thus, the authors named their method modified support structure tomography.
Findings
The study algorithm showed an error range of no more than 1.6% with exact volumes and 6.8% with slicing software. Moreover, the calculation time is only several minutes for tens of thousands of mesh triangles. The algorithm was verified for several meshes, including the cone, sphere, Stanford bunny and human manikin.
Originality/value
Simple hardware, such as a CPU, embedded system, Arduino or Raspberry Pi, can be used. This requires much less computational resources compared with the conventional g-code generation. Also, the global and local support structure is represented both quantitatively and graphically via tomographs.
Details
Keywords
Yongxiang Wu, Yili Fu and Shuguo Wang
This paper aims to use fully convolutional network (FCN) to predict pixel-wise antipodal grasp affordances for unknown objects and improve the grasp detection performance through…
Abstract
Purpose
This paper aims to use fully convolutional network (FCN) to predict pixel-wise antipodal grasp affordances for unknown objects and improve the grasp detection performance through multi-scale feature fusion.
Design/methodology/approach
A modified FCN network is used as the backbone to extract pixel-wise features from the input image, which are further fused with multi-scale context information gathered by a three-level pyramid pooling module to make more robust predictions. Based on the proposed unify feature embedding framework, two head networks are designed to implement different grasp rotation prediction strategies (regression and classification), and their performances are evaluated and compared with a defined point metric. The regression network is further extended to predict the grasp rectangles for comparisons with previous methods and real-world robotic grasping of unknown objects.
Findings
The ablation study of the pyramid pooling module shows that the multi-scale information fusion significantly improves the model performance. The regression approach outperforms the classification approach based on same feature embedding framework on two data sets. The regression network achieves a state-of-the-art accuracy (up to 98.9%) and speed (4 ms per image) and high success rate (97% for household objects, 94.4% for adversarial objects and 95.3% for objects in clutter) in the unknown object grasping experiment.
Originality/value
A novel pixel-wise grasp affordance prediction network based on multi-scale feature fusion is proposed to improve the grasp detection performance. Two prediction approaches are formulated and compared based on the proposed framework. The proposed method achieves excellent performances on three benchmark data sets and real-world robotic grasping experiment.
Details
Keywords
Shahidha Banu S. and Maheswari N.
Background modelling has played an imperative role in the moving object detection as the progress of foreground extraction during video analysis and surveillance in many real-time…
Abstract
Purpose
Background modelling has played an imperative role in the moving object detection as the progress of foreground extraction during video analysis and surveillance in many real-time applications. It is usually done by background subtraction. This method is uprightly based on a mathematical model with a fixed feature as a static background, where the background image is fixed with the foreground object running over it. Usually, this image is taken as the background model and is compared against every new frame of the input video sequence. In this paper, the authors presented a renewed background modelling method for foreground segmentation. The principal objective of the work is to perform the foreground object detection only in the premeditated region of interest (ROI). The ROI is calculated using the proposed algorithm reducing and raising by half (RRH). In this algorithm, the coordinate of a circle with the frame width as the diameter is considered for traversal to find the pixel difference. The change in the pixel intensity is considered to be the foreground object and the position of it is determined based on the pixel location. Most of the techniques study their updates to the pixels of the complete frame which may result in increased false rate; The proposed system deals these flaw by controlling the ROI object (the region only where the background subtraction is performed) and thus extracts a correct foreground by exactly categorizes the pixel as the foreground and mines the precise foreground object. The broad experimental results and the evaluation parameters of the proposed approach with the state of art methods were compared against the most recent background subtraction approaches. Moreover, the efficiency of the authors’ method is analyzed in different situations to prove that this method is available for real-time videos as well as videos available in the 2014 challenge change detection data set.
Design/methodology/approach
In this paper, the authors presented a fresh background modelling method for foreground segmentation. The main objective of the work is to perform the foreground object detection only on the premeditated ROI. The region for foreground extraction is calculated using proposed RRH algorithm. Most of the techniques study their updates to the pixels of the complete frame which may result in increased false rate; most challenging case is that, the slow moving object is updated quickly to detect the foreground region. The anticipated system deals these flaw by controlling the ROI object (the region only where the background subtraction is performed) and thus extracts a correct foreground by exactly categorizing the pixel as the foreground and mining the precise foreground object.
Findings
Plum Analytics provide a new conduit for documenting and contextualizing the public impact and reach of research within digitally networked environments. While limitations are notable, the metrics promoted through the platform can be used to build a more comprehensive view of research impact.
Originality/value
The algorithm used in the work was proposed by the authors and are used for experimental evaluations.
Details
Keywords
Kuniaki Kawabata, Kanako Saitoh, Mutsunori Takahashi, Hajime Asama, Taketoshi Mishima, Mitsuaki Sugahara and Masashi Miyano
The purpose of this paper is to present classification schemes for the crystallization state of proteins utilizing image processing.
Abstract
Purpose
The purpose of this paper is to present classification schemes for the crystallization state of proteins utilizing image processing.
Design/methodology/approach
Two classification schemes shown here are combined sequentially.
Findings
The correct ratio of experimental result using the method presented here is approximately 70 per cent.
Originality/value
The paper is a contribution to automated evaluation crystal growth, combining two classifiers based on specific visual feature, sequentially.
Details
Keywords
Zhixin Wang, Peng Xu, Bohan Liu, Yankun Cao, Zhi Liu and Zhaojun Liu
This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some…
Abstract
Purpose
This paper aims to demonstrate the principle and practical applications of hyperspectral object detection, carry out the problem we now face and the possible solution. Also some challenges in this field are discussed.
Design/methodology/approach
First, the paper summarized the current research status of the hyperspectral techniques. Then, the paper demonstrated the development of underwater hyperspectral techniques from three major aspects, which are UHI preprocess, unmixing and applications. Finally, the paper presents a conclusion of applications of hyperspectral imaging and future research directions.
Findings
Various methods and scenarios for underwater object detection with hyperspectral imaging are compared, which include preprocessing, unmixing and classification. A summary is made to demonstrate the application scope and results of different methods, which may play an important role in the application of underwater hyperspectral object detection in the future.
Originality/value
This paper introduced several methods of hyperspectral image process, give out the conclusion of the advantages and disadvantages of each method, then demonstrated the challenges we face and the possible way to deal with them.
Details
Keywords
R.S. Vignesh and M. Monica Subashini
An abundance of techniques has been presented so forth for waste classification but, they deliver inefficient results with low accuracy. Their achievement on various repositories…
Abstract
Purpose
An abundance of techniques has been presented so forth for waste classification but, they deliver inefficient results with low accuracy. Their achievement on various repositories is different and also, there is insufficiency of high-scale databases for training. The purpose of the study is to provide high security.
Design/methodology/approach
In this research, optimization-assisted federated learning (FL) is introduced for thermoplastic waste segregation and classification. The deep learning (DL) network trained by Archimedes Henry gas solubility optimization (AHGSO) is used for the classification of plastic and resin types. The deep quantum neural networks (DQNN) is used for first-level classification and the deep max-out network (DMN) is employed for second-level classification. This developed AHGSO is obtained by blending the features of Archimedes optimization algorithm (AOA) and Henry gas solubility optimization (HGSO). The entities included in this approach are nodes and servers. Local training is carried out depending on local data and updations to the server are performed. Then, the model is aggregated at the server. Thereafter, each node downloads the global model and the update training is executed depending on the downloaded global and the local model till it achieves the satisfied condition. Finally, local update and aggregation at the server is altered based on the average method. The Data tag suite (DATS_2022) dataset is used for multilevel thermoplastic waste segregation and classification.
Findings
By using the DQNN in first-level classification the designed optimization-assisted FL has gained an accuracy of 0.930, mean average precision (MAP) of 0.933, false positive rate (FPR) of 0.213, loss function of 0.211, mean square error (MSE) of 0.328 and root mean square error (RMSE) of 0.572. In the second level classification, by using DMN the accuracy, MAP, FPR, loss function, MSE and RMSE are 0.932, 0.935, 0.093, 0.068, 0.303 and 0.551.
Originality/value
The multilevel thermoplastic waste segregation and classification using the proposed model is accurate and improves the effectiveness of the classification.
Details
Keywords
A. Reyana, Sandeep Kautish, A.S. Vibith and S.B. Goyal
In the traffic monitoring system, the detection of stirring vehicles is monitored by fitting static cameras in the traffic scenarios. Background subtraction a commonly used method…
Abstract
Purpose
In the traffic monitoring system, the detection of stirring vehicles is monitored by fitting static cameras in the traffic scenarios. Background subtraction a commonly used method detaches poignant objects in the foreground from the background. The method applies a Gaussian Mixture Model, which can effortlessly be contaminated through slow-moving or momentarily stopped vehicles.
Design/methodology/approach
This paper proposes the Enhanced Gaussian Mixture Model to overcome the addressed issue, efficiently detecting vehicles in complex traffic scenarios.
Findings
The model was evaluated with experiments conducted using real-world on-road travel videos. The evidence intimates that the proposed model excels with other approaches showing the accuracy of 0.9759 when compared with the existing Gaussian mixture model (GMM) model and avoids contamination of slow-moving or momentarily stopped vehicles.
Originality/value
The proposed method effectively combines, tracks and classifies the traffic vehicles, resolving the contamination problem that occurred by slow-moving or momentarily stopped vehicles.
Details
Keywords
Yuanyuan Chen, Xiufeng He, Jia Xu, Lin Guo, Yanyan Lu and Rongchun Zhang
As one of the world's most productive ecosystems, ecological land plays an important role in regional and global environments. Utilizing advanced optical and synthetic aperture…
Abstract
Purpose
As one of the world's most productive ecosystems, ecological land plays an important role in regional and global environments. Utilizing advanced optical and synthetic aperture radar (SAR) data for land cover/land use research becomes increasingly popular. This research aims to investigate the complementarity of fully polarimetric SAR and optical imaging for ecological land classification in the eastern coastal area of China.
Design/methodology/approach
Four polarimetric decomposition methods, namely, H/Alpha, Yamaguchi3, VanZyl3 and Krogager, were applied to Advanced Land Observing Satellite (ALOS) SAR image for scattering parameter extraction. These parameters were merged with ALOS optical parameters for subsequent classification using the object-based quick, unbiased, efficient statistical tree decision tree method.
Findings
The experimental results indicate that an improved classification performance was obtained in the decision level when merging the two data sources. In fact, unlike classification using only optical images, the proposed approach allowed to distinguish ecological land with similar spectrum but different scattering. Moreover, unlike classification using only polarimetric information, the integration of polarimetric and optical data allows to accurately distinguish reed from artemisia and sand from salt field and therefore achieve a detailed classification of the coastal area characteristics.
Originality/value
This research proposed an integrated classification method for coastal ecological land with polarimetric SAR and optical data. The object-based and decision-level fusion enables effective ecological land classification in coastal area was verified.
Details