Search results

1 – 10 of over 5000
Article
Publication date: 1 June 2003

Jaroslav Mackerle

This paper gives a bibliographical review of the finite element and boundary element parallel processing techniques from the theoretical and application points of view. Topics…

1203

Abstract

This paper gives a bibliographical review of the finite element and boundary element parallel processing techniques from the theoretical and application points of view. Topics include: theory – domain decomposition/partitioning, load balancing, parallel solvers/algorithms, parallel mesh generation, adaptive methods, and visualization/graphics; applications – structural mechanics problems, dynamic problems, material/geometrical non‐linear problems, contact problems, fracture mechanics, field problems, coupled problems, sensitivity and optimization, and other problems; hardware and software environments – hardware environments, programming techniques, and software development and presentations. The bibliography at the end of this paper contains 850 references to papers, conference proceedings and theses/dissertations dealing with presented subjects that were published between 1996 and 2002.

Details

Engineering Computations, vol. 20 no. 4
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 1 December 1997

Dan Givoli, Joseph E. Flaherty and Mark S. Shephard

Describes a new finite element scheme for the large‐scale analysis of compressible and incompressible viscous flows. The scheme is based on a combined compressible‐ incompressible…

Abstract

Describes a new finite element scheme for the large‐scale analysis of compressible and incompressible viscous flows. The scheme is based on a combined compressible‐ incompressible Galerkin least‐squares (GLS) space‐time variational formulation. Three‐ dimensional unstructured meshes are employed, with piecewise‐constant temporal interpolation, local time‐stepping for steady flows, and linear continuous spatial interpolation in all the variables. The scheme incorporates automatic adaptive mesh refinement, with a choice of various error indicators. It is implemented on a distributed‐memory parallel computer, and includes an automatic load‐balancing procedure. Demonstrates the ability to solve both compressible and incompressible viscous flow problems using the parallel adaptive framework via numerical examples. These include Mach 3 flow over a flat plate, and a divergence‐free buoyancy‐driven flow in a cavity. The latter is a model for the steady melt flow in a Czochralski crystal growth process.

Details

International Journal of Numerical Methods for Heat & Fluid Flow, vol. 7 no. 8
Type: Research Article
ISSN: 0961-5539

Keywords

Article
Publication date: 1 December 2001

Jaroslav Mackerle

Gives a bibliographical review of the finite element meshing and remeshing from the theoretical as well as practical points of view. Topics such as adaptive techniques for meshing…

1895

Abstract

Gives a bibliographical review of the finite element meshing and remeshing from the theoretical as well as practical points of view. Topics such as adaptive techniques for meshing and remeshing, parallel processing in the finite element modelling, etc. are also included. The bibliography at the end of this paper contains 1,727 references to papers, conference proceedings and theses/dissertations dealing with presented subjects that were published between 1990 and 2001.

Details

Engineering Computations, vol. 18 no. 8
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 1 August 2001

Jaroslav Mackerle

Gives a bibliographical review of the error estimates and adaptive finite element methods from the theoretical as well as the application point of view. The bibliography at the…

1667

Abstract

Gives a bibliographical review of the error estimates and adaptive finite element methods from the theoretical as well as the application point of view. The bibliography at the end contains 2,177 references to papers, conference proceedings and theses/dissertations dealing with the subjects that were published in 1990‐2000.

Details

Engineering Computations, vol. 18 no. 5/6
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 14 March 2019

Lin Fu, Zhe Ji, Xiangyu Y. Hu and Nikolaus A. Adams

This paper aims to develop a parallel fast neighbor search method and communication strategy for particle-based methods with adaptive smoothing-length on distributed-memory…

Abstract

Purpose

This paper aims to develop a parallel fast neighbor search method and communication strategy for particle-based methods with adaptive smoothing-length on distributed-memory computing systems.

Design/methodology/approach

With a multi-resolution-based hierarchical data structure, the parallel neighbor search method is developed to detect and construct ghost buffer particles, i.e. neighboring particles on remote processor nodes. To migrate ghost buffer particles among processor nodes, an undirected graph is established to characterize the sparse data communication relation and is dynamically recomposed. By the introduction of an edge coloring algorithm from graph theory, the complex sparse data exchange can be accomplished within optimized frequency. For each communication substep, only efficient nonblocking point-to-point communication is involved.

Findings

Two demonstration scenarios are considered: fluid dynamics based on smoothed-particle hydrodynamics with adaptive smoothing-length and a recently proposed physics-motivated partitioning method [Fu et al., JCP 341 (2017): 447-473]. Several new concepts are introduced to recast the partitioning method into a parallel version. A set of numerical experiments is conducted to demonstrate the performance and potential of the proposed parallel algorithms.

Originality/value

The proposed methods are simple to implement in large-scale parallel environment and can handle particle simulations with arbitrarily varying smoothing-lengths. The implemented smoothed-particle hydrodynamics solver has good parallel performance, suggesting the potential for other scientific applications.

Article
Publication date: 16 April 2018

Xuanhua Fan, Keying Wang and Shifu Xiao

As a practical engineering method, earthquake response spectra play an important role in seismic hazard assessment and in seismic design of structures. However, the computing…

Abstract

Purpose

As a practical engineering method, earthquake response spectra play an important role in seismic hazard assessment and in seismic design of structures. However, the computing scale and the efficiency of commercial software restricted the solution of complex structures. There is a clear need of developing large-scale and highly efficient finite element procedures for response spectrum analysis.

Design/methodology/approach

In this paper, the kernel theories for earthquake response spectra are deduced and the corresponding parallel solution flow via the modal superposition method is presented. Based on the algorithm and the parallel data structure of JAUMIN framework, a parallel finite element (FE) solution module is established. Using the solution procedure on a supercomputer equipped with up to thousands of processors, the correctness and parallel scalability of the algorithm are evaluated via numerical experiments of typical engineering examples.

Findings

The results show that the solution module has the same precision as the commercial FE software ANSYS; the maximum solution scale achieves 154 million degrees of freedom (DOFs) with a favorable parallel computing efficiency, going far beyond the computing ability of the commercial FE software.

Originality/value

The solution scale in this paper is very challenging for the large-scale parallel computing of structural dynamics and will promote the dynamic analysis ability of complex facilities greatly.

Details

Engineering Computations, vol. 35 no. 2
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 3 July 2017

Alex A. Schmidt, Alice de Jesus Kozakevicius and Stefan Jakobsson

The current work aims to present a parallel code using the open multi-processing (OpenMP) programming model for an adaptive multi-resolution high-order finite difference scheme…

Abstract

Purpose

The current work aims to present a parallel code using the open multi-processing (OpenMP) programming model for an adaptive multi-resolution high-order finite difference scheme for solving 2D conservation laws, comparing efficiencies obtained with a previous message passing interface formulation for the same serial scheme and considering the same type of 2D formulations laws.

Design/methodology/approach

The serial version of the code is naturally suitable for parallelization because the spatial operator formulation is based on a splitting scheme per direction for which the flux components are numerically computed by a Lax–Friedrichs factorization independently for each row or column. High-order approximations for numerical fluxes are computed by the third-order essentially non-oscillatory (ENO) and fifth-order weighted essentially non-oscillatory (WENO) interpolation schemes, assuming sparse grids in each direction. The grid adaptivity is obtained by a cubic interpolating wavelet transform applied in each space dimension, associated to a threshold operator. Time is evolved by a third order TVD Runge–Kutta method.

Findings

The parallel formulation is implemented automatically at compiling time by the OpenMP library routines, being virtually transparent to the programmer. This over simplifies any concerns about managing and/or updating the adaptive grid when compared to what is necessary to be done when other parallel approaches are considered. Numerical simulations results and the large speedups obtained for the Euler equations in gas dynamics highlight the efficiency of the OpenMP approach.

Research limitations/implications

The resulting speedups reflect the effectiveness of the OpenMP approach but are, to a large extension, limited by the hardware used (2 E5-2620 Intel Xeon processors, 6 cores, 2 threads/core, hyper-threading enabled). As the demand for OpenMP threads increases, the code starts to make explicit use of the second logical thread available in each E5-2620 processor core and efficiency drops. The speedup peak is reached near the possible maximum (24) at about 22, 23 threads. This peak reflects the hardware configuration and the true software limit should be located way beyond this value.

Practical implications

So far no attempts have been made to parallelize other possible code segments (for instance, the ENO|-WENO-TVD code lines that process the different data components which could potentially push the speed up limit to higher values even further. The fact that the speedup peak is located close to the present hardware limit reflects the scalability properties of the OpenMP programming and of the splitting scheme as well. Consequently, it is likely that the speedup peak with the OpenMP approach for this kind of problem formulation will be close to the physical (and/or logical) limit of the hardware used.

Social implications

This work is the result of a successful collaboration among researchers from two different institutions, one internationally well-known and with a long-term experience in applied mathematics for industrial applications and the other in a starting process of international academic insertion. In this way, this scientific partnership has the potential of promoting further knowledge exchange, involving students and other collaborators.

Originality/value

The proposed methodology (use of OpenMP programming model for the wavelet adaptive splitting scheme) is original and contributes to a very active research area in the past years, namely, adaptive methods for conservation laws and their parallel formulations, which is of great interest for the entire scientific community.

Details

International Journal of Numerical Methods for Heat & Fluid Flow, vol. 27 no. 7
Type: Research Article
ISSN: 0961-5539

Keywords

Article
Publication date: 1 February 1996

Jaroslav Mackerle

Presents a review on implementing finite element methods on supercomputers, workstations and PCs and gives main trends in hardware and software developments. An appendix included…

Abstract

Presents a review on implementing finite element methods on supercomputers, workstations and PCs and gives main trends in hardware and software developments. An appendix included at the end of the paper presents a bibliography on the subjects retrospectively to 1985 and approximately 1,100 references are listed.

Details

Engineering Computations, vol. 13 no. 1
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 11 October 2019

Yaxin Peng, Naiwu Wen, Chaomin Shen, Xiaohuang Zhu and Shihui Ying

Partial alignment for 3 D point sets is a challenging problem for laser calibration and robot calibration due to the unbalance of data sets, especially when the overlap of data…

Abstract

Purpose

Partial alignment for 3 D point sets is a challenging problem for laser calibration and robot calibration due to the unbalance of data sets, especially when the overlap of data sets is low. Geometric features can promote the accuracy of alignment. However, the corresponding feature extraction methods are time consuming. The purpose of this paper is to find a framework for partial alignment by an adaptive trimmed strategy.

Design/methodology/approach

First, the authors propose an adaptive trimmed strategy based on point feature histograms (PFH) coding. Second, they obtain an initial transformation based on this partition, which improves the accuracy of the normal direction weighted trimmed iterative closest point (ICP) method. Third, they conduct a series of GPU parallel implementations for time efficiency.

Findings

The initial partition based on PFH feature improves the accuracy of the partial registration significantly. Moreover, the parallel GPU algorithms accelerate the alignment process.

Research limitations/implications

This study is applicable to rigid transformation so far. It could be extended to non-rigid transformation.

Practical implications

In practice, point set alignment for calibration is a technique widely used in the fields of aircraft assembly, industry examination, simultaneous localization and mapping and surgery navigation.

Social implications

Point set calibration is a building block in the field of intelligent manufacturing.

Originality/value

The contributions are as follows: first, the authors introduce a novel coarse alignment as an initial calibration by PFH descriptor similarity, which can be viewed as a coarse trimmed process by partitioning the data to the almost overlap part and the rest part; second, they reduce the computation time by GPU parallel coding during the acquisition of feature descriptor; finally, they use the weighted trimmed ICP method to refine the transformation.

Details

Assembly Automation, vol. 40 no. 2
Type: Research Article
ISSN: 0144-5154

Keywords

Article
Publication date: 12 April 2022

Sunil Budhiraja and Neerpal Rathi

The study aims to examine the association between learning culture and adaptive performance of delivery employees during crises situation. The study develops and tests a model…

Abstract

Purpose

The study aims to examine the association between learning culture and adaptive performance of delivery employees during crises situation. The study develops and tests a model that explains how learning culture, through change-efficacy and meaningful work, influences employees' adaptive performance (including how they handle crisis situations and deal with uncertainty).

Design/methodology/approach

Data was collected from 298 delivery employees working in e-commerce companies throughout India in a time-lagged manner. Regression analysis and structural equation modeling were performed to assess the influence of learning culture, change-efficacy and meaningful work on adaptive performance using SPSS 24. Further, PROCESS macro was used to test the parallel mediation effects through bootstrapping approach.

Findings

The study establishes a significant direct and indirect relationship between learning culture and adaptive performance for employees. Further, underpinning the transformative learning theory (Mezirow, 1997), and job characteristics theory (1976), this study came across two pathways for organizations to transform their learning efforts into improved adaptive performance for employees.

Practical implications

Organizations, particularly in crisis situations, can leverage employees' change-efficacy and meaningful work to connect learning efforts with employees' adaptive performance.

Originality/value

The study contributes significantly to existing theory on transformative learning and job characteristics theory while strengthening the literature on antecedents of employees' adaptive performance, particularly in crises situation.

Details

International Journal of Productivity and Performance Management, vol. 72 no. 8
Type: Research Article
ISSN: 1741-0401

Keywords

1 – 10 of over 5000