Search results

1 – 10 of over 4000
Article
Publication date: 1 January 1990

Edward T. Lee

Average turnaround time is an important consideration in selecting or evaluating computing systems. The greatest lower bound and the least upper bound of average turnaround time…

Abstract

Average turnaround time is an important consideration in selecting or evaluating computing systems. The greatest lower bound and the least upper bound of average turnaround time are investigated. New results are presented with illustrative examples.

Details

Kybernetes, vol. 19 no. 1
Type: Research Article
ISSN: 0368-492X

Keywords

Article
Publication date: 22 June 2010

Imam Machdi, Toshiyuki Amagasa and Hiroyuki Kitagawa

The purpose of this paper is to propose general parallelism techniques for holistic twig join algorithms to process queries against Extensible Markup Language (XML) databases on a…

Abstract

Purpose

The purpose of this paper is to propose general parallelism techniques for holistic twig join algorithms to process queries against Extensible Markup Language (XML) databases on a multi‐core system.

Design/methodology/approach

The parallelism techniques comprised data and task parallelism. As for data parallelism, the paper adopted the stream‐based partitioning for XML to partition XML data as the basis of parallelism on multiple CPU cores. The XML data partitioning was performed in two levels. The first level was to create buckets for creating data independence and balancing loads among CPU cores; each bucket was assigned onto a CPU core. Within each bucket, the second level of XML data partitioning was performed to create finer partitions for providing finer parallelism. Each CPU core performed the holistic twig join algorithm on each finer partition of its own in parallel with other CPU cores. In task parallelism, the holistic twig join algorithm was decomposed into two main tasks, which were pipelined to create parallelism. The first task adopted the data parallelism technique and their outputs were transferred to the second task periodically. Since data transfers incurred overheads, the size of each data transfer needed to be estimated cautiously for achieving optimal performance.

Findings

The data and task parallelism techniques contribute to good performance especially for queries having complex structures and/or higher values of query selectivity. The performance of data parallelism can be further improved by task parallelism. Significant performance improvement is attained by queries having higher selectivity because more outputs computed by the second task is performed in parallel with the first task.

Research limitations/implications

The proposed parallelism techniques primarily deals with executing a single long‐running query for intra‐query parallelism, partitioning XML data on‐the‐fly, and allocating partitions on CPU cores statically. During the parallel execution, presumably there are no such dynamic XML data updates.

Practical implications

The effectiveness of the proposed parallel holistic twig joins relies fundamentally on some system parameter values that can be obtained from a benchmark of the system platform.

Originality/value

The paper proposes novel techniques to increase parallelism by combining techniques of data and task parallelism for achieving high performance. To the best of the author's knowledge, this is the first paper of parallelizing the holistic twig join algorithms on a multi‐core system.

Details

International Journal of Web Information Systems, vol. 6 no. 2
Type: Research Article
ISSN: 1744-0084

Keywords

Article
Publication date: 4 January 2021

Carlos Enrique Torres-Aguilar, Pedro Moreno-Bernal, Jesús Xamán, Ivett Zavala Guillen and Irving Osiris Hernández-López

This paper aims to present an evolutionary algorithm (EA) to accelerate the convergence for the radiative transfer equation (RTE) numerical solution using high-order and…

128

Abstract

Purpose

This paper aims to present an evolutionary algorithm (EA) to accelerate the convergence for the radiative transfer equation (RTE) numerical solution using high-order and high-resolution schemes by the relaxation coefficients optimization.

Design methodology/approach

The objective function minimizes the residual value difference between iterations in each control volume until its difference is lower than the convergence criterion. The EA approach is evaluated in two configurations, a two-dimensional cavity with scattering media and absorbing media.

Findings

Experimental results show the capacity to obtain the numerical solution for both cases on all interpolation schemes tested by the EA approach. The EA approach reduces CPU time for the RTE numerical solution using SUPERBEE, SWEBY and MUSCL schemes until 97% and 135% in scattering and absorbing media cases, respectively. The relaxation coefficients optimized every two numerical solution iterations achieve a significant reduction of the CPU time compared to the deferred correction procedure with fixed relaxation coefficients.

Originality/value

The proposed EA approach for the RTE numerical solution effectively reduces the CPU time compared to the DC procedure with fixed relaxation coefficients.

Details

Engineering Computations, vol. 38 no. 6
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 25 February 2020

Shengquan Wang, Chao Wang, Yong Cai and Guangyao Li

The purpose of this paper is to improve the computational speed of solving nonlinear dynamics by using parallel methods and mixed-precision algorithm on graphic processing units…

Abstract

Purpose

The purpose of this paper is to improve the computational speed of solving nonlinear dynamics by using parallel methods and mixed-precision algorithm on graphic processing units (GPUs). The computational efficiency of traditional central processing units (CPUs)-based computer aided engineering software has been difficult to satisfy the needs of scientific research and practical engineering, especially for nonlinear dynamic problems. Besides, when calculations are performed on GPUs, double-precision operations are slower than single-precision operations. So this paper implemented mixed precision for nonlinear dynamic problem simulation using Belytschko-Tsay (BT) shell element on GPU.

Design/methodology/approach

To minimize data transfer between heterogeneous architectures, the parallel computation of the fully explicit finite element (FE) calculation is realized using a vectorized thread-level parallelism algorithm. An asynchronous data transmission strategy and a novel dependency relationship link-based method, for efficiently solving parallel explicit shell element equations, are used to improve the GPU utilization ratio. Finally, this paper implements mixed precision for nonlinear dynamic problems simulation using the BT shell element on a GPU and compare it to the CPU-based serially executed program and a GPU-based double-precision parallel computing program.

Findings

For a car body model containing approximately 5.3 million degrees of freedom, the computational speed is improved 25 times over CPU sequential computation, and approximately 10% over double-precision parallel computing method. The accuracy error of the mixed-precision computation is small and can satisfy the requirements of practical engineering problems.

Originality/value

This paper realized a novel FE parallel computing procedure for nonlinear dynamic problems using mixed-precision algorithm on CPU-GPU platform. Compared with the CPU serial program, the program implemented in this article obtains a 25 times acceleration ratio when calculating the model of 883,168 elements, which greatly improves the calculation speed for solving nonlinear dynamic problems.

Details

Engineering Computations, vol. 37 no. 6
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 29 January 2020

Carlos Enrique Torres-Aguilar, Jesús Xamán, Pedro Moreno-Bernal, Iván Hernández-Pérez, Ivett Zavala-Guillén and Irving Osiris Hernández-López

The purpose of this study is to propose a novel relaxation modified factor to accelerate the numerical solution of the radiative transfer equation (RTE) with several…

Abstract

Purpose

The purpose of this study is to propose a novel relaxation modified factor to accelerate the numerical solution of the radiative transfer equation (RTE) with several high-resolution total variation diminishing schemes. The methodology proposed is denoted as the X-factor method.

Design/methodology/approach

The X-factor method was compared with the technique deferred-correction (DC) for the calculations of a two-dimensional cavity with absorting-emiting-scatteting gray media using the discrete ordinates method. Four parameters were considered to evaluate: the absorption coefficient, the emissivity of boundary surface, the scattering albedo and under-relaxation factor.

Findings

The results showed the central processing unit (CPU) time of X-factor method was lower than DC. The reductions of CPU time with the X-factor method were observed from 0.6 to 75.4%.

Originality/value

The superiority of the X-factor method over DC was showed with the reduction of CPU time of the numerical solution of RTE for evaluated cases.

Details

Engineering Computations, vol. 37 no. 5
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 1 January 1983

S. SITHARAMA IYENGAR, PAUL O'NEILL and AVIS O'NEILL

The last decade has witnessed a growing concern among computer scientists to understand the complex interactions between humans and computer hardware. The work described in this…

Abstract

The last decade has witnessed a growing concern among computer scientists to understand the complex interactions between humans and computer hardware. The work described in this paper is an experimental study of a user‐computer interaction on a time‐sharing computer terminal network over a period of 1 year. The user‐system interaction described in this paper refers to a university environment. The user‐system performance variables considered are arrival patterns of jobs, inter‐arrival time, connect time, cpu time and think time. The users of the systems are grouped into on‐ and off‐campus users; a two‐way analysis of variance without replications established that arrival volume depended upon the weekday but not upon the user group. The pattern of arrivals throughout one day required an empirical distribution. Coefficient of variation indicated hyper‐exponential distributions for inter‐arrival time, connect time and cpu time, but an exponential distribution for think time. Furthermore, the experimental research described in this paper supports the fact that a hypothesis to characterize the interaction between the user and the computing system can be developed for an efficient use of the system.

Details

Kybernetes, vol. 12 no. 1
Type: Research Article
ISSN: 0368-492X

Article
Publication date: 4 July 2016

José I.V. Sena, Cedric Lequesne, L Duchene, Anne-Marie Habraken, Robertt A.F. Valente and Ricardo J Alves de Sousa

Numerical simulation of the single point incremental forming (SPIF) processes can be very demanding and time consuming due to the constantly changing contact conditions between…

Abstract

Purpose

Numerical simulation of the single point incremental forming (SPIF) processes can be very demanding and time consuming due to the constantly changing contact conditions between the tool and the sheet surface, as well as the nonlinear material behaviour combined with non-monotonic strain paths. The purpose of this paper is to propose an adaptive remeshing technique implemented in the in-house implicit finite element code LAGAMINE, to reduce the simulation time. This remeshing technique automatically refines only a portion of the sheet mesh in vicinity of the tool, therefore following the tool motion. As a result, refined meshes are avoided and consequently the total CPU time can be drastically reduced.

Design/methodology/approach

SPIF is a dieless manufacturing process in which a sheet is deformed by using a tool with a spherical tip. This dieless feature makes the process appropriate for rapid-prototyping and allows for an innovative possibility to reduce overall costs for small batches, since the process can be performed in a rapid and economic way without expensive tooling. As a consequence, research interest related to SPIF process has been growing over the last years.

Findings

In this work, the proposed automatic refinement technique is applied within a reduced enhanced solid-shell framework to further improve numerical efficiency. In this sense, the use of a hexahedral finite element allows the possibility to use general 3D constitutive laws. Additionally, a direct consideration of thickness variations, double-sided contact conditions and evaluation of all components of the stress field are available with solid-shell and not with shell elements. Additionally, validations by means of benchmarks are carried out, with comparisons against experimental results.

Originality/value

It is worth noting that no previous work has been carried out using remeshing strategies combined with hexahedral elements in order to improve the computational efficiency resorting to an implicit scheme, which makes this work innovative. Finally, it has been shown that it is possible to perform accurate and efficient finite element simulations of SPIF process, resorting to implicit analysis and continuum elements. This is definitively a step-forward on the state-of-art in this field.

Details

Engineering Computations, vol. 33 no. 5
Type: Research Article
ISSN: 0264-4401

Keywords

Article
Publication date: 1 April 2006

Rajanand Rajagopalan and C. Richard Cassady

The purpose of this paper is to develop an improved, enumerative solution procedure for solving the original selective maintenance problems. Selective maintenance refers to the…

1037

Abstract

Purpose

The purpose of this paper is to develop an improved, enumerative solution procedure for solving the original selective maintenance problems. Selective maintenance refers to the process of identifying the set of maintenance actions to perform from a desirable set of maintenance actions.

Design/methodology/approach

A series of four improvements to a previously proposed enumerative solution procedure are presented. The improvements are defined and tested sequentially on an experimental set of problem instances. The improvements are characterized relative to the achieved reduction in CPU time for a software application.

Findings

The improved enumerative procedure reduces the CPU time required to solve the selective maintenance problems by as much as 99 per cent. There is a corresponding increase in practical problem size of more than 200 per cent.

Practical implications

Almost all organizations use a variety of repairable systems to achieve their mission. Typically, these systems have to share the limited maintenance resources possessed by the organization. Therefore, an improved ability to solve selective maintenance problem is relevant to many industries.

Originality/value

The body of knowledge relative to selective maintenance continues to grow. However, this is the first study aimed at improving the capability of engineers to solve practically‐sized selective maintenance problems.

Details

Journal of Quality in Maintenance Engineering, vol. 12 no. 2
Type: Research Article
ISSN: 1355-2511

Keywords

Open Access
Article
Publication date: 5 September 2023

Ali Akbar Izadi and Hamed Rasam

Efficient thermal management of central processing unit (CPU) cooling systems is vital in the context of advancing information technology and the demand for enhanced data…

Abstract

Purpose

Efficient thermal management of central processing unit (CPU) cooling systems is vital in the context of advancing information technology and the demand for enhanced data processing speeds. This study aims to explore the thermal performance of a CPU cooling setup using a cylindrical porous metal foam heat sink.

Design/methodology/approach

Nanofluid flow through the metal foam is simulated using the Darcy–Brinkman–Forschheimer equation, accounting for magnetic field effects. The temperature distribution is modeled through the local thermal equilibrium equation, considering viscous dissipation. The problem’s governing partial differential equations are solved using the similarity method. The CPU’s hot surface serves as a solid wall, with nanofluid entering the heat sink as an impinging jet. Verification of the numerical results involves comparison with existing research, demonstrating strong agreement across numerical, analytical and experimental findings. Ansys Fluent® software is used to assess temperature, velocity and streamlines, yielding satisfactory results from an engineering standpoint.

Findings

Investigating critical parameters such as Darcy number (10−4DaD ≤ 10−2), aspect ratio (0.5 ≤ H/D ≤ 1.5), Reynolds number (5 ≤ ReD,bf ≤ 3500), Eckert number (0 ≤ ECbf ≤ 0.1) , porosity (0.85 ≤ ε ≤ 0.95), Hartmann number (0 ≤ HaD,bf ≤ 300) and the volume fraction of nanofluid (0 ≤ φ ≤ 0.1) reveals their impact on fluid flow and heat sink performance. Notably, Nusselt number will reduce 45%, rise 19.2%, decrease 14.1%, and decrease 0.15% for Reynolds numbers of 600, with rising porosity from 0.85 to 0.95, Darcy numbers from 10−4 to 10−2, Eckert numbers from 0 to 0.1, and Hartman numbers from 0 to 300.

Originality/value

Despite notable progress in studying thermal management in CPU cooling systems using porous media and nanofluids, there are still significant gaps in the existing literature. First, few studies have considered the Darcy–Brinkman–Forchheimer equation, which accounts for non-Darcy effects and the flow and geometric interactions between coolant and porous medium. The influence of viscous dissipation on heat transfer in this specific geometry has also been largely overlooked. Additionally, while nanofluids and impinging jets have demonstrated potential in enhancing thermal performance, their utilization within porous media remains underexplored. Furthermore, the unique thermal and structural characteristics of porous media, along with the incorporation of a magnetic field, have not been fully investigated in this particular configuration. Consequently, this study aims to address these literature gaps and introduce novel advancements in analytical modeling, non-Darcy flow, viscous dissipation, nanofluid utilization, impinging jets, porous media characteristics and the impact of a magnetic field. These contributions hold promising prospects for improving CPU cooling system thermal management and have broader implications across various applications in the field.

Details

International Journal of Numerical Methods for Heat & Fluid Flow, vol. 34 no. 1
Type: Research Article
ISSN: 0961-5539

Keywords

Article
Publication date: 17 October 2018

Sura Nawfal and Fakhrulddin Ali

The purpose of this paper is to achieve the acceleration of 3D object transformation using parallel techniques such as multi-core central processing unit (MC CPU) or graphic…

Abstract

Purpose

The purpose of this paper is to achieve the acceleration of 3D object transformation using parallel techniques such as multi-core central processing unit (MC CPU) or graphic processing unit (GPU) or even both. Generating 3D animation scenes in computer graphics requires applying a 3D transformation on the vertices of the objects. These transformations consume most of the execution time. Hence, for high-speed graphic systems, acceleration of vertex transform is very much sought for because it requires many matrix operations (need) to be performed in a real time, so the execution time is essential for such processing.

Design/methodology/approach

In this paper, the acceleration of 3D object transformation is achieved using parallel techniques such as MC CPU or GPU or even both. Multiple geometric transformations are concatenated together at a time in any order in an interactive manner.

Findings

The performance results are presented for a number of 3D objects with paralleled implementations of the affine transform on the NVIDIA GPU series. The maximum execution time was about 0.508 s to transform 100 million vertices using LabVIEW and 0.096 s using Visual Studio. Other results also showed the significant speed-up compared to CPU, MC CPU and other previous work computations for the same object complexity.

Originality/value

The high-speed execution of 3D models is essential in many applications such as medical imaging, 3D games and robotics.

Details

Journal of Engineering, Design and Technology, vol. 16 no. 6
Type: Research Article
ISSN: 1726-0531

Keywords

1 – 10 of over 4000