دانلود رایگان مقاله الگوریتم متناوب موازی با و بدون تداخل در معماریهای چند هسته ای

عنوان فارسی
الگوریتم های تکرار متناوب موازی با و بدون تداخل در معماریهای چند هسته ای
عنوان انگلیسی
Parallel alternating iterative algorithms with and without overlapping on multicore architectures
صفحات مقاله فارسی
0
صفحات مقاله انگلیسی
10
سال انتشار
2015
نشریه
الزویر - Elsevier
فرمت مقاله انگلیسی
PDF
کد محصول
6E105
رشته های مرتبط با این مقاله
مهندسی کامپیوتر
گرایش های مرتبط با این مقاله
مهندسی الگوریتم ها و محاسبات، مهندسی نرم افزار و معماری سیستم های کامپیوتری
مجله
پیشرفت در مهندسی نرم افزار
دانشگاه
گروه فیزیک و معماری کامپیوتر، دانشگاه میگل هرناندز، آلیکانته، اسپانیا
کلمات کلیدی
روش دو مرحله ای بلوک ، تکرار متناوب، تداخل، محاسبات موازی، حافظه اشتراک گذاشته شده، حافظه توزیع شده، معادله لاپلاس، زنجیره مارکوف
چکیده

Abstract


We consider the problem of solving large sparse linear systems where the coefficient matrix is possibly singular but the equations are consistent. Block two-stage methods in which the inner iterations are performed using alternating methods are studied. These methods are ideal for parallel processing and provide a very general setting to study parallel block methods including overlapping. Convergence properties of these methods are established when the matrix in question is either M-matrix or symmetric matrix. Different parallel versions of these methods and implementation strategies, with and without overlapping blocks, are explored. The reported experiments show the behavior and effectiveness of the designed parallel algorithms by exploiting the benefits of shared memory inside the nodes of current SMP supercomputers.

نتیجه گیری

5. Conclusions


In this paper we have studied the problem of solving large consistent linear systems by means of parallel alternating two-stage algorithms with and without overlapping. These algorithms have been applied to both singular and nonsingular large linear systems. In the nonsingular case, the problem to be solved comes from the discretization of the Laplace’s equation while in the singular case the test problems arise from Markov chain modeling. The algorithms have been implemented and tested on distributed and shared memory, and using a distributed shared memory model, obtaining a good scalability and efficiency. Generally, the PALU algorithms behave better than the PAGS algorithms. On the other hand, the overlapping algorithms have sped up the convergence time of the non-overlapping algorithms. The amount of overlap needed to improve the convergence rate is problem specific and depends on the characteristics of the matrix and the block diagonal structure considered in the corresponding parallel algorithm.


بدون دیدگاه