Research and implementation of scalable parallel computing based on Map-Reduce

As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no cle...

Full description

Saved in:
Bibliographic Details
Published inJournal of Shanghai University Vol. 15; no. 5; pp. 426 - 429
Main Author 阮青强 沈文枫 柴亚辉 徐炜民
Format Journal Article
LanguageEnglish
Published Heidelberg Shanghai University Press 01.10.2011
School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P.R.China%School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P.R.China
School of Information Engineering, East China Jiaotong University, Nanchang 330013, P.R.China
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no clear idea about distributed and parallel programming, this interface can be utilized to save development time. This paper introduces the method of using Hadoop, the open-source Map-Reduce software platform, to combine PCs to carry out scalable parallel computing. Our experiment using 12 PCs to compute N-body problem based on Map-Reduce model shows that we can get a 9.8x speedup ratio. This work indicates that the Map-Reduce can be applied in scalable parallel computing.
Bibliography:As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no clear idea about distributed and parallel programming, this interface can be utilized to save development time. This paper introduces the method of using Hadoop, the open-source Map-Reduce software platform, to combine PCs to carry out scalable parallel computing. Our experiment using 12 PCs to compute N-body problem based on Map-Reduce model shows that we can get a 9.8x speedup ratio. This work indicates that the Map-Reduce can be applied in scalable parallel computing.
NGUYEN Thanh-cuong, SHEN Wen-feng, CHAI Ya-hui , XU Wei-min ( l. School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P. R. China ;2. School of Information Engineering, East China Jiaotong University, Nanehang 330013, P. R. China)
31-1735/N
Map-Reduce, distributed computing, N-body problem
ObjectType-Article-2
SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 23
ISSN:1007-6417
1863-236X
DOI:10.1007/s11741-011-0763-3