Research and implementation of scalable parallel computing based on Map-Reduce
As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no cle...
Saved in:
Published in | Journal of Shanghai University Vol. 15; no. 5; pp. 426 - 429 |
---|---|
Main Author | |
Format | Journal Article |
Language | English |
Published |
Heidelberg
Shanghai University Press
01.10.2011
School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P.R.China%School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P.R.China School of Information Engineering, East China Jiaotong University, Nanchang 330013, P.R.China |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no clear idea about distributed and parallel programming, this interface can be utilized to save development time. This paper introduces the method of using Hadoop, the open-source Map-Reduce software platform, to combine PCs to carry out scalable parallel computing. Our experiment using 12 PCs to compute N-body problem based on Map-Reduce model shows that we can get a 9.8x speedup ratio. This work indicates that the Map-Reduce can be applied in scalable parallel computing. |
---|---|
Bibliography: | As a parallel programming model, Map-Reduce is used for distributed computing of massive data. Map-Reduce model encapsulates the details of parallel implementation, fault-tolerant processing, local computing and load balancing, etc., provides a simple but powerful interface. In case of having no clear idea about distributed and parallel programming, this interface can be utilized to save development time. This paper introduces the method of using Hadoop, the open-source Map-Reduce software platform, to combine PCs to carry out scalable parallel computing. Our experiment using 12 PCs to compute N-body problem based on Map-Reduce model shows that we can get a 9.8x speedup ratio. This work indicates that the Map-Reduce can be applied in scalable parallel computing. NGUYEN Thanh-cuong, SHEN Wen-feng, CHAI Ya-hui , XU Wei-min ( l. School of Computer Engineering and Science, Shanghai University, Shanghai 200072, P. R. China ;2. School of Information Engineering, East China Jiaotong University, Nanehang 330013, P. R. China) 31-1735/N Map-Reduce, distributed computing, N-body problem ObjectType-Article-2 SourceType-Scholarly Journals-1 ObjectType-Feature-1 content type line 23 |
ISSN: | 1007-6417 1863-236X |
DOI: | 10.1007/s11741-011-0763-3 |