The full text of this article

 

Problems with using MPI 1.1 and 2.0 as compilation targets for parallel language implementations
by Dan Bonachea, Jason Duell
International Journal of High Performance Computing and Networking (IJHPCN), Vol. 1, No. 1/2/3, 2004

 

Abstract: MPI support is nearly ubiquitous on high-performance systems today and is generally highly tuned for performance. It would thus seem to offer a convenient 'portable network assembly language' to developers of parallel programming languages who wish to target different network architectures. Unfortunately, neither the traditional MPI 1.1 API nor the newer MPI 2.0 extensions for one-sided communication provide an adequate compilation target for global address space languages, and this is likely to be the case for many other parallel languages as well. Simulating one-sided communication under the MPI 1.1 API is too expensive, while the MPI 2.0 one-sided API imposes a number of significant restrictions on memory access patterns that would need to be incorporated at the language level, as a compiler cannot effectively hide them given current conflict and alias detection algorithms.

Online publication date: Fri, 05-Aug-2005

 

is only available to individual subscribers or to users at subscribing institutions.

 
Existing subscribers:
Go to Inderscience Online Journals to access the Full Text of this article.

 
Pay per view:
If you are not a subscriber and you just want to read the full contents of this article, buy online access here.

 
Complimentary Subscribers, Editors or Members of the Editorial Board of the International Journal of High Performance Computing and Networking (IJHPCN):
Login with your Inderscience username and password:

 

    Username:        Password:         

Forgotten your password?


 
Want to subscribe?
A subscription gives you complete access to all articles in the current issue, as well as to all articles in the previous three years (where applicable). See our Orders page to subscribe.

 
If you still need assistance, please email subs@inderscience.com