High-performance computing: paradigm and infrastructure
Gespeichert in:
1. Verfasser: | |
---|---|
Format: | Elektronisch E-Book |
Sprache: | English |
Veröffentlicht: |
Hoboken, NJ
Wiley-Interscience
c2006
|
Schriftenreihe: | Wiley series on parallel and distributed computing
|
Schlagworte: | |
Online-Zugang: | FAW01 FAW02 Volltext |
Beschreibung: | Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002 Includes bibliographical references and index Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming - 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching - through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results - 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream |
Beschreibung: | 1 Online-Ressource (xxxviii, 778 p.) |
ISBN: | 047165471X 0471732702 0471732710 9780471654711 9780471732709 9780471732716 |
Internformat
MARC
LEADER | 00000nmm a2200000zc 4500 | ||
---|---|---|---|
001 | BV043108019 | ||
003 | DE-604 | ||
005 | 00000000000000.0 | ||
007 | cr|uuu---uuuuu | ||
008 | 151126s2006 |||| o||u| ||||||eng d | ||
020 | |a 047165471X |9 0-471-65471-X | ||
020 | |a 0471732702 |c electronic bk. : Adobe Reader |9 0-471-73270-2 | ||
020 | |a 0471732710 |c electronic bk. |9 0-471-73271-0 | ||
020 | |a 9780471654711 |9 978-0-471-65471-1 | ||
020 | |a 9780471732709 |c electronic bk. : Adobe Reader |9 978-0-471-73270-9 | ||
020 | |a 9780471732716 |c electronic bk. |9 978-0-471-73271-6 | ||
035 | |a (OCoLC)71353018 | ||
035 | |a (DE-599)BVBBV043108019 | ||
040 | |a DE-604 |b ger |e aacr | ||
041 | 0 | |a eng | |
049 | |a DE-1046 |a DE-1047 | ||
082 | 0 | |a 004/.35 |2 22 | |
100 | 1 | |a Yang, Laurence Tianruo |e Verfasser |4 aut | |
245 | 1 | 0 | |a High-performance computing |b paradigm and infrastructure |c edited by Laurence T. Yang, Minyi Guo |
264 | 1 | |a Hoboken, NJ |b Wiley-Interscience |c c2006 | |
300 | |a 1 Online-Ressource (xxxviii, 778 p.) | ||
336 | |b txt |2 rdacontent | ||
337 | |b c |2 rdamedia | ||
338 | |b cr |2 rdacarrier | ||
490 | 0 | |a Wiley series on parallel and distributed computing | |
500 | |a Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002 | ||
500 | |a Includes bibliographical references and index | ||
500 | |a Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming | ||
500 | |a - 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching | ||
500 | |a - through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results | ||
500 | |a - 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr | ||
500 | |a With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream | ||
650 | 7 | |a COMPUTERS / Systems Architecture / Distributed Systems & Computing |2 bisacsh | |
650 | 7 | |a Sistemas distribuídos |2 larpcal | |
650 | 7 | |a Arquitetura e organização de computadores |2 larpcal | |
650 | 7 | |a Processamento eletrônico de dados |2 larpcal | |
650 | 7 | |a Grid Computing |2 swd | |
650 | 7 | |a Supercomputer |2 swd | |
650 | 7 | |a Electronic data processing / Distributed processing |2 fast | |
650 | 7 | |a High performance computing |2 fast | |
650 | 7 | |a Parallel processing (Electronic computers) |2 fast | |
650 | 4 | |a Datenverarbeitung | |
650 | 4 | |a High performance computing | |
650 | 4 | |a Parallel processing (Electronic computers) | |
650 | 4 | |a Electronic data processing |x Distributed processing | |
650 | 0 | 7 | |a Grid Computing |0 (DE-588)4743648-7 |2 gnd |9 rswk-swf |
650 | 0 | 7 | |a Supercomputer |0 (DE-588)4128144-5 |2 gnd |9 rswk-swf |
689 | 0 | 0 | |a Supercomputer |0 (DE-588)4128144-5 |D s |
689 | 0 | 1 | |a Grid Computing |0 (DE-588)4743648-7 |D s |
689 | 0 | |8 1\p |5 DE-604 | |
700 | 1 | |a Guo, Minyi |e Sonstige |4 oth | |
856 | 4 | 0 | |u http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698 |x Aggregator |3 Volltext |
912 | |a ZDB-4-EBA | ||
999 | |a oai:aleph.bib-bvb.de:BVB01-028532210 | ||
883 | 1 | |8 1\p |a cgwrk |d 20201028 |q DE-101 |u https://d-nb.info/provenance/plan#cgwrk | |
966 | e | |u http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698 |l FAW01 |p ZDB-4-EBA |q FAW_PDA_EBA |x Aggregator |3 Volltext | |
966 | e | |u http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698 |l FAW02 |p ZDB-4-EBA |q FAW_PDA_EBA |x Aggregator |3 Volltext |
Datensatz im Suchindex
_version_ | 1804175524636917760 |
---|---|
any_adam_object | |
author | Yang, Laurence Tianruo |
author_facet | Yang, Laurence Tianruo |
author_role | aut |
author_sort | Yang, Laurence Tianruo |
author_variant | l t y lt lty |
building | Verbundindex |
bvnumber | BV043108019 |
collection | ZDB-4-EBA |
ctrlnum | (OCoLC)71353018 (DE-599)BVBBV043108019 |
dewey-full | 004/.35 |
dewey-hundreds | 000 - Computer science, information, general works |
dewey-ones | 004 - Computer science |
dewey-raw | 004/.35 |
dewey-search | 004/.35 |
dewey-sort | 14 235 |
dewey-tens | 000 - Computer science, information, general works |
discipline | Informatik |
format | Electronic eBook |
fullrecord | <?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>07226nmm a2200709zc 4500</leader><controlfield tag="001">BV043108019</controlfield><controlfield tag="003">DE-604</controlfield><controlfield tag="005">00000000000000.0</controlfield><controlfield tag="007">cr|uuu---uuuuu</controlfield><controlfield tag="008">151126s2006 |||| o||u| ||||||eng d</controlfield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">047165471X</subfield><subfield code="9">0-471-65471-X</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">0471732702</subfield><subfield code="c">electronic bk. : Adobe Reader</subfield><subfield code="9">0-471-73270-2</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">0471732710</subfield><subfield code="c">electronic bk.</subfield><subfield code="9">0-471-73271-0</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9780471654711</subfield><subfield code="9">978-0-471-65471-1</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9780471732709</subfield><subfield code="c">electronic bk. : Adobe Reader</subfield><subfield code="9">978-0-471-73270-9</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9780471732716</subfield><subfield code="c">electronic bk.</subfield><subfield code="9">978-0-471-73271-6</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(OCoLC)71353018</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(DE-599)BVBBV043108019</subfield></datafield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">DE-604</subfield><subfield code="b">ger</subfield><subfield code="e">aacr</subfield></datafield><datafield tag="041" ind1="0" ind2=" "><subfield code="a">eng</subfield></datafield><datafield tag="049" ind1=" " ind2=" "><subfield code="a">DE-1046</subfield><subfield code="a">DE-1047</subfield></datafield><datafield tag="082" ind1="0" ind2=" "><subfield code="a">004/.35</subfield><subfield code="2">22</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Yang, Laurence Tianruo</subfield><subfield code="e">Verfasser</subfield><subfield code="4">aut</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">High-performance computing</subfield><subfield code="b">paradigm and infrastructure</subfield><subfield code="c">edited by Laurence T. Yang, Minyi Guo</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="a">Hoboken, NJ</subfield><subfield code="b">Wiley-Interscience</subfield><subfield code="c">c2006</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">1 Online-Ressource (xxxviii, 778 p.)</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="490" ind1="0" ind2=" "><subfield code="a">Wiley series on parallel and distributed computing</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">Includes bibliographical references and index</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming </subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a"> - 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching </subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a"> - through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results </subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a"> - 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">COMPUTERS / Systems Architecture / Distributed Systems & Computing</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Sistemas distribuídos</subfield><subfield code="2">larpcal</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Arquitetura e organização de computadores</subfield><subfield code="2">larpcal</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Processamento eletrônico de dados</subfield><subfield code="2">larpcal</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Grid Computing</subfield><subfield code="2">swd</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Supercomputer</subfield><subfield code="2">swd</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Electronic data processing / Distributed processing</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">High performance computing</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Parallel processing (Electronic computers)</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Datenverarbeitung</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">High performance computing</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Parallel processing (Electronic computers)</subfield></datafield><datafield tag="650" ind1=" " ind2="4"><subfield code="a">Electronic data processing</subfield><subfield code="x">Distributed processing</subfield></datafield><datafield tag="650" ind1="0" ind2="7"><subfield code="a">Grid Computing</subfield><subfield code="0">(DE-588)4743648-7</subfield><subfield code="2">gnd</subfield><subfield code="9">rswk-swf</subfield></datafield><datafield tag="650" ind1="0" ind2="7"><subfield code="a">Supercomputer</subfield><subfield code="0">(DE-588)4128144-5</subfield><subfield code="2">gnd</subfield><subfield code="9">rswk-swf</subfield></datafield><datafield tag="689" ind1="0" ind2="0"><subfield code="a">Supercomputer</subfield><subfield code="0">(DE-588)4128144-5</subfield><subfield code="D">s</subfield></datafield><datafield tag="689" ind1="0" ind2="1"><subfield code="a">Grid Computing</subfield><subfield code="0">(DE-588)4743648-7</subfield><subfield code="D">s</subfield></datafield><datafield tag="689" ind1="0" ind2=" "><subfield code="8">1\p</subfield><subfield code="5">DE-604</subfield></datafield><datafield tag="700" ind1="1" ind2=" "><subfield code="a">Guo, Minyi</subfield><subfield code="e">Sonstige</subfield><subfield code="4">oth</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="u">http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698</subfield><subfield code="x">Aggregator</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">ZDB-4-EBA</subfield></datafield><datafield tag="999" ind1=" " ind2=" "><subfield code="a">oai:aleph.bib-bvb.de:BVB01-028532210</subfield></datafield><datafield tag="883" ind1="1" ind2=" "><subfield code="8">1\p</subfield><subfield code="a">cgwrk</subfield><subfield code="d">20201028</subfield><subfield code="q">DE-101</subfield><subfield code="u">https://d-nb.info/provenance/plan#cgwrk</subfield></datafield><datafield tag="966" ind1="e" ind2=" "><subfield code="u">http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698</subfield><subfield code="l">FAW01</subfield><subfield code="p">ZDB-4-EBA</subfield><subfield code="q">FAW_PDA_EBA</subfield><subfield code="x">Aggregator</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="966" ind1="e" ind2=" "><subfield code="u">http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698</subfield><subfield code="l">FAW02</subfield><subfield code="p">ZDB-4-EBA</subfield><subfield code="q">FAW_PDA_EBA</subfield><subfield code="x">Aggregator</subfield><subfield code="3">Volltext</subfield></datafield></record></collection> |
id | DE-604.BV043108019 |
illustrated | Not Illustrated |
indexdate | 2024-07-10T07:17:40Z |
institution | BVB |
isbn | 047165471X 0471732702 0471732710 9780471654711 9780471732709 9780471732716 |
language | English |
oai_aleph_id | oai:aleph.bib-bvb.de:BVB01-028532210 |
oclc_num | 71353018 |
open_access_boolean | |
owner | DE-1046 DE-1047 |
owner_facet | DE-1046 DE-1047 |
physical | 1 Online-Ressource (xxxviii, 778 p.) |
psigel | ZDB-4-EBA ZDB-4-EBA FAW_PDA_EBA |
publishDate | 2006 |
publishDateSearch | 2006 |
publishDateSort | 2006 |
publisher | Wiley-Interscience |
record_format | marc |
series2 | Wiley series on parallel and distributed computing |
spelling | Yang, Laurence Tianruo Verfasser aut High-performance computing paradigm and infrastructure edited by Laurence T. Yang, Minyi Guo Hoboken, NJ Wiley-Interscience c2006 1 Online-Ressource (xxxviii, 778 p.) txt rdacontent c rdamedia cr rdacarrier Wiley series on parallel and distributed computing Master and use copy. Digital master created according to Benchmark for Faithful Digital Reproductions of Monographs and Serials, Version 1. Digital Library Federation, December 2002 Includes bibliographical references and index Cover HIGH-PERFORMANCE COMPUTING Contents Preface Contributors PART 1 Programming Model 1 ClusterGOP: A High-Level Programming Environment for Clusters 1.1 Introduction 1.2 GOP Model and ClusterGOP Architecture 1.3 VisualGOP 1.4 The ClusterGOP Library 1.5 MPMD programming Support 1.6 Programming Using ClusterGOP 1.7 Summary 2 The Challenge of Providing A High-Level Programming Model for High-Performance Computing 2.1 Introduction 2.2 HPC Architectures 2.3 HPC Programming Models: The First Generation 2.4 The Second Generation of HPC Programming Models 2.5 OpenMP for DMPs 2.6 Experiments with OpenMP on DMPs 2.7 Conclusions 3 SAT: Toward Structured Parallelism Using Skeletons 3.1 Introduction 3.2 SAT: A Methodology Outline 3.3 Skeletons and Collective Operations 3.4 Case Study: Maximum Segment SUM (MSS) 3.5 Performance Aspect in SAT 3.6 Conclusions and Related Work 4 Bulk-Synchronous Parallelism: An Emerging Paradigm of High-Performance Computing 4.1 The BSP Model 4.2 BSP Programming - 4.3 Conclusions 5 Cilk Versus MPI: Comparing Two Parallel Programming Styles on Heterogeneous Systems 5.1 Introduction 5.2 Experiments 5.3 Results 5.4 Conclusion 6 Nested Parallelism and Pipelining in OpenMP 6.1 Introduction 6.2 OpenMP Extensions for Nested Parallelism 6.3 OpenMP Extensions For Thread Synchronization 6.4 Summary 7 OpenMP for Chip Multiprocessors 7.1 Introduction 7.2 3SoC Architecture Overview 7.3 The OpenMp Compiler/Translator 7.4 Extensions to OpenMP for DSEs 7.5 Optimization for OpenMP 7.6 Implementation 7.7 Performance Evaluation 7.8 Conclusions PART 2 Architectural and System Support 8 Compiler and Run-Time Parallelization Techniques for Scientific Computations on Distributed-Memory Parallel Computers 8.1 Introduction 8.2 Background Material 8.3 Compiling Regular Programs on DMPCs 8.4 Compiler and Run-Time Support for Irregular Programs 8.5 Library Support for Irregular Applications 8.6 Related Works 8.7 Concluding Remarks 9 Enabling Partial-Cache Line Prefetching - through Data Compression 9.1 Introduction 9.2 Motivation of Partial Cache-Line Prefetching 9.3 Cache Design Details 9.4 Experimental Results 9.5 Related Work 9.6 Conclusion 10 MPI Atomicity and Concurrent Overlapping I/O 10.1 Introduction 10.2 Concurrent Overlapping I/O 10.3 Implementation Strategies 10.4 Experiment Results 10.5 Summary 11 Code Tiling: One Size Fits All 11.1 Introduction 11.2 Cache Model 11.3 Code Tiling 11.4 Data Tiling 11.5 Finding Optimal Tile Sizes 11.6 Experimental Results 11.7 Related Work 11.8 Conclusion 12 Data Conversion for Heterogeneous Migration/Checkpointing 12.1 Introduction 12.2 Migration and Checkpointing 12.3 Data Conversion 12.4 Coarse-Grain Tagged RMR in MigThread 12.5 Microbenchmarks and Experiments 12.6 Related Work 12.7 Conclusions and Future Work 13 Receiving-Message Prediction and Its Speculative Execution 13.1 Background 13.2 Receiving-Message Prediction Method 13.3 Implementation of the Method in the MIPI Libraries 13.4 Experimental Results - 13.5 Conclusing Remarks 14 An Investigation of the Applicability of Distributed FPGAs to High-Performance Computing 14.1 Introduction 14.2 High Performance Computing with Cluster Computing 14.3 Reconfigurable Computing with EPGAs 14.4 DRMC: A Distributed Reconfigurable Metacomputer 14.5 Algorithms Suited to the Implementation on FPGAs/DRMC 14.6 Algorithms Not Suited to the Implementation on FPGAs/DRMC 14.7 Summary PART 3 Scheduling and Resource Management 15 Bandwidth-Aware Resource Allocation for Heterogeneous Computing Systems to Maximize Throughput 15.1 Introduction 15.2 Related Work 15.3 System Model and Problem Statement 15.4 Resource Allocation to Maximize System Throughput 15.5 Experimental Results 15.6 Conclusion 16 Scheduling Algorithms with Bus Bandwidth Considerations for SMPs 16.1 Intr With hyperthreading in Intel processors, hypertransport links in next generation AMD processors, multi-core silicon in today's high-end microprocessors from IBM and emerging grid computing, parallel and distributed computers have moved into the mainstream COMPUTERS / Systems Architecture / Distributed Systems & Computing bisacsh Sistemas distribuídos larpcal Arquitetura e organização de computadores larpcal Processamento eletrônico de dados larpcal Grid Computing swd Supercomputer swd Electronic data processing / Distributed processing fast High performance computing fast Parallel processing (Electronic computers) fast Datenverarbeitung High performance computing Parallel processing (Electronic computers) Electronic data processing Distributed processing Grid Computing (DE-588)4743648-7 gnd rswk-swf Supercomputer (DE-588)4128144-5 gnd rswk-swf Supercomputer (DE-588)4128144-5 s Grid Computing (DE-588)4743648-7 s 1\p DE-604 Guo, Minyi Sonstige oth http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698 Aggregator Volltext 1\p cgwrk 20201028 DE-101 https://d-nb.info/provenance/plan#cgwrk |
spellingShingle | Yang, Laurence Tianruo High-performance computing paradigm and infrastructure COMPUTERS / Systems Architecture / Distributed Systems & Computing bisacsh Sistemas distribuídos larpcal Arquitetura e organização de computadores larpcal Processamento eletrônico de dados larpcal Grid Computing swd Supercomputer swd Electronic data processing / Distributed processing fast High performance computing fast Parallel processing (Electronic computers) fast Datenverarbeitung High performance computing Parallel processing (Electronic computers) Electronic data processing Distributed processing Grid Computing (DE-588)4743648-7 gnd Supercomputer (DE-588)4128144-5 gnd |
subject_GND | (DE-588)4743648-7 (DE-588)4128144-5 |
title | High-performance computing paradigm and infrastructure |
title_auth | High-performance computing paradigm and infrastructure |
title_exact_search | High-performance computing paradigm and infrastructure |
title_full | High-performance computing paradigm and infrastructure edited by Laurence T. Yang, Minyi Guo |
title_fullStr | High-performance computing paradigm and infrastructure edited by Laurence T. Yang, Minyi Guo |
title_full_unstemmed | High-performance computing paradigm and infrastructure edited by Laurence T. Yang, Minyi Guo |
title_short | High-performance computing |
title_sort | high performance computing paradigm and infrastructure |
title_sub | paradigm and infrastructure |
topic | COMPUTERS / Systems Architecture / Distributed Systems & Computing bisacsh Sistemas distribuídos larpcal Arquitetura e organização de computadores larpcal Processamento eletrônico de dados larpcal Grid Computing swd Supercomputer swd Electronic data processing / Distributed processing fast High performance computing fast Parallel processing (Electronic computers) fast Datenverarbeitung High performance computing Parallel processing (Electronic computers) Electronic data processing Distributed processing Grid Computing (DE-588)4743648-7 gnd Supercomputer (DE-588)4128144-5 gnd |
topic_facet | COMPUTERS / Systems Architecture / Distributed Systems & Computing Sistemas distribuídos Arquitetura e organização de computadores Processamento eletrônico de dados Grid Computing Supercomputer Electronic data processing / Distributed processing High performance computing Parallel processing (Electronic computers) Datenverarbeitung Electronic data processing Distributed processing |
url | http://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&db=nlabk&AN=148698 |
work_keys_str_mv | AT yanglaurencetianruo highperformancecomputingparadigmandinfrastructure AT guominyi highperformancecomputingparadigmandinfrastructure |