Unit name | High Performance Computing |
---|---|
Unit code | COMSM2400 |
Credit points | 10 |
Level of study | M/7 |
Teaching block(s) |
Teaching Block 2 (weeks 13 - 24) |
Unit director | Professor. McIntosh-Smith |
Open unit status | Not open |
Pre-requisites |
Ability to program competently in C. |
Co-requisites |
None |
School/department | Department of Computer Science |
Faculty | Faculty of Engineering |
The aim of this unit is to introduce and explore technologies relating to high performance, high throughput, and high availability computing, and to offer practical hands-on use of and experience with said technologies. Students completing the unit should have had an opportunity to integrate content from other units in the programme, for example implementing high performance parallel versions of algorithms from COMS21103 or COMS21202 based on theory introduced in COMS22101. The syllabus will include: • Algorithmic models (the view from Berkeley) • Computational models (PRAM, Flynn's taxonomy). • Communication models (interconnects, message passing). • Memory models (NUMA, COMA). • Single-computer technologies (vector computing via SSE, multi-core computing via OpenMP, stream computing via OpenCL/CUDA). • Multi-computer technologies (cluster computing via MPI, cloud/grid computing via Hadoop/MapReduce). • Other approaches (batch processing via Condor, distributed computing via BOINC, distributed and redundant file systems, e.g., RAID/GFS, load balancing, check-pointing). • Design and implementation of parallel algorithms and libraries.
On successful completion of this unit, students will be able to:
Specific learning outcomes will be tackled through focused coursework activities, including:
• Many-core parallel programming through stream languages such as OpenCL and Cuda
Roughly 2/3 of teaching in lecture format, 1/3 in laboratory or problem class format.
Assessment for the unit is 100% via coursework assignments based on hands-on use of high performance computing platforms (e.g., BlueCrystal or similar). The assignments will turn the theory developed in this and previous units into practical experience.
David A. Patterson and John L. Hennessy. Computer Organization and Design: The Hardware/Software Interface. Morgan Kaufman. 2004. ISBN: 1-558-60604-1 Background A. Grama, G. Karypis, V. Kumar and A. Gupta Introduction to Parallel Computing Addison Wesley, 2nd Edition. ISBN: 0201648652 Background B. Chapman, G. Jost and R. van der Pas Using OpenMP: Portable Shared Memory Parallel Programming MIT Press ISBN: 0262533022 Background P. Pacheco Parallel Programming with MPI Morgan Kaufmann ISBN: 1558603395 Background