EPCC and the MSc in High Performance Computing run an annual Guest Lecture series, with speakers drawn from industry and academia. Talks focus on state-of-the-art applications of HPC in real world situations, with many reflecting the type of work our MSc
Iain Bethune (ibethune@exseed.ed.ac.uk)
Bioinformatics and more widely Computational Biology is a largely data-driven Science. The array of high-throughput technology platforms in the last 10 years mean that the amount of data being generated in this field is likely to enter into Exabytes by 2020. The challenges associated with this are quite different from the data sets generated by High Energy Physics or Astrophysics in that they tend to gathered from a wide variety of different providers. Meta-analyses of these data sets can give startling new insights but come with many caveats - in particular that the quality of the data from each provider can be highly variable. I will spend some time talking about one set of experiences I have dealing with one specific technology platform and in particular how it is clear that the detection of bias in data sets is a key element of any high-throughput analysis. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
Performing complex solar shading analysis to take into account the sun's path and solar penetration on large buildings has historically consumed very many CPU cycles for IES "Virtual Environment" (3D building physics) simulation users. One particularly complex model took almost 2 weeks to process. A description of the collaboration between IES and EPCC to apply HPC practices to the solar analysis algorithm and benchmark results will be presented. Additionally we will cover the pitfalls found when integrating the MPI solution in a commercial environment and how we addressed these. The presentation will conclude with a brief description of in-progress and potential future developments with EPCC. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
Intel will provide an insight into future HPC technology development looking at hardware trends, ecosystem support and the challenges around ExaScale computing. The talk will also touch upon the convergence of High Performance Computing and High Performance Data Analytics, examining where the effective use of this rapidly maturing capability can provide industry and academia with a competitive advantage. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
PrimeGrid is a volunteer computing project that gives participants the chance to be the discoverer of a new world record prime number! In addition, we are working towards the solution of several mathematical problems which have remained unsolved for over 50 years. The talk will cover some basic facts about prime numbers, the history of the search for large primes (and a little of the maths!), and show the audience how they can use their computers to join PrimeGrid and find new primes of their own. This talk was given at the British Science Festival on Sat 7th September 2013 by Iain Bethune of EPCC.
There are several important science and engineering problems that require the coordinated execution of multiple high-performance simulations. Some common scenarios include but are not limited to, "an ensemble of tasks", "loosely-coupled simulations of tightly-coupled simulations" or "multi-component multi-physics simulations". However, historically supercomputing centers, have supported and priortised the execution of single "jobs" on supercomputers. Not suprisingly, the tools and capabilities to support coordinated multiple simulations are limited. A promising way to overcome this common limitation is the use of a Pilot-Job --- which can be defined as a container or placeholder job to provide multi-level scheduling via an application-level scheduling overlay over the system scheduler. We discuss both the theory and practise of Pilot-Jobs: Specifically, we introduce the P* Model of Pilot-Jobs and present "BigJob" as a SAGA-based extensible, interopable and scalable implementation of the P* Model. We then discuss several science problems that have/are using BigJob to execute multiple simulations at unprecedented scales on a range of supercomputers and distributed supercomputing infrastructure such as XSEDE. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
In this talk I will give a brief history of parallel processing in games and how the industry has responded to hardware changes in its constant race to create games with more, better and faster. I then consider some of the lessons we have learned so far and finish with my opinion on how a future game engine might be structured to target many-core architectures. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course. Talk slides Videos
The Met Office hosts some the largest computers in the UK to predict the weather and changes in the climate. Best known for the Public Weather Service, the Met Office also advises UK government and other organisations world wide. The difficulties of modelling such a complex system as the Earth's atmosphere, how it is actually done and what future computational challenges lie ahead will be discussed. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
This talk will look at the drivers leading the way to Exascale computing. It will examine technology trends, and use these to infer some of the characteristics of future large scale HPC systems. It will also look in to what this means for the software environment, reliability, use and affordability of IT systems at the high end. Recently, the IT industry has become more focused on problems involving the use and management of large data sets (Big Data) and we map this trend on to the consequences of Exascale computing. What will future architectures look like, how will they be used, and what does that mean for applications development today? This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
Building design has been revolutionised in recent years. New materials and better construction techniques have allowed bespoke and impressive public spaces to be created. Consider the Millennium Dome, Hong Kong airport, or the 2012 Olympic stadium. These bespoke spaces represent a significant challenge for fire safety. Unlike for conventional buildings, with regular layouts and dimensions, there is very little experience to suggest how fires might develop in these spaces, and how people might react in an emergency situation. An Edinburgh-led consortium, called FireGrid, is developing technology designed to tackle this specific problem head-on. By combining advanced modelling with sensor data, in a high performance computing (HPC) context, the FireGrid team are developing a system that can provide fire fighters with better information to support their response, to control the impact of a fire emergency, and to reduce the risk to life and property. In this talk we will review progress of the FireGrid consortium and describe a technology prototype that has been built to demonstrate the potential for such a system, and to stress some of the key technologies that are required. This talk was given as part of our MSc in HPC's 'HPC Ecosystem' course.
The past five years have seen the use of graphical processing units for computation grow from being the interest of handful of early adopters to a mainstream technology used in the world’s largest supercomputers. The CUDA GPU programming ecosystem today provides all that a developer needs to accelerate scientific applications with GPUs. The architecture of a GPU has much to offer to the future of large-scale computing where energy-efficiency is paramount. NVIDIA is the lead contractor for the DARPA-funded Echelon project investigating efficient parallel computer architectures for the exascale era. Timothy Lanfear is a Solution Architect in NVIDIA’s Professional Solutions Group, promoting the use of the NVIDIA Tesla(TM) computing solution for high-performance computing. He has twenty years’ experience in HPC, starting as a computational scientist in British Aerospace’s corporate research centre, and then moving to technical pre-sales roles with Hitachi, ClearSpeed, and most recently NVIDIA. He has a degree in Electrical Engineering and a PhD for research in the field of graph theory, both from Imperial College London.
We will review some basic properties and theorems regarding prime numbers, and take a quick trip through the history of prime number searching. Secondly, we will discuss two classes of algorithms of importance for computational primality testing - Sieving and the Lucas-Lehmer (and similar) tests - and their implementations on modern CPUs and GPUs. Finally, we will introduce GIMPS and PrimeGrid, two large and well-known distributed prime search projects. Links:
There is a plethora of version control systems and it is not obvious which to choose. Do you pine for the warm, comforting blanket of CVS? Fear not! I shall give an overview of the zoo of version control systems available to the modern programmer. I'll describe the design principles (there is much overlap) and how these influence the use patterns. I hope to convince you that modern systems are at least as good as CVS and Subversion, and to indicate some of the areas in which they make a programmer's life much, much easier. Links: Talk slides - Talk slides
Cloud computing has become one of the most advertised and talked about technologies for the past few years since many ISPs/IT firms have begun to adopt virtualisation technologies. As cloud technologies mature, more and more businesses are moving their services into “the cloud†or building internal clouds to cater for their IT needs. Gihan Munasinghe & Tabassum Sharif of Flexiant discuss the advantages of having cloud infrastructures, introduce Flexiscale and look into how you can build your own internal cloud infrastructure. Links: Slides - Talk slides
Fluidity is a powerful Computational Fluid Dynamics framework developed at Imperial College over the last 20 years. Fluidity can be applied to a number of scientific applications, ranging from classic CFD, to oceans and multi-material problems. It uses a variety of discretisations on an unstructured mesh and includes a number of novel and innovative features, such as adaptive re-meshing, a user friendly interface, and a python interface. The main advantage Fluidity has is the adaptive re-meshing, which allows resolution to be focused where and when it is required during a simulation, depending on the simulation state at that time. This talk will briefly cover some of the features and science behind Fluidity before covering details of AMCG’s software development methodologies and the parallel scaling behaviour of Fluidity on HECToR. Finally, future developments will be covered. Links: Slides - Talk slides
The Materials Chemistry Consortium (MCC) is the single largest consumer of resources on the HECToR National Supercomputer. In this talk I will give an overview of how MCC members exploit HECToR in a wide range of materials chemistry research and the types of applications that we are interested in. Links: Slides - Talk slides
Traditional disk technology has lagged behind processor and memory technology for some time. During this session we will discuss what the problems are and how new technologies such as enterprise class Solid State Drives (SSD) are helping. Until now, high performance storage systems have been designed around the characteristics of spinning magnetic media. This session will also discuss design considerations for such high performance systems, an overview of storage virtualization, and how these concepts are increasingly important with the move towards solid state storage media. Links: Slides - Talk slides
The Met Office has long used high performance computing to support its activities in weather prediction and climate research. This talk starts by giving an overview of the numerical weather prediction process and continues with a look at the HPC platforms used to make solving these problems possible. Some of the challenges currently faced, particularly in the area of scalability, are presented and recent progress in these areas discussed. Links: Slides - Talk slides
This talk traces the development of the different types of instruction-level parallelism that have been incorporated into the hardware of processors from the early 1960s to the present day. We will see how the use of parallel function units in the CDC 6600 eventually led to the design of the Cray-1 with its vector processing instructions and superscalar processors such as the Alpha. The talk also covers Very Long Instruction Word systems and SIMD systems. The talk is accompanied by visual demonstrations of the activities that occur within processors using simulation models developed using the HASE computer architecture simulation environment. Links: Slides - Talk slides
Lattice QCD is a very important tool for understanding the Standard Model of particle physics. It presents enormous computational challenges, both in carrying out simulations and in making sure they are as efficient as possible. This lecture explains why we need lattice QCD; describes how the UKQCD Collaboration sets out to overcome these computational challenges; and illustrates how GPUs are used here in Edinburgh. Links: Slides - Talk slides