Riding on Moore's Law
Start date: Apr 1, 2013,
End date: Mar 31, 2018
The most common interpretation of Moore's Law is that the number of components on a chip and accordingly the computer performance doubles every two years. At the end of the 20th century, when clock frequencies stagnated at ~3 GHz, and instruction level parallelism reached the phase of diminishing returns, industry turned towards multiprocessors, and thread level parallelism. However, too much of the technological complexity of multicore architectures is exposed to the programmers, leading to a software development nightmare.We propose a radically new concept of parallel computer architectures, using a higher level of abstraction, Instead of expressing algorithms as a sequence of instruction, we will group instructions into higher-level tasks that will be automatically managed by the architecture, much in the same way superscalar processors managed instruction level parallelism.We envision a holistic approach where the parallel architecture is partially implemented as a software runtime, and the reminder in hardware. The hardware gains the freedom to deliver performance at the expense of additional complexity, as long as it provides the required support primitives for the runtime software to hide complexity from the programmer. Moreover, it offers a single solution that could solve most of the problems we encounter in the current approaches: handling parallelism, the memory wall, the power wall, and the reliability wall in a wide range of application domains from mobile up to supercomputers .We will focus our research on a most efficient form of multicore architecture coupled with vector accelerators for exploiting both thread and data level parallelism.All together, this novel approach toward future parallel architectures is the way to ensure continued performance improvements, getting us out of the technological mess that computers have turned into, once more riding on Moore's Law.
Get Access to the 1st Network for European Cooperation