Traditional data structure designs, whether lock-based or lock-free, provide parallelism via fine grained synchronization among threads. We introduce a new synchronization paradi...
Danny Hendler, Itai Incze, Nir Shavit, Moran Tzafr...
We present a new cache oblivious scheme for iterative stencil computations that performs beyond system bandwidth limitations as though gigabytes of data could reside in an enormou...
Robert Strzodka, Mohammed Shaheen, Dawid Pajak, Ha...
Modern computers have taken advantage of the instruction-level parallelism (ILP) available in programs with advances in both architecture and compiler design. Unfortunately, large...
Exploiting locality at run-time is a complementary approach to a compiler approach for those applications with dynamic memory access patterns. This paper proposes a memory-layout ...
The mpC language was developed to write efJicientand portable programsfor wide range of distributed memory machines. It supports both task and data parallelism, allows both static...
Dmitry Arapov, Alexey Kalinov, Alexey L. Lastovets...