We investigate the secret connection between class layout and software performance. And of course, how to make your software faster.
All posts tagged memory wall
Decreasing the Number of Memory Accesses 1/2
In this post, we are investigating a few common ways to decrease the number of memory accesses in your program.
Frugal Programming: Saving Memory Subsystem Bandwidth
We investigate techniques of frugal programming: how to program so you don’t waste the limited memory resources in your computer system.
For Software Performance, the Way Data is Accessed Matters!
In our experiments with the memory access pattern, we have seen that good data locality is a key to good software performance. Accessing memory sequentially and splitting the data set into small-sized pieces which are processed individually improves data locality and software speed. In this post, we will present a few techniques to improve the…
The memory subsystem from the viewpoint of software: how memory subsystem affects software performance 2/3
We continue the investigation from the previous post, trying to measure how the memory subsystem affects software performance. We write small programs (kernels) to quantify the effects of cache line, memory latency, TLB cache, cache conflicts, vectorization and branch prediction.
The memory subsystem from the viewpoint of software: how memory subsystem affects software performance 1/3
In this post we investigate the memory subsystem of a desktop, server and embedded system from the software viewpoint. We use small kernels to illustrate various aspects of the memory subsystem and how it effects performance and runtime.
Instruction-level parallelism in practice: speeding up memory-bound programs with low ILP
We talk about instruction level parallelism: what instruction-level parallelism is, why is it important for your code’s performance and how you can add instruction-level parallelism to improve the performance of your memory-bound program.
Memory consumption, dataset size and performance: how does it all relate?
We investigate how memory consumption, dataset size and software performance correlate…
When vectorization hits the memory wall: investigating the AVX2 memory gather instruction
For all the engineers who like to tinker with software performance, vectorization is the holy grail: if it vectorizes, this means that it runs faster. Unfortunately, many times this is not the case, and the results of forcing vectorization by any means can mean lower performance. This happens when vectorization hits the memory wall: although…