I decided to try a new thing at the beginning of 2019. You probably heard about competitive programming and about popular websites like Topcoder, Codility, and others. The idea is that you can try yourself in solving programming puzzles. Usually, it’s not enough to just solve the puzzle, your solution must qualify for certain criteria such as algorithmic complexity and memory usage. Also in competitive challenges, it is important how fast you solved the problem.
Those sorts of challenges usually test your problem-solving skills as well as your knowledge of algorithms and data structures. It is the kind of task we usually see in programming job interviews. You need to write the code from scratch to a complete solution.
What I am starting is also a form of a contest but aims at developing a different set of skills. We will train to optimize already optimized code. We will learn hardware optimizations. This is briefly how it will look like: I will pick one existing benchmark of my choice and will send it to all my subscribers. You will have the time to play with the benchmark and find all the performance headroom you can find. You then send all your findings to me (modified sources/assemblies/compiler patches?/whatever speeds up the benchmark) and I check it and run it on my hardware. In the end, I will announce the winners.
Before I explain it in more detail I feel a need of writing a disclaimer:
Disclaimer: This is absolutely a non-profit effort. I’m not looking to make any money from it or use someone’s knowledge for my own purposes. I will not use submissions to extract any intellectual property they might have. Also, this is not aiming to advertise any particular software product including benchmarks/compilers. That’s why all the benchmarks and compilers I will use require to be open sourced. Neither does it aim to advertise any particular HW.
In this contest, I try to emulate the situation in performance-critical projects in the final stages. When all the functionality is delivered and tested. But before shipping the binary to the customers you were asked to tune the app to its peak performance. You know the hardware it will be deployed to. You have the sources in your hands and the freedom to modify them (without introducing any bugs :) ). You are the build master, so you can add any compiler flags you want.
For the challenges that I will send I’m not looking for optimizations that fall into the category: “Oh, I just used quicksort instead of bubble sort”. Expect the benchmark to be already optimized to some degree. Your task is to tune it for particular hardware to peak performance.
A0: You will learn/practice how to do optimizations for HW. This might include eliminating cache misses by inserting prefetch instructions, and getting rid of Code alignment issues, improving performance by helping the compiler to vectorize/unroll the loop better. You will learn different techniques as you go.
A1: It will be an open sourced benchmark written in C/C++. Usually several source files. It should be easy to build and require minimal dependencies. Preferably it should have some form of validation.
A2: Most likely it will be 64-bit Linux with Intel x86 CPU. For a start, I will not bother with disabling CPU dynamic frequency scaling features or setting thread affinity. I might do this in the future.
A3: It doesn’t matter much. If you don’t have Intel CPU or you are on Windows/Mac, just optimize for whatever you have. I would be happy to know about optimizations that help other CPUs, operating systems, etc. I don’t have real prizes to give, so it’s mostly practicing and learning.
A4: Use all of your knowledge. Start with profiling the benchmark. You can browse through posts on this blog. Additionally, I will write a separate post that might help beginners.
A5: Good news! All the dirty tricks allowed! The goal of this contest is to learn how to squeeze as much performance as possible from the hardware using any means available. You can modify sources and insert any compiler hints like pragmas, builtins, function attributes, etc. Also, you can generate an assembly listing (
-S compiler option) and modify it. Finally, you can add some compiler options that might speed up the benchmark.
A6: 1) Do not rewrite the benchmark completely or introduce major changes in algorithms. The good judging rule is “everything interesting should be calculated in runtime”.
2) Do not manually parallelize the benchmark, e.g converting it from single- to multi-threaded or offload computations to the GPU. I mean, I’m glad that you can do it and I will be happy to take a look at what you did, but it’s not the intent of the contest.
3) Using PGO is allowed, however you can use it only for driving your optimizations, not for the submission. So, you can check how the benchmark gets better with PGO and understand why. And then make this optimization manually. Again, the purpose is practicing and learning.
A7: 1) I will not accept binaries for security reasons.
2) You can send patch files that I can apply to the sources of the benchmark or just assembly listing. If you did multiple optimizations please split them into separate patches.
git format-patch is the right tool for that. This will save me a lot of time.
3) If you send assembly listing files please do also include a diff file from the baseline (what you changed in the assembly). If you provide modified assembly listings they should be generated only with open-sourced C/C++ compilers like GCC and clang. You can of course cheat and generate assembly with some other compiler that is better for the benchmark, but I will probably easily detect that. And it’s not about tricks, it’s about learning.
4) Make sure you specify clearly the compiler options if they are different from the baseline.
5) If you are capable of hacking a compiler that’s also acceptable. You can send me patches for gcc/llvm compilers which I can apply and use for building the benchmark. Please use the top of tree revisions because it will be easier for me to apply them. We can then use it for improving our open-source compilers. Also, it would be very nice if you can send me a textual description of all the optimizations you made.
A8: I will take your sources, build them on my machine and run the benchmark. I will run your binary multiple times (depending on the running time) and take the minimum. I’m thinking about testing all the solutions on some cloud machine, but that’s not settled yet.
A9: I will calculate your score as a ratio between the execution time of the binary with your optimizations and the baseline.
I know there are a lot more concerns you might have. This is just a first attempt with a focus on learning how to do HW optimizations. In the end, I don’t have real prizes to give out. :) After each contest, I will share all the findings people did, so there is a big opportunity to learn from others! I know that a number of really experienced guys read this blog, so I encourage everyone to participate. Everyone is welcome!
All communication (including sending benchmarks and score submissions) will happen through emails, so make sure to subscribe using the form at the bottom of the page! I am planning to start the first contest at the end of February 2019.
Let me know what you think about it or if you have any ideas or comments. You can also vote if you like it using the buttons below.
Information presented in llvm documentation: Benchmarking tips might also be helpful.