GPU Profiling ============= This tutorial will guide you step-by-step through how to conduct profiling and performance tuning using :code:`nvprof` and :code:`nvvp`. - What is profiling? - Why we need profiling? - How to do profiling? - Profile tools - Hands-on Tutorial - Profiling tips What's profiling? ================= In software engineering, profiling is a form of dynamic program analysis that measures the space (memory) or time complexity of a program, the usage of particular instructions, or the frequency and duration of function calls. Most commonly, profiling information serves to aid program optimization. Briefly, profiler is used to measure application performance. Program analysis tools are extremely important for understanding program behavior. Simple profiling can tell you that how long does an operation take? For advanced profiling, it can interpret why does an operation take a long time? Why we need profiling? ====================== Since training deep neural network typically take a very long time to get over, performance is gradually becoming the most important thing in deep learning field. The first step to improve performance is to understand what parts are slow. No point in improving performance of a region which doesn’t take much time! How to do profiling? ==================== To achieve maximum performance, there are five steps you can take to reach your goals. - Profile the code - Find the slow parts - Work out why they’re slow - Make them fast - Profile the code again Usually, processor has two key performance limits include float point throughput and memory throughput. For GPU, it also need more parallelism to fulfill its potential. This is why they can be so fast. Profiler Tools ============== For general GPU profiling, a bunch of tools are provided from both NVIDIA and third party. :code:`nvprof` is Nvidia profiler and :code:`nvvp` is (GUI based) Nvidia visual profiler. In this tutorial, we will focus on nvprof and nvvp. :code:`test_GpuProfiler` from :code:`paddle/math/tests` directory will be used to evaluate above profilers. .. code-block:: c++ TEST(Profiler, BilinearFwdBwd) { hl_profiler_start(); auto numSamples = 10; auto channels = 16; auto imgSize = 64; testBilinearFwdBwd(numSamples, imgSize, imgSize, channels); hl_profiler_end(); } :code:`hl_profiler_start` and :code:`hl_profiler_end` can be used to profile only regions of interest in PaddlePaddle. They are wrapper functions of :code:`cudaProfilerStart` and :code:`cudaProfilerStop` respectively to avoid program crashes when CPU version of PaddlePaddle invokes them. Hands-on Approach ================= To use this command line profiler :code:`nvprof`, you can simply issue the command: .. code-block:: bash nvprof ./paddle/math/tests/test_GpuProfiler Then, you can get the following profiling result: .. image:: nvprof.png :align: center :scale: 30% For visual profiler :code:`nvvp`, you can either import the output of :code:`nvprof –o ...` or run application through GUI. .. image:: nvvp1.png :align: center :scale: 30% From the perspective of kernel functions, :code:`nvvp` can even illustrate why does an operation take a long time? As shown in the following figure, kernel's block usage, register usage and shared memory usage from :code:`nvvp` allow us to fully utilize all warps on the GPU. .. image:: nvvp2.png :align: center :scale: 30% From the perspective of application, :code:`nvvp` can give you some suggestions to address performance bottleneck. For instance, some advice in data movement and compute utilization from the below figure can guide you to tune performance. .. image:: nvvp3.png :align: center :scale: 30% .. image:: nvvp4.png :align: center :scale: 30% Profiling tips ============== - The :code:`nvprof` and :code:`nvvp` output is a very good place to start - The timeline is a good place to go next - Only dig deep into a kernel if it’s taking a significant amount of your time. - Where possible, try to match profiler output with theory. 1) For example, if I know I’m moving 1GB, and my kernel takes 10ms, I expect the profiler to report 100GB/s. 2) Discrepancies are likely to mean your application isn’t doing what you thought it was. - Know your hardware: If your GPU can do 6 TFLOPs, and you’re already doing 5.5 TFLOPs, you won’t go much faster! Profiling is a key step in optimisation. Sometimes quite simple changes can lead to big improvements in performance. Your mileage may vary! Reference ========= Jeremy Appleyard, `GPU Profiling for Deep Learning `_, 2015