Closed maxyanghu closed 9 months ago
did you measure compilation speedup?
did you measure compilation speedup?
32 core i913900K + 4090
didn't you notice how many candidates are in matmul and bert?
https://docs.google.com/spreadsheets/d/1uVBh1hCafwhOEeRdrFnuoaz4HOFECyr5cl7WnuMm7Ns/edit?usp=sharing I did a compilation time breakdown which shows the improvement this optimization has brought to the whole compliation time. This breakdown contains measurements of several common operators. I measured three parts of the compilation time:
a. 'lower' refers to the time spent on IR lowering (kernel level optimiztion b. 'codegen' refers to the time spent on converting IRModule to .cu source file c. 'compile' refers to the time spent on compiling .cu source file to object files
'compile_opt' is the optimized branch.
Useful data, thx @hunbssfy Could you make some clarifications
@vadiklyutiy
python/hidet/drivers/build_module.py
: build_ir_module
. if you look into the code in this function, you can easily find the three functions(lower/codegen/compile) I measured. I used python module 'time' to measure the function calling time. In the multiprocessing scenario, I use a shared value(with a lock) to accumulate all the time used. The time in the sheet is the total time summed from all the processor's individual time.build_ir_module
's signature, it can now take a sequence of IRModule
and do lower/codegen/compile in a batch. Previously, this function can only take an IRModule at a time. I assume there might be some overhead in the measurement itself and python interpreter. In the batch compilation case, we measure the time to lower/codegen/compile in a batch style while in the original code we measure them for each task.' 1. Thanks for clarification. How many CPU thread did you have on machine where you did measure? ' 3. Seems like packing in a batches impact also speed up "lower". For me a bit unexpected but possible.
BTW From you spreadsheet. It is clear that if we have more candidates we have bigger improvement. matmul, conv, attn have similar number of candidates. For matmul and conv improvement 8.05x and 7.22x. But attn has only 3.83x. Maybe it's worth to look into it. Maybe the reason just because attn has much more code in .cu file but maybe we have some misprint or didn't take into account something.
@vadiklyutiy I used 32 cores so that is 32 threads.
I also noticed this unbalanced acceleration percentage. I'll post new discoveries if I got any.
Thanks @hunbssfy @vadiklyutiy !
What if different IRModules have functions with the same name but different implementations? For example, we might have a function in the schedule template which will have different implementations for different hyper-parameters. When we merge the generated code for multiple IRModules, there might be name conflict issue.
Thanks @hunbssfy @vadiklyutiy !
What if different IRModules have functions with the same name but different implementations? For example, we might have a function in the schedule template which will have different implementations for different hyper-parameters. When we merge the generated code for multiple IRModules, there might be name conflict issue.
But as I understand finally they are linked to the lib.o. Does this issue exist in current implementation?
@yaoyaoding could you response
But as I understand finally they are linked to the lib.o. Does this issue exist in current implementation?
They are private functions in each lib.a
and the exposed public functions (e.g., those functions with API
macro at the head of function declaraion like launch_1
, launch_2
) have different names. Thus current implementation does not have this problem.
Thanks @hunbssfy @vadiklyutiy !
What if different IRModules have functions with the same name but different implementations? For example, we might have a function in the schedule template which will have different implementations for different hyper-parameters. When we merge the generated code for multiple IRModules, there might be name conflict issue.
@yaoyaoding I went through the generated .cu file during compilation. Here is one of them:
namespace candidate_0 {
static __device__ __forceinline__ void cuda_mma_sync_aligned_m16n8k8_row_col_f32_tf32_tf32_f32(tfloat32_t * __restrict__ a, tfloat32_t * __restrict__ b, float * __restrict__ c) {
...
}
static __global__ void __launch_bounds__(64) batch_matmul_kernel(float * __restrict__ a, float * __restrict__ b, float * __restrict__ c) {
...
}
void launch(float * __restrict__ a, float * __restrict__ b, float * __restrict__ c) {
...
}
}
namespace candidate_1 {
// two static 'internal' functions
// launch function
}
namespace candidate_2 {
...
}
....
As you can see, all candidates have a launch
function. But they dwell finely within separate namespaces. This patch didn't change the namespaces of the candidates which means even though candidates are 'grouped' into different .cu files, their namespace names still map to their initial candidate numbers. For candidate 1 that's namespace candidate_0. for candidate 700 that's namespace candidate_700. So there will be no namespace conflicts. I also looked into the generated .so file's symbols, here is what they looked like:
0000000000040c60 T _ZN11candidate_06launchEPfS0S0 000000000003eb60 T _ZN11candidate_16launchEPfS0S0 0000000000040000 T _ZN11candidate_26launchEPfS0S0 0000000000040420 T _ZN11candidate_36launchEPfS0S0 0000000000040840 T _ZN11candidate_46launchEPfS0S0 000000000003ef80 T _ZN11candidate_56launchEPfS0S0 0000000000041ce0 T _ZN11candidate_66launchEPfS0S0 000000000003fbe0 T _ZN11candidate_76launchEPfS0S0 000000000003f7c0 T _ZN11candidate_86launchEPfS0S0 000000000003f3a0 T _ZN11candidate_96launchEPfS0S0 0000000000043180 T _ZN12candidate_106launchEPfS0S0 0000000000041080 T _ZN12candidate_116launchEPfS0S0 000000000003e320 T _ZN12candidate_126launchEPfS0S0 0000000000042d60 T _ZN12candidate_136launchEPfS0S0 00000000000435a0 T _ZN12candidate_146launchEPfS0S0 0000000000042520 T _ZN12candidate_156launchEPfS0S0 000000000003e740 T _ZN12candidate_166launchEPfS0S0 00000000000414a0 T _ZN12candidate_176launchEPfS0S0 0000000000042940 T _ZN12candidate_186launchEPfS0S0 00000000000418c0 T _ZN12candidate_196launchEPfS0S0 0000000000042100 T _ZN12candidate_206launchEPfS0S0 00000000000478e0 T _ZN12candidate_216launchEPfS0S0 00000000000457e0 T _ZN12candidate_226launchEPfS0S0 0000000000046c80 T _ZN12candidate_236launchEPfS0S0 00000000000470a0 T _ZN12candidate_246launchEPfS0S0 00000000000474c0 T _ZN12candidate_256launchEPfS0S0 0000000000045c00 T _ZN12candidate_266launchEPfS0S0 0000000000048960 T _ZN12candidate_276launchEPfS0S0 0000000000046860 T _ZN12candidate_286launchEPfS0S0 0000000000046440 T _ZN12candidate_296launchEPfS0S0 0000000000046020 T _ZN12candidate_306launchEPfS0S0 ...
They look normal. So I don't think there will be a function name conflict issue.
But as I understand finally they are linked to the lib.o. Does this issue exist in current implementation?
They are private functions in each
lib.a
and the exposed public functions (e.g., those functions withAPI
macro at the head of function declaraion likelaunch_1
,launch_2
) have different names. Thus current implementation does not have this problem.
@yaoyaoding I inspected the final lib.so file. There are two types of APIs which are GLOBAL. The first one is the launch
function of each candidates:
3171: 00000000000cb5d0 1043 FUNC GLOBAL DEFAULT 14 candidate_460::launch(float, float, float*)
the second one is these hidet_launch functions:
3199: 00000000001190f0 9 FUNC GLOBAL DEFAULT 14 hidet_launch_229
I read all comments again and seems there is some misunderstanding. Max in his patch compiles together several candidates of the same operator. Hence all it happens for the same IRModule. He did not try to compile several ops (IRModules) together.
So, this
What if different IRModules have functions with the same name but different implementations? For example, we might have a function in the schedule template which will have different implementations for different hyper-parameters. When we merge the generated code for multiple IRModules, there might be name conflict issue.
cannot happen. Does it make sense?
Hence all it happens for the same IRModule. He did not try to compile several ops (IRModules) together.
Even the candidates are fot the same operator, they are represented by different IRModules.
The problem that I previously worried about will not happen in the case that all IRModules that will be batch compiled are coming from the same operator as we will set different namespace for different IRModules. I used namespace to avoid name conflict in the shared library when I group all the kernels in the same shared library (previously, they are in diffferent shared libraries). This alleviates the potential problem of function name conflict problem.
There is still potential name conflict problem if the batch compile API (i.e., build_ir_module
function) is used to compile IRModules with name conflict. As there is no such use case in hidet for now, we can proceed without dealing with this problem. But, @hunbssfy could you add some assertion in the function to make sure that there is no name conflict before we group (i.e., make sure each "namespace::function_name" is unique).
Hence all it happens for the same IRModule. He did not try to compile several ops (IRModules) together.
Even the candidates are fot the same operator, they are represented by different IRModules.
The problem that I previously worried about will not happen in the case that all IRModules that will be batch compiled are coming from the same operator as we will set different namespace for different IRModules. I used namespace to avoid name conflict in the shared library when I group all the kernels in the same shared library (previously, they are in diffferent shared libraries). This alleviates the potential problem of function name conflict problem.
There is still potential name conflict problem if the batch compile API (i.e.,
build_ir_module
function) is used to compile IRModules with name conflict. As there is no such use case in hidet for now, we can proceed without dealing with this problem. But, @hunbssfy could you add some assertion in the function to make sure that there is no name conflict before we group (i.e., make sure each "namespace::function_name" is unique).
Hi @yaoyaoding , I added some unqiueness function name check after the regrouping of IRModule
optimize the compilation behavior for each core. Instead of generating a cuda source file for each core, congragate them into a single source file to reduce compilation overhead.
Initial test reveals a 20% decrease in the total compilation+run time of bert model (222s -> 177s). Also the same percentage of decrease in operator tests.