Even though a short one, a short-circuit of multiple logical AND/OR/XOR/etc operations in same line causes warp divergence right?
bool a, b, c, d, e;
init(a,b,c,d, warpLane);
// warp lanes 0,...3: short circuits at a
// warp lanes 4,...7: short circuits at b
// warp lanes 8,...15: short circuits at c
// warp lanes 16,...31: short circuits at d
bool something = a && b && c && d && e;
This has 4-way divergence right?
Is it enough to just convert it to bitwise operations to evade the warp divergence? Does it have any side effects or performance penalties under some cases?
unsigned int a, b, c, d;
// initialize different per warp lane
init(a,b,c,d, warpLane);
unsigned int something = a & b & c & d & e; // assuming its faster than a * b * c * d * e