I just started playing with GPUs. For the most part I need to do linear operations and have coded up much of the stuff using CUBLAS. In one case I need to loop over an entire matrix in device memory and compute exp(x) for each element. Of course this can be done in parallel. What’s an easy way to perform this operation?? Is there a one liner or some sample code??