I’m trying to do a comparison between running a Gaussian Blur algorithm with CUDA and OpenCV. I’m using Visual Studio 2015 and CUDA Toolkit V8 to run things.
My GPU is a NVidia GeForce 840M and my CPU is a Intel Core i7-4510U @ 2Ghz. I’m also having 16GB of RAM and 512 GB SSD, but I doubt this is useful information. Yes I am running Visual Studio on a laptop. Main OS is Windows 10.
I tried doing a comparison today between OpenCV and CUDA and it stuck me that the OpenCV program ran in 300 milliseconds while the CUDA one ran in 1500 milliseconds.
I added counters to strictly before/after each filter. Should it be because of the time data is transferred between CPU/GPU?
Here are the algorithms I took as reference for this comparison: