I have a little problem. Let me explain briefly and sorry for my english:
I have an app (APP2) developed in CUDA that uses some data (stored in an array in global memory).
This data is from an external hardware and I’m able to copy it in GPU memory using other app (APP1). Both programs works fine separately.
But the question is how tell the other program where the data is located, there is a way to do this? There exist some function that allow me to read from an specific address of GPU memory? Because the data is refreshing every time and I have to manipulate this and show it at the screen. So the rule maybe will be:
APP1: Capture data and copy to device memory (forever loop)
APP2: Get data from the same location in device memory and make things…
How can I copy and get from the same location of memory? Do you know how to achieve this? Using one thread for each app and sharing a global variable?
I have no idea and I’m getting crazy :-P
Thank you very much in advance!