I am pretty new here. Just started coding in CUDA :rolleyes:
I have a few lower level questions that I am curious about:
How does the CPU assign work to GPU? Does it send an address of the instructions and data to GPU and later GPU fetches it? or Does it send the all the instructions and data at once?
If the full code is being sent, then what happened if the code i wrote exceeds the GPU memory size?
If the GPU only gets the addresses of the instructions then, where will it be fetching from?
Is the “pinned memory” used to store data and instructions or data only?
What are the places inside the GPU where the code resides?