Just a quick question here. Has anyone written up an on-card OSS licensed Malloc/free routine for the GPU?
This library could be built using cudaMalloc to initially fill a pool of memory on the video card, then doll it out to threads/blocks on request. I might find myself needing to write a library that does this in the near-future and I was wondering if anyone else has gone down this path as well.
Also a question for Nvidia: Is an on-card memory allocation library forthcoming in a future release of CUDA? I don’t want to wind up duplicating work that is going to be released as part of CUDA in the future.