GPU out of memory when the total ram usage is 2.8G

Hi,

The fix is available now.
Please check JetPack SDK | NVIDIA Developer

Hi,

JetPack 3.1 really can sovle this problem.

Thanks a lot.

Hi everyone,
Im trying to use my Jetson Tx2 with Jetpack 3.2 gpu with tensorflow wihch works but Its too slow and I got the following errors:

018-04-26 13:14:40.069245: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.53GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.078779: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 442.69MiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.091886: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 798.28MiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.110919: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.04GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.152810: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.07GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.331434: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.07GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.350748: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.07GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.408377: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 2.13GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.601111: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.13GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.
2018-04-26 13:14:40.634236: W tensorflow/core/common_runtime/bfc_allocator.cc:219] Allocator (GPU_0_bfc) ran out of memory trying to allocate 1.14GiB. The caller indicates that this is not a failure, but may mean that there could be performance gains if more memory were available.

I sees there is RAM issue. Can I solve this with allocating swap like what you suggested?

Hi,

Please try the following configuration:

config = tf.ConfigProto()
config.gpu_options.allow_growth = True

session = tf.Session(config=config, ...)

If the error keeps, please share your tegrastats results with us.

sudo ./tegrastats

Thanks.

  • I am using JetPack3.2 and CUDA9

  • I try to deploy a Lightweight Mask RCNN Model on the Jetson.
    On my host machine (with 4gb GPU mem) takes the model <2GB memory.
    So a lot less then the 8GB available on the Jetson.

Still it gives me memory errors stating that too many resources were requested to launch.

Hi,

It’s recommended to apply some experiment to figure out the real cause first.
Could you test the maximal accessible memory chunk with CUDA API directly?
[url]https://devtalk.nvidia.com/default/topic/1013464/jetson-tx2/gpu-out-of-memory-when-the-total-ram-usage-is-2-8g/post/5170376/#5170376[/url]

We can allocate around 7.xGb memory on TX2.

If you also can allocate 7Gb memory but hit out of memory issue with TensorFlow API.
This limitation may come from TensorFlow implementation rather than Jetson system.

Thanks.

is there any fix one this?

Hi, civilman

What is your error?
This is a pretty old issue and is already fixed in JetPack 3.1.

The original bug is that CUDA cannot allocate more than half of the physical memory since we separate it for CPU and GPU.
This issue won’t occur in our latest JetPack3.3.

Thanks.