Optimizing Energy Consumption by Adjusting GPU Frequency

Hello everyone,

I’m currently exploring an energy optimization problem but find myself unsure how to proceed effectively.

To reduce the energy consumption of a memory-bound task, one idea is to lower the computational frequency (or computational capability) of the device. This approach would decrease power consumption while having a minimal effect on latency, thereby reducing overall energy usage (E = P × t).

Does this idea hold merit?

If so, how can we determine the optimal trade-off point? In cases where power cannot be directly measured, could specific PMU events or metrics help identify the transition point between memory-bound and compute-bound as the computational frequency decreases? For instance, in CUDA, metrics such as smsp__average_warps_issue_stalled_long_scoreboard_per_issue_active.ratio and smsp__average_warps_issue_stalled_math_pipe_throttle_per_issue_active.ratio might be relevant.

Additionally, could a similar strategy be applied to CPUs? Alternatively, for compute-bound tasks, would reducing the EMC frequency (memory bandwidth) be an effective approach?

I would greatly appreciate any insights or guidance on this topic.

Thank you in advance for your help!

Best regards,
yc-ruan

Hi,
We have the default power modes for each Orin platform. Please check

Jetson Orin Nano Series, Jetson Orin NX Series and Jetson AGX Orin Series — NVIDIA Jetson Linux Developer Guide 1 documentation

And you can use power estimator tool to customize one set fitting your use-case. Please check it and try.

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.