While training a Neural Network model on a Jetson Orin module, I observed that the energy consumption during the first epoch is noticeably higher compared to subsequent epochs. Here’s in below table from my training logs, indicating Time_total and Energy_total:
This pattern emerges despite a constant model architecture and training methodology. I’m using Python, PyTorch, and operating on a Linux Ubuntu system.
Could the community shed light on why this might be happening? Is it a common occurrence or specific to neural networks? Any insights would be greatly appreciated.