We are evaluating NVIDIA GPUs for a production VOD transcoding system focused on H.264 and H.265 (HEVC) at 1080p resolution.
The primary requirement is high-density concurrent NVENC throughput, rather than GPU compute or AI workloads.
We are currently comparing NVIDIA A16 (high NVENC density) with NVIDIA L40S (fewer NVENC engines but higher compute and memory bandwidth).
While we understand NVIDIA does not publish fixed FPS benchmarks, we would appreciate architectural or practical guidance from engineers or users with real-world experience:
-
Can L40S reasonably substitute A16 for high-density 1080p H.264/H.265 VOD transcoding?
-
In practice, how does aggregate NVENC throughput of L40S compare to A16 under such workloads?
-
Are there specific scenarios where L40S becomes a better choice despite lower NVENC density?
Any insights, high-level comparisons, or real-world observations would be very helpful.