Benchmarking LLMs on AI-Generated CUDA Code with ComputeEval 2025.2

Originally published at: Benchmarking LLMs on AI-Generated CUDA Code with ComputeEval 2025.2 | NVIDIA Technical Blog

Can AI coding assistants write efficient CUDA code? To help measure and improve their capabilities, we created ComputeEval, a robust, open source benchmark for evaluating AI models and agents on CUDA programming tasks.  A few months ago, we announced the first release of ComputeEval and today, we’re introducing its first major expansion by adding more…