I have looked around and cannot find a definitive lamers simplified comparison on AI specific cards benchmarks. By AI I mean RTX cards that can executed AI specific jobs like running Ollama models, Yolo/Darknet etc and how can these cards be compared when executing such workloads.
It would be great for instance if one could understands, again in the simplest terms possible:
why Nvidia releases new generation cards ie and how they perform… ex. RTX 1060 vs 2060 vs 3060 vs 4060 … 2070 vs 3070 vs … etc
how these performs when using a relatively basic model/workload ie a simple ollama model job
how big of a job/model etc they can perform and how fast, for instance card XYZ can run ollama deepseek-r1:latest and get an answer in 3 seconds … vs deepseek-r1:70b and get an answer in 10 seconds or other parameters if possible.
i have previously looked at sites such as RTX 5070 vs RTX 2070 [6-Benchmark Showdown] but I was told by peers on IRC Libera #hardware that these benchmarks are not reliable. Is there something Nvidia provides to give such comparision before evaluating a card purchase?