Llama 3 8b hardware

What would it take to make 10k inferences per second of more than 2000 input tokens and 50 output tokens?

Data here may be of interest.