With Nvidia’s Volta architecture having been skipped on the consumer GeForce front, only finding its way into the prosumer-grade Titan V, workstation-grade Quadro GV100, and the Xavier generation of Tegra SoCs that focus on self-driving cars, the next green team architecture gamers can expect to use after Pascal is Turing, which will be powering the recently announced GeForce RTX 2000 series of graphics cards.
Today, we’re able at last to bring you full details of the Turing architecture and the specifications of the RTX 2080 Ti, RTX 2080, and RTX 2070, although performance data is still bound by NDA. As a reminder, the RTX 2080 Ti and RTX 2080 launch this month, while the RTX 2070 isn’t launching until October.
As you’ll see, there’s plenty to cover; this is a summary of information provided by Nvidia in an 83-page whitepaper and a full day of press-only presentations, so you may want to grab a cup of [insert caffeinated beverage of choice] before getting into the nitty gritty. However, we reckon a good place to start is with the full specifications of the three upcoming cards. After all, who doesn’t love a bloody good spec table?
(Note that Nvidia Founders Edition card specifications will differ slightly to these reference specs on account of their higher boost clocks, which are used to inform the TFLOPS, TIPS, RTX-OPS, and TDP calculations.)
|Nvidia GeForce RTX 2080 Ti||Nvidia GeForce GTX 1080 Ti||Nvidia GeForce RTX 2080||Nvidia GeForce GTX 1080||Nvidia GeForce RTX 2070||Nvidia GeForce GTX 1070|
|Layout||6 GPCs, 68 SMs||6 GPCs, 28 SMs||6 GPCs, 46 SMs||4 GPCs, 20 SMs||3 GPCs, 36 SMs||3 GPCs, 15 SMs|
|Peak TFLOPS (FP32)||13.4||11.3||10||8.9||7.5||6.5|
|Peak TIPS (INT32)||13.4||N/A||10||N/A||7.5||N/A|
|Peak FP16 Tensor TFLOPS (FP16 Accumulate)||107.6||N/A||80.5||N/A||59.7||N/A|
|Transistors||18.6 billion||12 billion||13.6 billion||7.2 billion||10.8 billion||7.2 billion|
|Process||12nm FFN||16nm||12nm FFN||16nm||12nm FFN||16nm|
|Memory||11GB GDDR6||11GB GDDR5X||8GB GDDR6||8GB GDDR5X||8GB GDDR6||8GB GDDR5|
|Memory Data Rate||14Gbps||11Gbps||14Gbps||10Gbps||14Gbps||8Gbps|
We have included the relevant Pascal parts for reference, but comparing raw specs across different architectures rarely leads to a great deal of insight, and Turing is a pretty big change relative to Pascal. While Nvidia is still using CUDA cores for classic floating point shader workloads, the addition of a separate and independent datapath for integer math means the Turing Streaming Multiprocessor is now very different in design. You’ll also notice a move from GDDR5 and GDDR5X memory to GDDR6, which has resulted in a redesign of Turing’s memory system. Add to this the dedicated hardware for AI-related workloads (Tensor Cores) and ray tracing workloads (RT Cores), and you end up with a vastly different family of GPUs, as evidenced by the number of ‘N/A’ entries in the spec table for Pascal parts.
Additional new hardware features include updated display and video engines along with VirtualLink support and a move to NVLink for multi-GPU configurations. Add to that new advanced shading techniques which Nvidia is leveraging to boost performance and give developers new flexibility, and you start to see why Nvidia refers to Turing as ‘the biggest architectural leap forward in over a decade’. How it actually performs in games remains to be seen, but for now we’ll be digging into exactly what informs that claim.
Just quickly, you may have seen and been confused by the new term RTX-OPS. This is a term invented by Nvidia, but there is logic in how the figure is derived. It simply serves to more accurately compare the differences between (and only between) Nvidia RTX GPUs when truly taking into account the new capabilities of their GPUs (based on how much work each part does in generating a “typical” frame) in a way that the classic TFLOPS figure simply misses. There’s maths behind it, but it’s not especially interesting since it’s really just a marketing/categorisation tool, and at least you know now what its purpose is.
October 14 2021 | 15:04