Difference between revisions of "Talk:GPU"
(→embedded CPU controller) |
(→Nvidia architectures: new section) |
||
Line 1: | Line 1: | ||
+ | == Nvidia architectures == | ||
+ | Afaik Nvidia did never official mention SIMD in their papers as hardware architecture, with Tesla they only referred to as SIMT. | ||
+ | |||
+ | Nevertheless, my own conclusions are: | ||
+ | |||
+ | Tesla has 8 wide SIMD, executing a Warp of 32 threads over 4 cycles. | ||
+ | |||
+ | Fermi has 16 wide SIMD, executing a Warp of 32 threads over 2 cycles. | ||
+ | |||
+ | Kepler is somehow odd, not sure how the compute units are partitioned. | ||
+ | |||
+ | Maxwell and Pascal have 32 wide SIMD, executing a Warp of 32 threads over 1 cycle. | ||
+ | |||
+ | Volta and Turing seem to have 16 wide FPU SIMDs, but my own experiments show 32 wide VALU. |
Revision as of 11:45, 18 April 2021
Nvidia architectures
Afaik Nvidia did never official mention SIMD in their papers as hardware architecture, with Tesla they only referred to as SIMT.
Nevertheless, my own conclusions are:
Tesla has 8 wide SIMD, executing a Warp of 32 threads over 4 cycles.
Fermi has 16 wide SIMD, executing a Warp of 32 threads over 2 cycles.
Kepler is somehow odd, not sure how the compute units are partitioned.
Maxwell and Pascal have 32 wide SIMD, executing a Warp of 32 threads over 1 cycle.
Volta and Turing seem to have 16 wide FPU SIMDs, but my own experiments show 32 wide VALU.