Changes

Jump to: navigation, search

GPU

145 bytes removed, 08:44, 20 April 2021
Floating Point Instruction Throughput
* FP64
: Consumer GPUs have in general a lower ratio (FP32:FP64) for double-precision (64 bit) floating point operations than server brand GPUs, like 4:1 down to 32:1 compared to 2:1 to 4:1.
* FP16
: Some GPGPU architectures offer half-precision (16 bit) floating point operation throughput with an FP32:FP16 ratio of 1:2. Older architectures migth not support FP16 at all, at the same rate as FP32, or at very low rates.
==Tensors==
422
edits

Navigation menu