Single instruction, multiple thread (SIMT) is an "execution model" and abstraction on top of the hardware paradigm, single instruction, multiple data, introduced by Nvidia:[1][2]
[The G80 Nvidia GPU architecture] introduced the single-instruction multiple-thread (SIMT) execution model where multiple independent threads execute concurrently using a single instruction.
SIMT is intended to limit instruction fetching overhead,[3] and is used in modern GPUs in combination with 'latency hiding' to enable high-performance execution despite considerable latency in memory-access operations.[4]
See also
References
- ^ "Nvidia Fermi Compute Arcitecture Whitepaper" (PDF). http://www.nvidia.com/. NVIDIA Corporation. 2009. Retrieved 2014-07-17.
{{cite web}}
: External link in
(help)|website=
- ^ "NVIDIA Tesla: A Unified Graphics and Computing Architecture". http://www.ieee.org/. IEEE. 2008. p. 6 (Subscription required.). Retrieved 2014-08-07.
{{cite web}}
: External link in
(help)|website=
- ^ Rul, Sean; Vandierendonck, Hans; D’Haene, Joris; De Bosschere, Koen (2010). An experimental study on performance portability of OpenCL kernels. Symp. Application Accelerators in High Performance Computing (SAAHPC).
- ^ "Advanced Topics in CUDA" (PDF). cc.gatech.edu. 2011. Retrieved 2014-08-28.