Content deleted Content added
AMD's upcoming hUMA architecture |
Guy Harris (talk | contribs) →hUMA: Use a citation template. |
||
(38 intermediate revisions by 30 users not shown) | |||
Line 1:
{{Short description|Parallel computing memory architecture}}
There are three types of UMA architectures:
==hUMA==
In April 2013, the term hUMA (''heterogeneous uniform memory access'') began to appear in [[AMD]] promotional material to refer to [[CPU]] and [[GPU]] sharing the same system memory via [[Cache coherence|cache coherent]] views. Advantages include an easier programming model and less copying of data between separate memory pools.<ref>{{cite web |author=Peter Bright |url=https://arstechnica.com/information-technology/2013/04/amds-heterogeneous-uniform-memory-access-coming-this-year-in-kaveri/ |title=AMD's "heterogeneous Uniform Memory Access" coming this year in Kaveri |website=[[Ars Technica]] |date=April 30, 2013}}</ref>
▲In the UMA architecture, each processor may use a private cache. Peripherals are also shared in some fashion, The UMA model is suitable for general purpose and time sharing applications by multiple users. It can be used to speed up the execution of a single large program in time critical applications.<ref>Advanced Computer Architecture, Kai Hwang, ISBN 0-07-113342-9</ref>
▲==Types of UMA architectures==
▲# UMA using bus-based [[Symmetric Multi-Processing]] (SMP) architectures
▲# UMA using crossbar switches
▲# UMA using [[Multistage interconnection networks]]
==See also==
* [[Non-
* [[Cache-
* [[Heterogeneous System Architecture]]
==
{{reflist}}
{{Parallel computing}}
[[Category:Computer memory]]
[[Category:Parallel computing]]
|