'''Uniform memory access''' ('''UMA''') is a [[shared-memory architecture|shared memory]] architecture used in [[parallel computer]]s. All the processors in the UMA model share the physical memory uniformly. In aan UMA architecture, access time to a memory ___location is independent of which processor makes the request or which memory chip contains the transferred data. Uniform memory access computer architectures are often contrasted with [[non-uniform memory access]] (NUMA) architectures. In the UMANUMA architecture, each processor may use a private cache. Peripherals are also shared in some fashion. The UMA model is suitable for general purpose and [[time sharing]] applications by multiple users. It can be used to speed up the execution of a single large program in [[real-time computing|time-critical]] applications.<ref>{{cite book |title=Advanced Computer Architecture, |author=Kai Hwang, ISBN |isbn=0-07-113342-9}}</ref>
==Types of architectures==
Line 8 ⟶ 9:
==hUMA==
In April 2013, the term hUMA (''heterogeneous uniform memory access'') began to appear in [[AMD]] promotional material to refer to [[CPU]] and [[GPU]] sharing the same system memory via [[Cache coherence|cache coherent]] views. Advantages include an easier programming model and less copying of data between separate memory pools.<ref>{{cite web |author=Peter Bright.[http|url=https://arstechnica.com/information-technology/2013/04/amds-heterogeneous-uniform-memory-access-coming-this-year-in-kaveri/ |title=AMD's "heterogeneous Uniform Memory Access" coming this year in Kaveri], |website=[[Ars Technica,]] |date=April 30, 2013.}}</ref>