Surface computing: Difference between revisions

Content deleted Content added
Included more information regarding the uses of surface computing
GidSlHs (talk | contribs)
Added Surface Types and Technological Components sections
Line 2:
 
Early work in this area was done at the University of Toronto, Alias Research, and MIT.<ref>http://www.billbuxton.com/multitouchOverview.html</ref> Surface work has included customized solutions from vendors such as [http://LM3LABS.com LM3LABS] or GestureTek, Applied Minds for [[Northrop Grumman]].<ref>[http://www.is.northropgrumman.com/products/touchtable/index.html Applied Minds & Northrop Grumman]</ref> Major computer vendor platforms are in various stages of release: the iTable by PQLabs,<ref>[http://multi-touch-screen.net/ PQLabs]</ref> [[Linux MPX]],<ref>[http://gizmodo.com/gadgets/touch-me/linux-mpx-multi+touch-table-may-become-free-diy-microsoft-surface-one-day-278613.php Linux MPX Multitouch]</ref> the [[Ideum]] [[MT-50 Multitouch Table|MT-50]], interactive bar by spinTOUCH, and [[Pixelsense|Microsoft PixelSense]] (formerly known as Microsoft Surface).
 
== Surface Types ==
 
==== Flat ====
Flat surface types refer to two-dimensional surfaces such as tabletops. This is the most common form of Surface Computing in the commercial space as seen by products like Microsoft’s PixelSense and  iTable. The aforementioned commercial products utilize a multi-touch LCD screen as a display, but other implementations use projectors. Part of the appeal of two-dimensional surface computing is the ease and reliability of interaction. Since the advent of tablet computing, a set of intuitive gestural interactions have been developed to compliment two-dimensional surfaces. However, the two-dimensional plane limits the range of interactions a user is able to perform. Furthermore, interactions are only detected when making direct contact with the surface. In order to afford the user a wider range of interaction, research has been done to augment the interaction schemes for two-dimensional surfaces. This research involves using the space above the screen as another dimension for interaction, so, for example, the height of a user’s hands above the surface becomes a meaningful distinction for interaction. This particular system would qualify as a hybrid that uses a flat surface, but a three-dimensional space for interaction. <ref name=":0">{{Cite web|url = http://research.microsoft.com/pubs/132463/hilliges2009uist.pdf|title = Interaction in the Air: Adding Further Depth to Interactive Tabletops|date = |accessdate = |website = |publisher = |last = |first = }}</ref>
 
==== Non-Flat ====
While most work with Surface Computing has been done with flat surfaces, non-flat surfaces have become an interest with researchers. The eventual goal of Surface Computing itself is tied to the notion of [[Ubiquitous Computing]] “where everyday surfaces in our environment are made interactive”.<ref name=":1" /> These everyday surfaces are often non-flat, so researchers have begun exploring curved and three-dimensional modes. Some of these include spherical, cylindrical and parabolic surfaces. Including a third dimension to Surface Computing presents both benefits and challenges. One of these benefits is an extra dimension of interaction. Unlike flat surfaces, three dimensional surfaces allow for a sense of depth and are thus classified as “depth-aware” surfaces. This allows for more diverse gestural interactions. However, one of the main challenges  is designing intuitive gestural actions to facilitate interaction with these non-flat surfaces. Furthermore, three dimensional shapes such as spheres and cylinders require viewing from all angles, also known as Omnidirectional displays. Designing compelling views from every angle is a difficult task, as is designing applications that make sense for these display types.<ref name=":1" />
 
==Technological Components==
 
==== Display ====
Displays for Surface Computing can range from [[LCD]] and [[Projection]] screens to physical object surfaces. Displays can be divided into single-viewpoint and multi-viewpoint displays. Single-viewpoints include any flat screen or surface where viewing is typically done from one angle. A multi-viewpoint display would include any three-dimensional object surface like a sphere or cylinder that allows viewing from any angle. <ref name=":0" />
 
==== Projectors ====
If a projection screen or a physical object surface is being used, a projector is needed to superimpose the image on the display. A wide range of projectors are used including DLP, LCD, and LED. Front and rear projection techniques are also utilized. The advantage of a projector is that it can project onto any arbitrary surface. However, a user will end up casting shadows onto the display itself, making it harder to identify high detail.
 
==== Infrared Cameras ====
Infrared or [[Thermographic]] cameras are used to facilitate gestural detection. Unlike digital cameras, infrared cameras operate independent of light, instead relying on the heat signature of an object. This is beneficial because it allows for gesture detection in all lighting conditions. However, cameras are victim to occlusion by other objects that may result in a loss of gesture tracking.
 
==Interaction Methods==
Line 10 ⟶ 27:
•'''Two Dimensional''' Typically, traditional surface types are two-dimensional and only require two-dimensional touch interactions. Depending on the system, multi-touch gestures, such as [[pinch to zoom]], are supported. These gestures allow the user to manipulate what they see on the surface by physically touching it and moving their fingers them across the surface.
 
•'''Three Dimensional''' Using depth aware cameras it is possible to make three dimensional gestures. Such gestures allow the user to move in three dimensions of space without having to come into contact with the surface itself, such as the methods used in [[DepthTouch]].<ref name=":1">[http://research.microsoft.com/en-us/um/people/benko/publications/2009/benko_mm09bnt236.pdf Beyond Flat Surface Computing: Challenges of Depth-Aware and Curved Interfaces]</ref>
==Applications==