ZeroTouch demonstrated at human-computer interaction conference

The Texas A&M Interface Ecology Lab introduced the latest generation of the high fidelity ZeroTouch Sensor for multitouch and free-air interaction at the interactivity exhibition of the ACM SIGCHI Conference on Human Factors in Computing Systems (CHI 2012) in Austin.

The Interface Ecology Lab is led by Dr. Andruid Kerne in the Graduate student Jon Moeller with the ZeroTouch technologyDepartment of Computer Science and Engineering.(Pictured: Graduate student Jon Moeller with the ZeroTouch technology)

ZeroTouch is a disruptive technology that brings the body into human-centered computing, enabling new forms of expressive, yet precise and powerful interactivity. ZeroTouch turns any display or any area into a multitouch surface. The latest ZeroTouch sensor’s data transfer rate is eight times faster than before, enabling construction of sensors of any practical size.

ZeroTouch is the highest performance and potentially most cost effective large-scale multitouch sensor currently available. When integrated into a 42-inch television, ZeroTouch is expected to add as little as $500 to cost at mass production introduction. Meanwhile, development kits are available to a limited number of early adopting researchers at somewhat higher costs.

ZeroTouch has been developed with support from the Human-Centered Computing program of the National Science Foundation.

The Interface Ecology Lab exhibit features one ZeroTouch integrated with a 55-inch LCD, heralding a new era of responsive televisions. Also featured are three ZeroTouch sensors for eSports competition, two of them integrated with pen-based computing, for Pen-In-Hand Command. Another pen + multi-touch sensor drives Embodied InfoComposer, a single display groupware for helping people curate more connected and rich collections than those afforded by Pintrest. Zooming Bookmarker uses ZeroTouch to bring multitouch into Google’s Chrome browser. The free-air interaction of intangibleCanvas, reminiscent of the technology seen in Minority Report and Iron Man, invites attendees to collaborate in a painting space.

Pen-In-Hand Command provides a more direct and entertaining means of interaction for real time strategy games in the vein of Starcraft. eSports spectators see the players gesture and the action unfold. As pairs of participants go head to head in a competitive game of skill, commentators can use ZeroTouch to change the view, with greater ease than live sports cameramen.

Ph.D. student researcher and developer Bill Hamilton said, “New fluid means to manipulate and control the battlefield, enabling more precise and complex control than ever before.”

Embodied InfoComposer enables more expressive and natural interactions for exploring and organizing information collections of rich bookmarks. Participants use fluid pen and multitouch interactions to compose visual overviews that express and develop relationships, supporting creativity. To mitigate the inevitable information overload as collections grow large, Embodied InfoComposer enables grouping rich bookmarks into piles that can be collapsed or fanned out. A Microsoft Kinect tracks the participant’s hands and stylus, distinguishing touches performed by the dominant and non-dominant hands. Using the non-dominant hand as a reference frame, like positioning paper for the dominant hand, exciting new natural forms of pen + multitouch interaction are made possible.

ZoomBrowser utilizes rich bookmarks and a zoomable bookmarking space to transform Google Chrome into a fluid multi-touch browser. Web surfers see their active web pages represented as rich bookmarks in an infinite, zoomable space. As you zoom out, an individual web page transitions fluidly into a series of overviews, reminiscent of the Eames’ brothers’ Powers of Ten. A space connecting rich bookmarks for all tabs brings new cohesion to the browsing experience.

intangibleCanvas uses the ZeroTouch sensor as a precision free-air interaction input modality, allowing users to reach through the sensor and paint on a projected screen. The embodied interaction enables painting with the elbows, the arms, and the head as well as the fingers. intangibleCanvas affords control over brush style, color, and ink flow through a multi-modal phone interface held in the non-dominant hand. intangibleCanvas 2.0 uses a visible laser plane to overcome free-air interaction’s common problem of invisible activation. Participants can take their paintings home by email, and publish to a shared online space.

For more information, email