Colorado Springs, CO.-based independent software developer Kevin Connolly became a minor YouTube celebrity over the past several months with the demonstration of his natural user interface (NUI) hack of Microsoft's Kinect software development kit (SDK). Connolly demonstrated moving different images on a small bank of screens up and down, in and out, and sorted through three-dimensional image arrays, all through gesture alone. It was, he notes, very similar to the image manipulation featured in Steven Spielberg's 2002 futuristic film, Minority Report.
"I'm just some guy," Connolly says. "I made that work in a matter of hours. Imagine what we have the technology to do if one guy in his apartment can do that in a few hours."
Indeed, Microsoft's decision to release the Kinect SDK in June has garnered much attention in the technical and technology trade press and among researchers and enthusiasts. Anoop Gupta, a Microsoft distinguished scientist, says more than 100,000 individuals downloaded the SDK in the first six weeks after its release. However, the terms of the release forbid any commercial use of the SDK, and Connolly says he halted work on his nascent NUI after he got it working to his satisfaction. Yet, the release of the SDK was a signal that low-cost motion and depth-sensing technology may soon herald epochal changes in the way humans and computers interact. The Kinect hardware, for instance, is manufactured by Tel Aviv-based PrimeSense, and lists for about $200. Researchers in numerous disciplines call such technology for such a price absolutely revolutionary.
John Underkoffler's experience with NUI research goes far beyond hobbyist SDKs. In fact, Underkoffler is the man behind the Minority Report interface and has commercialized it via Los Angeles-based Oblong Industries, at which he is chief scientist. Underkoffler calls the Microsoft SDK release a "rhetorical event we all love." He says these events, like the 2006 release of Nintendo's Wii, "puts in the foreground for different sets of eyesthe end consumer for the Wii, the home or dorm room hacker for the Kinect SDKthe idea that it isn't going to be mouse and keyboard forever. We've seen the dialogue go from one of doubt or questioning to a kind of acceptance. Everyone now knows it isn't going to be mouse and keyboard forever, but the real question is, What is it going to be?"
Some of the next steps of NUI research are likely to be through gaming applications, which are increasingly dependent upon motion-sensing technology. Recently, four California Institute of Technology students undertook a study of motion-sensing technologies originating in gaming, and their possible uses in other computationally intensive fields. The students explored the depth camera technology in Kinect, the inertial sensors of the Wii, and electromagnetic sensing technology developed by Sixense Entertainment and Razer. One of the students, Peter Ngo, believes the idea of the NUI as a fully toolless interface is overstated, as does Amir Rubin, CEO of Sixense.
"People don't buy motion control," Rubin says. "They don't buy PCs. They don't buy consoles. They buy the experience being delivered to them. The best input device is something you don't remember is on you, and that is true even for the slogan sold by Microsoft'You are the controller.' I agree, but if I have to alter the way I use my body in order to interact with Kinect, then it's the worst controller."
Even the most ardent NUI advocates agree with Rubin. Oblong's Underkoffler, for instance, says that writers will likely be well served by the keyboard for the foreseeable future, but that those who design ship hulls or airplane wings would be better served by three-dimensional NUIs. Moreover, he says it is vital not to graft a notion of a new interface design by simply extending two-dimensional GUI concepts onto the prototypes of three-dimensional applications. These applications will need computational capabilities along not just the flat x and y axesan example might be a wall-sized but still two-dimensional application for designing the very three-dimensional ship's hull he mentionedbut will also need to compute the depth of the z axis. Oblong's g-speak platform, based on work Underkoffler pioneered in the 1990s at the Massachusetts Institute of Technology's Media Lab, computes this spatial environment via networked computers and screens that allow rich three-dimensional interaction. Ultimately, Underkoffler thinks a hybrid UI ecosystem will evolve.
"We're not out to replace the keyboard; let it do what it's best at," he says, "but when it comes to designing airplane wings, you do need a spatial UI. So, it's about situating the right activities in the right interaction zone."
Since the Microsoft SDK precludes commercial use, many early academic and enterprise projects using Prime-Sense and/or stripped down Kinect hardware use either homebrewed algorithms or open source drivers and middleware released by consortia such as OpenCV or OpenNI, the natural interface forge formed in November 2010, by PrimeSense and robotics pioneer Willow Garage. OpenNI leverages the PrimeSense depth-sensing technology, which is processed in parallel by PrimeSense's system-on-a-chip processor after receiving coded near-infrared light from its partnered CMOS sensor.
OpenNI supplies a set of APIs to be implemented by the sensor devices, and a set of APIs to be implemented by the middleware components. Thus, OpenNI's API enables applications to be written and ported to operate on top of different middleware modules. It also enables middleware developers to write algorithms on top of raw data formats, regardless of which sensor device has produced them, and offers sensor manufacturers the capability to build sensors that power any OpenNI-compliant application.
"People don't buy motion control," Amir Rubin says. "They buy the experience being delivered to them. The best input device is something you don't remember is on you."
In fact, one nascent healthcare industry application partially built on open source stacks by a team of surgeons and engineers at Sunnybrook Health Sciences Center in Toronto for the Kinect camera is already drawing attention.
Allowing surgeons access to medical images while not having to touch a controllerand thereby saving them the necessity to re-scrub in order to preserve sterility around the patientis an early enterprise triumph for the NUI concept. Computer vision specialist Jamie Tremaine says the gesture-based UI he and his colleagues developed has proven exceptionally robust and enables surgeons to view through MRI and CT scan samples that can run from 4,000 to 10,000 slides without ever having to re-scrub.
For such an application, the hand and arm gestures recognized by the Kinect camera are suitable, but Tremaine says "a lot of the work we've done hasn't even been on the technical side as much as creating gestures in the operating room that allow very fine-grained control, but which have to be larger."
Another NUI developer, Evan Lang of Seattle-based UI design firm IdentityMine, says his work with trying to develop Kinect NUIs (on PrimeSense drivers) similar to current GUI commands revealed vexing user issues. In developing a Web button, for instance, Lang says, "I programmed it to recognize a poking gesture, where you move your hand quickly forward and quickly back. When I got some test users to try it out, and said 'Poke it or press it,' everybody had a very different idea of what that actually meant. Some did a kind of poking thing. Other people moved their hand forward but wouldn't move it back, and others, who were very cautious and deliberate about it, the machine wouldn't register as a poke."
Oblong's Underkoffler says problems such as Lang encountered are emblematic of grafting current GUI-based mechanics on an idea that needs something else.
"We believe it's not appropriate to start talking about NUIs until you have a complete solution," Underkoffler says. "If you flash back 30 years, it's like dropping an early prototype of a mouse in everybody's lap and saying, 'We have a new interface.' You don't, because you've just got a new input device. So, really it's a full loop proposition. What's the input modality? What shows up on screen? What's the analogue of the windows and scroll bars and radio buttons? Until that's not only been answered in a way to allow real work to happen, but has become kind of a standard, and more in the cognitive sense, recognizably and pervasively present, then you don't have a new interface."
Sixense's Rubin predicts the next-generation standard UI device would not be a question of which technology is most elegant, but rather, that which meets three criteria: a consumer-friendly price, an intuitive UI design, and ease of software development on top of that device.
"If you can meet the combination of those three," he says, "then you will have the next-generation standard of input devices."
Robotics researchers such as Nicholas Roy, associate professor of aeronautics at the Massachusetts Institute of Technology, were among the first to adopt the PrimeSense sensor, and it may be their work that shows the longest-range potential for, and a new concept of, what a NUI will be.
Allowing surgeons access to medical images while not having to touch a controllerand thereby saving them the need to re-scrub in order to preserve sterility around the patientis an early enterprise triumph for the NUI concept.
Vehicles equipped with the three-dimensional sensorsamong them a helicopter Roy and his students programmedgain what Roy calls a "whole new sense of the human-centered environment," and are able to sense things such as drop-offs in stairs, table legs, and so on. And now, with more researchers exploiting the extremely cheap sensor technology, it is likely that more UI work will be done exploring how a robot, or any other computer, will "think" its way to interacting with humans.
"A lot of the estimation and planning algorithms my students have developed for the helicopter, we actually reused in the context of interacting with robots," Roy says. "If the problem is no longer how the vehicle plans to get from Point A to Point B, but the problem is how does the vehicle understand what the human wants in terms of some task, then our research programs have a lot of commonality between those two seemingly very different domains."
Cohn, G., Morris, D., Patel, S.N., and Tan, D.S.
Your noise is my command: Sensing gestures using the body as an antenna, Proceedings of the 2011 Annual Conference on human factors in computing systems, Vancouver, British Columbia, Canada, May 712, 2011.
Gallo, L., Placitelli, A.P., and Ciampi, M.
Controller-free exploration of medical image data: Experiencing the Kinect, Proceedings of the 24th IEEE International Symposium on Computer-Based Medical Systems, Los Alamitos, CA, June 2730, 2011.
Henry, P., Krainin, M., Herbst, E., Ren, X., and Fox, D.
RGB-D mapping: Using depth cameras for dense 3D modeling of indoor environments, Proceedings of the International Symposium on Experimental Robotics, New Delhi, India, Dec. 1821, 2010.
Shotton, J., et al.
Real-time human pose recognition in parts from single depth images, Proceedings of the 24th IEEE Computer Vision and Pattern Recognition, Colorado Springs, CO, June 2025, 2011.
Underkoffler, J., Ullmer, B., and Ishii, H.
Emancipated pixels: Real-world graphics in the luminous room, Proceedings of the Special Interest Group on Computer Graphics, Los Angeles, CA, August 813, 1999.
©2011 ACM 0001-0782/11/1200 $10.00
Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and full citation on the first page. Copyright for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, to republish, to post on servers, or to redistribute to lists, requires prior specific permission and/or fee. Request permission to publish from firstname.lastname@example.org or fax (212) 869-0481.
The Digital Library is published by the Association for Computing Machinery. Copyright © 2011 ACM, Inc.
Some great coverage here. But a key point is missed - when we coined the term 'Brave NUI World' in our book (http://www.amazon.com/Brave-NUI-World-Designing-Interfaces/dp/0123822319), we were referring specifically to *new* niches for technology, not about supplanting highly efficient input technologies designed for a particular context. The iPhone and other similar devices did not replace the desktop, they simply enabled elements of an experience previously reserved for the desktop to be experienced elsewhere. The projects that you have described (and all manner of others in the HCI community) are great examples of this as well.
John's an incredibly creative guy, but I'm sure he too would agree that gesturing in-air to generate text (or even dictating) is a far worse experience for desktop publishing or spreadsheet accounting than using a keyboard and mouse. What technologies like Surface and Kinect have enabled researchers to explores are those new contexts, where maybe a PC is used now but it isn't a great experience, or the PC simply can't work well. The hospital environment is one where the sanitation and other elements of context of use are simply ill-suited to using the current input technologies. I welcome you to read the book, written by myself and CHI Academy member Dennis Wixon, both of the Surface and later company-wide NUI gurus at MSFT, where we make this point in detail, and describe a clear design process for how to achieve the sorts of experiences you have written about here.
Displaying 1 comment