Abstract
Natural and intuitive human-computer interfaces, while striving for immersive experience, often suffer from relatively cumbersome input methods. Mice and keyboards are widely accepted yet limited to point-and-click and basic textual editing operations, distancing the user from the world presented to them. Touch-screen based gestures have introduced a partial solution to this problem. However, they are constrained due to size, weight and positioning and currently provide limited gesture-based operations. In this paper, we propose a gesture model by meta-action formulation based on observations of existing touch gestures as experienced from different perspectives such as type, multiplicity, stage-of-control and control space. We then propose a kernel gesture set which forms the basis of a unified user experience across a large variety of applications and devices. In order to show the effectiveness, intuitiveness and adaptability of the proposed kernel gesture set, a user study is given. We further implement and extend the kernel gesture set through a case study in a very practical domain: garment modeling.
Original language | English |
---|---|
Title of host publication | Proceedings of VRCAI 2011 |
Subtitle of host publication | ACM SIGGRAPH Conference on Virtual-Reality Continuum and its Applications to Industry |
Pages | 279-286 |
Number of pages | 8 |
DOIs | |
Publication status | Published - 1 Dec 2011 |
Event | 10th International Conference on Virtual Reality Continuum and Its Applications in Industry, VRCAI'11 - Hong Kong, China Duration: 11 Dec 2011 → 12 Dec 2011 |
Conference
Conference | 10th International Conference on Virtual Reality Continuum and Its Applications in Industry, VRCAI'11 |
---|---|
Country/Territory | China |
City | Hong Kong |
Period | 11/12/11 → 12/12/11 |
Keywords
- Garment modeling
- Gesture modeling
- Kernel gestures
- Meta-action
- Multi-touch interaction
ASJC Scopus subject areas
- Artificial Intelligence
- Computer Science Applications