Special Section on Touching the 3rd DimensionA survey of 3D object selection techniques for virtual environments
Graphical abstract
Highlights
► We review major 3D object selection techniques for virtual environments. ► Important findings in human control models for 3D object selection are reviewed. ► We analyze major factors influencing selection performance.
Introduction
In the last decades we have witnessed enormous improvements in spatial input devices and motion tracking systems. These advances have motivated the development of a plethora of interaction techniques relying on six DoFs (Degrees of Freedom) input devices and user gestures. Interaction through natural gestures is gaining further popularity since the recent mass commercialization of low-cost solutions for full-body tracking, which is enabling the deployment of natural interfaces outside virtual reality labs. We will use the term 3D interaction to refer to interaction tasks requiring users to make some gestures in free (unconstrained) 3D space. These gestures typically involve one or both hands, and might also involve the user's head and other parts of the body.
The design of appropriate 3D interaction techniques for virtual environments (VEs) is a challenging problem [19], [51]. On the positive side, interacting in free space with natural gestures opens a new world of possibilities for exploiting the richness and expressiveness of the interaction, allowing users to control simultaneously more DoFs and exploiting well-known real-world actions. On the negative side, 3D interaction is more physically demanding and might hinder user tasks by increasing the required dexterity. Compare for example the act of selecting an object using a mouse pointer to that of grasping a 3D object in free space. Mouse movement involves small, fast muscles whereas grasping often requires a complex arm movement involving larger and slower muscles [23], [48]. Furthermore, current immersive VEs, even the most sophisticated ones, neither fail to provide the same level of cues for understanding the environment, nor reproduce faithfully the physical constraints of the real world [74]. For this reason, although humans are used to perform 3D interaction gestures in the real world, users of IVEs often encounter difficulties in understanding 3D spatial relationships and controlling multiple DoFs simultaneously.
Object selection is one of the fundamental tasks in 3D user interfaces [19] and the initial task for most common user interactions in a VE. Manipulation tasks often depend on (and are preceded by) selection tasks. As a consequence, poorly designed selection techniques often have a significant negative impact on the overall user performance. In this survey, we review major 3D interaction techniques intended for 3D object selection tasks. We do not consider indirect selection techniques, e.g. selecting from a menu or performing semantic queries. A 3D object selection technique requires the user to gesture in 3D space, e.g. grabbing an object or pointing to something (see Fig. 1). Two main 3D selection metaphors can be identified: virtual hand [78] and virtual pointing [63], [54]. In the early days, virtual hand techniques were more popular as they map identically virtual tasks with real tasks, resulting in a more natural interaction. Lately, it has been shown that overcoming the physical constraints of the real world provides substantial benefits, e.g. letting the user select objects out of reach by enlarging the user's virtual arm [75], or using virtual pointing techniques such as raycasting [63]. In fact, raycasting selection is one of the most popular techniques for 3D object selection tasks [16]. A number of user studies in the literature have found that virtual pointing techniques often result in better selection effectiveness than competing 3D selection metaphors [19]. Unlike classical virtual hand techniques, virtual pointing techniques allow the user to select objects beyond their area of reach and require relatively less physical movement.
Selection through virtual pointing, though, is not free from difficulties. The selection of small or distant objects through virtual pointing remains to be a difficult task. Some techniques address the selection of small objects by increasing the size of the selection tool [36], [73], at the expense of requiring disambiguation mechanisms to guess the object the user aims to select [30]. Noise from tracking devices and the fact that the interaction takes place in free space with no physical support for the hands [55] further hinders the accurate selection of small targets [43]. The user also has to keep the tool orientation steady until the selection confirmation is triggered, for example, by a button press. Such a confirmation action is likely to produce a change in the tool orientation, nicknamed Heisenberg effect [20], potentially causing a wrong selection. Occlusion is another major handicap for accomplishing spatial tasks [33]. Most interaction techniques for 3D selection and manipulation require the involved objects to be visible. A common solution for selecting occluded objects is to navigate to an appropriate location so that the targets become unoccluded. However, this navigate-to-select approach is impractical for selection-intensive applications. Therefore occlusion management techniques are often essential for helping users discover and access potential targets.
A number of approaches have been proposed to improve user performance in terms of task completion times and error counts [15]. A common strategy is to apply human control models such as the optimized initial impulse model [62] and Fitts' Law [34], [35]. While the optimized initial impulse model refers to the accuracy a user can achieve given the movement required to perform an action, Fitts' Law estimates the time required to acquire a target. However, as users are bounded by human motor skills, there is a natural trade-off between speed and accuracy. In a typical scenario, high-accuracy rates will produce high task completion times and vice-versa.
In the context of the real usage of 3D interfaces, the subjective impressions of the users about an interaction technique can play a larger role than merely speed. The inability to select objects precisely may prove to be overly annoying and thus frustrate users. A performance increase might not be desirable if it is achieved at the expense of increasing the cognitive load of the task, or using techniques requiring extensive training.
The rest of this paper is organized as follows. Section 2 reviews existing human pointing models. In Section 3 we review major techniques for 3D object selection and extend previously proposed classifications [18], [76], [29] with a number of additional criteria to further elucidate the potential benefits and drawbacks of existing selection techniques. A comprehensive summary of the reviewed techniques is given in Table 1. Section 4 analyzes major factors influencing selection performance and proposes some usability guidelines. Finally, Section 5 provides some concluding remarks and future research directions.
Section snippets
Human pointing models
In order to point to (acquire) an object (the target), the user is required to perform a set of gestures (movements) to position the selection tool (e.g. his finger) over it. For each movement, the final position of the selection tool (endpoint) determines whether the acquisition is accomplished (the endpoint is inside the target) or not (the endpoint is outside the target). Once the target is acquired, the user has to trigger some selection mechanism to confirm the acquisition (e.g. pressing a
Classification of selection techniques
A number of taxonomies have been proposed to classify existing 3D selection techniques. In Bowman et al. [18] classification, interaction techniques are decomposed into subtasks and classified according to them (see Fig. 3). Following [18], a selection technique has to provide means to indicate an object (object indication), a mechanism to confirm its selection (confirmation of selection) and visual, haptic or audio feedback to guide the user during the selection task (feedback). One limitation
Factors influencing performance
A number of usability guidelines exist for 2D user interfaces, however, in general, they are not directly applicable to 3D user interfaces. 3D user interfaces are significantly more difficult to design, implement and use than their 2D counterparts. 3DUIs are based on real-world characteristics such as naive physics, body awareness, environmental awareness, social awareness and social skills [46].
There are a few works explicitly focusing on usability guidelines for 3D user interfaces, being the
Conclusions and future outlook
The act of pointing to graphical elements is one of the fundamental tasks in human–computer interaction. Although 3D interaction techniques for target selection have been used for many years, they still exhibit major limitations regarding effective, accurate selection of targets in real-world applications. Some of these limitations are concerned with visual feedback issues (occlusion, visibility mismatch, depth perception in stereoscopic displays) and the inherent features of the human motor
References (108)
- et al.
Anisomorphic ray-casting manipulation for interacting with 2D GUIs
Comput Graph
(2007) “Beating” Fitts' lawvirtual enhancements for pointing facilitation
Int J Hum–Comput Stud
(2004)- et al.
View-projection animation for 3D occlusion management
Comput Graph
(2007) - et al.
A human motor behavior model for distal pointing tasks
Int J Hum–Comput Stud
(2010) - et al.
JDcada highly interactive 3D modeling system
Comput Graph
(1994) - et al.
Manipulating objects in virtual worldscategorization and empirical evaluation of interaction
J Visual Lang Comput
(1999) - et al.
Fitts' law as the outcome of a dynamic noise filtering model of motor control
Hum Movement Sci
(1995) - Andujar C, Argelaguet F. Virtual pads: decoupling motor space and visual space for flexible manipulation of 2D windows...
- et al.
Hand-based disocclusion for the world-in-miniature metaphor
PresenceTeleop Virt Environ
(2010) - Argelaguet F. Pointing facilitation techniques for 3D object selection on virtual environments. PhD thesis, Universitat...
Efficient 3D pointing selection in cluttered virtual environments
IEEE Comput Graph Appl
A survey of usability evaluation in virtual environmentsclassification and comparison of methods
PresenceTeleop Virt Environ
The virtual venueuser-computer interaction in information-rich virtual environments
PresenceTeleop Virt. Environ.
3D user interfacestheory and practice
Multimodal feedback for the acquisition of small targets
Ergonomics
Adaptive cutaways for comprehensible rendering of polygonal scenes
ACM Trans Graph
A morphological analysis of the design space of input devices
ACM Trans Inf Syst
Dense and dynamic 3D selection for game-based virtual environments
IEEE Trans Visualization Comput Graph
A taxonomy of 3D occlusion management for visualization
IEEE Trans. Visualization Comput Graph
The information capacity of the human motor system is controlled by the amplitude of movement
J. Exp. Psychol.
Information capacity of discrete motor response
J. Exp. Psychol.
Precise and rapid interaction thought scaled manipulation in immersive virtual environments
IEEE Virtual Reality
PRISM interaction for enhancing control in immersive virtual environments
ACM Trans Comput–Hum Interact
The challenges of 3D interactiona CHI'94 workshop
SIGCHI Bull
Cited by (352)
Selection as Tapping: An evaluation of 3D input techniques for timing tasks in musical Virtual Reality
2024, International Journal of Human Computer StudiesResearch on a spatial–temporal characterisation of blink-triggered eye control interactions
2024, Advanced Engineering InformaticsA survey of immersive visualization: Focus on perception and interaction
2023, Visual InformaticsThe effect of hands synchronicity on users perceived arms Fatigue in Virtual reality environment
2023, International Journal of Human Computer StudiesTwo-step techniques for accurate selection of small elements in VR environments
2023, Graphical ModelsEnhancing weight perception in virtual reality: an analysis of kinematic features
2024, Virtual Reality