3D Interaction Techniques
3D Interaction Techniques
The 3D space used for interaction can be the real physical space, a virtual space
representation simulated in the computer, or a combination of both. When the real
space is used for data input, humans perform actions or give commands to the
machine using an input device that detects the 3D position of the human action. When
it is used for data output, the simulated 3D virtual scene is projected onto the real
environment through one output device or a combination of them.
3D Interaction Techniques
3D Interaction Techniques are the different ways that the user can interact with the 3D
virtual environment to execute different kind of tasks. The quality of these techniques
has a profound effect on the quality of the entire 3D User Interfaces. They can be
classified into three different groups: Navigation, Selection and manipulation and
System control.
Navigation
Navigation is the most used by the user in big 3D environments and presents different
challenges as supporting spatial awareness, giving efficient movements between
distant places and making navigation bearable so the user can focus on more
important tasks. These techniques can be divided into two components: travel and
wayfinding.
Travel
Travel is a conceptual technique that consists in the movement of the viewpoint from
one location to another. This orientation is usually handled in immersive virtual
environments by head tracking. Exists five types of travel interaction techniques:
Physical movement: uses the user's body motion to move through the virtual environment.
Is an appropriate technique when is required an augmented perception of the feeling of
being present or when is required physical effort form the user.
Manual viewpoint manipulation: the user's hands movements determine the displacement
on the virtual environment. One example could be when the user moves their hands in a way
that seems like is grabbing a virtual rope and pulls his self up. This technique could be easy to
learn and efficient, but can cause fatigue.
Steering: the user has to constantly indicate where to move. Is a common and efficient
technique. One example of this are the gaze-directed steering, where the head orientation
determines the direction of travel.
Target-based travel: user specifies a destination point and the system effectuates the
displacement. This travel can be executed by teleport, where the user is instantly moved to
the destination point or the system can execute some transition movement to the destiny.
These techniques are very simple from the user's point of view because he only has to
indicate the destination.
Route planning: the user specifies the path that should be taken through the environment
and the system executes the movement. The user may draw a path on a map of the virtual
environment to plan a route. This technique allows users to control travel while they have
the ability to do other tasks during motion.
Way finding
Is the cognitive process of defining a route for the virtual environment, using and
acquiring spatial knowledge to construct a cognitive map of the virtual environment.
In order for a good wayfinding, users should receive wayfinding supports during the
virtual environment travel to facilitate it because of the constraints from the virtual
world.
Selection
To avoid this limit, there are many techniques that have been suggested, like the Go-
Go technique.[14] This technique allows the user to extend the reach-area using a
non-linear mapping of the hand: when the user extends the hand beyond a fixed
threshold distance, the mapping becomes non-linear and the hand grows.
Many other techniques, relying on different input strategies, have also been
developed.[16]
Manipulation
3D Manipulations, like navigation, is one of the essential tasks with 3D data, objects
or environments. It is the basis of many 3D software (such as Blender, Autodesk,
VTK) which are widely used. These software, available mostly on computers, are thus
almost always combined with a mouse and keyboard. To provide enough DOFs (the
mouse only offers 2), these software rely on modding with a key in order to separately
control all the DOFs involved in 3D manipulations. With the recent avent of multi-
touch enabled smartphones and tablets, the interaction mappings of these software
have been adapted to multi-touch (which offers more simultaneous DOF
manipulations than a mouse and keyboard). A survey conducted in 2017 of 36
commercial and academic mobile applications on Android and iOS however
suggested that most applications did not provide a way to control the minimum 6
DOFs required,[7] but that among those which did, most made use of a 3D version of
the RST (Rotation Scale Translation) mapping: 1 finger is used for rotation around x
and y, while two-finger interaction controls rotation around z, and translation along x,
y, and z.
System Control
Also exists different hybrid techniques that combine some of the types.