[go: up one dir, main page]

0% found this document useful (0 votes)
54 views4 pages

3D Interaction Techniques

3D interaction allows users to interact with and move through 3D virtual environments. Users can navigate virtual spaces using techniques like physical movement, viewpoint manipulation, steering, or selecting destinations. Selection and manipulation techniques allow users to select and adjust 3D objects, with common methods including using a virtual hand, extending reach, or pointing with a virtual ray. System controls let users issue commands through menus, voice, gestures, or virtual tools.

Uploaded by

mohammed Biratu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
54 views4 pages

3D Interaction Techniques

3D interaction allows users to interact with and move through 3D virtual environments. Users can navigate virtual spaces using techniques like physical movement, viewpoint manipulation, steering, or selecting destinations. Selection and manipulation techniques allow users to select and adjust 3D objects, with common methods including using a virtual hand, extending reach, or pointing with a virtual ray. System controls let users issue commands through menus, voice, gestures, or virtual tools.

Uploaded by

mohammed Biratu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 4

3D interaction

From Wikipedia, the free encyclopedia

Jump to navigation Jump to search

In computing, 3D interaction is a form of human-machine interaction where users are


able to move and perform interaction in 3D space. Both human and machine process
information where the physical position of elements in the 3D space is relevant.

The 3D space used for interaction can be the real physical space, a virtual space
representation simulated in the computer, or a combination of both. When the real
space is used for data input, humans perform actions or give commands to the
machine using an input device that detects the 3D position of the human action. When
it is used for data output, the simulated 3D virtual scene is projected onto the real
environment through one output device or a combination of them.

3D Interaction Techniques
3D Interaction Techniques are the different ways that the user can interact with the 3D
virtual environment to execute different kind of tasks. The quality of these techniques
has a profound effect on the quality of the entire 3D User Interfaces. They can be
classified into three different groups: Navigation, Selection and manipulation and
System control.

Navigation

Navigation is the most used by the user in big 3D environments and presents different
challenges as supporting spatial awareness, giving efficient movements between
distant places and making navigation bearable so the user can focus on more
important tasks. These techniques can be divided into two components: travel and
wayfinding.

Travel

Travel is a conceptual technique that consists in the movement of the viewpoint from
one location to another. This orientation is usually handled in immersive virtual
environments by head tracking. Exists five types of travel interaction techniques:

 Physical movement: uses the user's body motion to move through the virtual environment.
Is an appropriate technique when is required an augmented perception of the feeling of
being present or when is required physical effort form the user.
 Manual viewpoint manipulation: the user's hands movements determine the displacement
on the virtual environment. One example could be when the user moves their hands in a way
that seems like is grabbing a virtual rope and pulls his self up. This technique could be easy to
learn and efficient, but can cause fatigue.
 Steering: the user has to constantly indicate where to move. Is a common and efficient
technique. One example of this are the gaze-directed steering, where the head orientation
determines the direction of travel.
 Target-based travel: user specifies a destination point and the system effectuates the
displacement. This travel can be executed by teleport, where the user is instantly moved to
the destination point or the system can execute some transition movement to the destiny.
These techniques are very simple from the user's point of view because he only has to
indicate the destination.
 Route planning: the user specifies the path that should be taken through the environment
and the system executes the movement. The user may draw a path on a map of the virtual
environment to plan a route. This technique allows users to control travel while they have
the ability to do other tasks during motion.

Way finding

Is the cognitive process of defining a route for the virtual environment, using and
acquiring spatial knowledge to construct a cognitive map of the virtual environment.

In order for a good wayfinding, users should receive wayfinding supports during the
virtual environment travel to facilitate it because of the constraints from the virtual
world.

These supports can be user-centered supports such as a large field-of-view or even


non-visual support such as audio, or environment-centered support, artificial cues and
structural organization to define clearly different parts of the environment. Some of
the most used artificial cues are maps, compasses and grids, or even architectural cues
like lighting, color and texture.

Selection and Manipulation

Selection and Manipulation techniques for 3D environments must accomplish at least


one of three basic tasks: object selection, object positioning and object rotation.

Selection

The task of selecting objects or 3D volumes in a 3D environments requires first being


able to find the desired target and then being able to select it. Most 3D
datasets/environments are severed by occlusion problems,[12] so the first step of
finding the target relies on manipulation of the viewpoint or of the 3D data itself in
order to properly identify the object or volume of interest. This initial step is then of
course tightly coupled with manipulations in 3D. Once the target is visually identified,
users have access to a variety of techniques to select it.
Usually, the system provides the user a 3D cursor represented as a human hand whose
movements correspond to the motion of the hand tracker. This virtual hand technique
[13] is rather intuitive because simulates a real-world interaction with objects but with
the limit of objects that we can reach inside a reach-area.

To avoid this limit, there are many techniques that have been suggested, like the Go-
Go technique.[14] This technique allows the user to extend the reach-area using a
non-linear mapping of the hand: when the user extends the hand beyond a fixed
threshold distance, the mapping becomes non-linear and the hand grows.

Another technique to select and manipulate objects in 3D virtual spaces consists in


pointing at objects using a virtual-ray emanating from the virtual hand.[15] When the
ray intersects with the objects, it can be manipulated. Several variations of this
technique has been made, like the aperture technique, which uses a conic pointer
addressed for the user's eyes, estimated from the head location, to select distant
objects. This technique also uses a hand sensor to adjust the conic pointer size.

Many other techniques, relying on different input strategies, have also been
developed.[16]

Manipulation

3D Manipulations occurs before a selection task (in order to visually identify a 3D


selection target) and after a selection has occurred, to manipulate the selected object.
3D Manipulations require 3 DOF for rotations (1 DOF per axis, namely x, y, z) and 3
DOF for translations (1 DOF per axis) and at least 1 additional DOF for uniform
zoom (or alternatively 3 additional DOF for non-uniform zoom operations).

3D Manipulations, like navigation, is one of the essential tasks with 3D data, objects
or environments. It is the basis of many 3D software (such as Blender, Autodesk,
VTK) which are widely used. These software, available mostly on computers, are thus
almost always combined with a mouse and keyboard. To provide enough DOFs (the
mouse only offers 2), these software rely on modding with a key in order to separately
control all the DOFs involved in 3D manipulations. With the recent avent of multi-
touch enabled smartphones and tablets, the interaction mappings of these software
have been adapted to multi-touch (which offers more simultaneous DOF
manipulations than a mouse and keyboard). A survey conducted in 2017 of 36
commercial and academic mobile applications on Android and iOS however
suggested that most applications did not provide a way to control the minimum 6
DOFs required,[7] but that among those which did, most made use of a 3D version of
the RST (Rotation Scale Translation) mapping: 1 finger is used for rotation around x
and y, while two-finger interaction controls rotation around z, and translation along x,
y, and z.
System Control

System control techniques allows the user to send commands to an application,


change the interaction mode or modify a parameter. The command sender always
includes the selection of an element from a set. System control techniques can be
categorized into four groups:

 Graphical menus: visual representations of commands.


 Voice commands: menus accessed via voice.
 Gestural interaction: command accessed via body gesture.
 Tools: virtual objects with an implicit function or mode.

Also exists different hybrid techniques that combine some of the types.

You might also like