WO2006104132A1 - ユーザインタフェイスシステム - Google Patents
ユーザインタフェイスシステム Download PDFInfo
- Publication number
- WO2006104132A1 WO2006104132A1 PCT/JP2006/306245 JP2006306245W WO2006104132A1 WO 2006104132 A1 WO2006104132 A1 WO 2006104132A1 JP 2006306245 W JP2006306245 W JP 2006306245W WO 2006104132 A1 WO2006104132 A1 WO 2006104132A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- unit
- finger
- gui
- user interface
- interface system
- Prior art date
Links
- 238000001514 detection method Methods 0.000 claims abstract description 72
- 230000006870 function Effects 0.000 claims description 73
- 238000000034 method Methods 0.000 claims description 36
- 238000003384 imaging method Methods 0.000 claims description 23
- 230000004044 response Effects 0.000 claims description 9
- 238000006073 displacement reaction Methods 0.000 claims description 4
- 230000002596 correlated effect Effects 0.000 abstract 1
- 210000003811 finger Anatomy 0.000 description 244
- 238000012545 processing Methods 0.000 description 24
- 210000003813 thumb Anatomy 0.000 description 21
- 238000010586 diagram Methods 0.000 description 20
- 230000015654 memory Effects 0.000 description 14
- 230000008569 process Effects 0.000 description 9
- 230000004048 modification Effects 0.000 description 7
- 238000012986 modification Methods 0.000 description 7
- 230000008859 change Effects 0.000 description 6
- 210000004932 little finger Anatomy 0.000 description 6
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 5
- 238000004590 computer program Methods 0.000 description 5
- 238000012937 correction Methods 0.000 description 4
- 238000011410 subtraction method Methods 0.000 description 3
- 210000000746 body region Anatomy 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000003825 pressing Methods 0.000 description 2
- 239000000700 radioactive tracer Substances 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000005057 finger movement Effects 0.000 description 1
- 210000004247 hand Anatomy 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/042—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means
- G06F3/0425—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by opto-electronic means using a single imaging device like a video camera for tracking the absolute position of a single or a plurality of objects with respect to an imaged reference surface, e.g. video camera imaging a display or a projection screen, a table or a wall surface, on which a computer generated image is displayed or projected
Definitions
- the present invention relates to a user interface system, and more particularly to a user interface system that allows a user to specify a function to be executed for a computer device.
- each subgroup is associated with the finger of the user, and the finger recognition system identifies the subgroup by recognizing the finger currently operating the key. To do. In other words, because the keypad does not display alphabets, numbers, and command descriptions, the user must remember which fingers are associated with which subgroups, which is inefficient and unusable. There are problems.
- an object of the present invention is to provide a user interface system that is easier to use.
- a first aspect of the present invention provides a user interface system. It is aimed at.
- the present invention includes an operation unit that receives a user's instruction input, a finger position detection unit that detects a first coordinate value group indicating a current position of a user's fingertip on the operation surface of the operation unit, and a graphical user interface (GUI)
- GUI graphical user interface
- a component storage unit that stores component data representing a component, an allocation unit that uniquely allocates component data held in the component storage unit to each of the first coordinate values detected by the finger position detection unit, and an allocation unit allocation
- the GUI image representing the user interface in which the GUI component represented by the component data stored in the component storage unit and each of the first coordinate values detected by the finger position detection unit are uniquely associated is generated.
- An image creation unit and a display control unit configured to display a GUI image generated by the image creation unit on a display screen.
- an imaging device installed so as to be capable of photographing the user's manual operation
- the finger position detection unit is configured to detect a current position of the user's fingertip on the operation surface of the operation unit from an image captured by the imaging device It is preferable to detect a coordinate value group indicating.
- the image creation unit extinguishes the GUI part associated with the finger bent by the user among the fingers of the hand image acquired by the imaging device, and the coordinates indicating the current position of the fingertip remaining last It is preferable to further include a function execution unit that executes the function of the GUI component assigned to the value.
- the image creation unit connects the position indicated by each of the first coordinate values detected by the finger position detection unit and the GUI component represented by the component data allocated by the allocation unit with a line.
- the image creation unit further includes a history information holding unit for storing functions executed in the past, and creates a GUI image to be added to the fingertip based on the history information. Is preferred.
- the image creating unit synthesizes a GUI component represented by the component data assigned by the assigning unit at or around the position indicated by each of the first coordinate values detected by the finger position detecting unit.
- the image creation unit further includes a left and right hand discriminating unit for discriminating left and right hands, and the GUI component added to the first coordinate value is changed when operating with the left hand and when operating with the right hand. It is preferable to do.
- the display control unit displays a hand image including first coordinate values indicating the positions of the plurality of types of fingertips identified by the finger identifying unit, and the allocating unit is configured to represent the plurality of types identified by the finger identifying unit. It is preferable to delete the displayed hand image after uniquely assigning the GUI component to each of the first coordinate values indicating the position of the fingertip.
- the display control unit displays the number of fingers corresponding to the GUI component assigned to each of the first coordinate values indicating the positions of the plurality of types of fingertips identified by the finger identifying unit by the assigning unit. It is preferable.
- the image creation unit creates a GUI image in which the size of the GUI component stored in the component storage unit is changed according to the type of the finger identified by the finger identification unit.
- the display control unit applies the fingertip represented by the specific first coordinate value detected by the finger position detection unit.
- the GUI component assigned by the assigning unit is displayed so as to follow the movement of the fingertip.
- the display control unit moves the GUI component assigned to the fingertip by the assigning unit according to the operation of the fingertip represented by the specific first coordinate value detected by the finger position detection unit. U, prefer to display.
- the assigning unit maintains only the assignment for the specific first coordinate value detected by the finger position detecting unit, and assigns the other first coordinate value detected by the finger position detecting unit. U, prefer to change.
- a touch position detection unit that detects a second coordinate value indicating a position where the user touches the operation surface of the operation unit, a second coordinate value detected by the touch position detection unit, and an allocation unit From the first coordinate value uniquely assigned by, the contact finger that determines how many fingers the user touches the operation surface and further identifies the GUI part assigned to the determined finger It is preferable to further include a determination unit and a function execution unit having a function assigned to the GUI component specified by the contact finger determination unit.
- the assigning unit preferably reassigns the component data once assigned to the first coordinate value detected by the finger position detecting unit to another GUI component.
- the assigning unit does not assign the part to the first coordinate value indicating the position of the specific type of finger, and the first coordinate value indicating the position to the remaining finger is interchanged between the plurality of devices. Different parts data are assigned uniquely, and the image creation unit can select multiple devices with a specific finger and respond to operations with a specific finger for the currently selected device.
- the G UI in which the first coordinate value indicating the position of the remaining finger whose type has been identified by the above and the part data stored in the part storage unit and assigned to the remaining finger by the assigning part are associated with each other Prefer to produce images.
- the contact position detection unit detects a second coordinate value indicating the first contact position in response to each contact, The second coordinate value indicating the second contact position is detected, and the contact finger determination unit detects the first second coordinate value detected by the contact position detection unit and the first coordinate value uniquely assigned by the assignment unit. From the coordinate value of 1 and GUI parts, it is determined how many fingers the user is touching the operation surface of the operation unit, and the second coordinate value of the second time detected by the contact position detection unit. Therefore, it is preferable that the GUI component specified by the user this time is specified.
- the contact position detection unit When the user traces between at least two points on the operation surface of the operation unit, the contact position detection unit periodically detects a second coordinate value indicating the current contact position in response to each contact. Then, the contact finger determination unit calculates the operation surface of the operation unit from the second coordinate value periodically detected by the contact position detection unit, the first coordinate value uniquely assigned by the assignment unit, and the GUI component. It is preferable to determine how many fingers the user is touching, and to identify the GUI component specified by the user this time.
- a second aspect of the present invention is directed to a display method used in a user interface system.
- the present invention provides a finger position detection step for detecting a first coordinate value group indicating a current position of a user's fingertip on an operation surface of an operation unit provided in a user interface system and receiving an instruction input from the user, and a user interface A component that is stored in the storage device of the system and represents a GUI (Graphical User Interface) component
- An assignment step that uniquely assigns data to each of the first coordinate values detected by the finger position detection step, and a GUI component stored in the storage device according to the assignment step assignment and the finger position detection step.
- An image creation step for generating a GUI image representing a user interface that is uniquely associated with each of the first coordinate values, and the GUI image generated by the image creation step are stored in the user interface system.
- a display control step for displaying on the provided display screen.
- each GUI component As described in each aspect above, a predetermined function is assigned to each GUI component. Each of these GUI parts is displayed on the display screen as part of the GUI image. The displayed GUI component is associated with each fingertip of the user, that is, assigned to each fingertip. Since each GUI component is displayed in this manner, the user interface system according to each aspect can present to the user in a visible manner at least what functions can be realized. This makes it possible to provide a user interface system that is easier to use.
- FIG. 1 is a schematic diagram showing a schematic configuration of a user interface system 1 according to an embodiment of the present invention.
- FIG. 2 is a schematic diagram showing a functional block configuration of the computer apparatus 3 shown in FIG. 1.
- FIG. 3 is a flowchart showing display processing of the computer apparatus 3 shown in FIG.
- FIG. 4 is a flowchart showing a detailed processing procedure of finger position detection unit 32 shown in FIG.
- FIG. 5A is a first schematic diagram showing an example of processing of the finger identification unit 33 shown in FIG.
- FIG. 5B is a second schematic diagram showing an example of processing of the finger identification unit 33 shown in FIG.
- FIG. 6A is a first schematic diagram showing a GUI image created by the image creation unit 36 shown in FIG.
- FIG. 6B is a second schematic diagram showing a GUI image created by the image creation unit 36 shown in FIG.
- FIG. 6C is a third schematic diagram showing a GUI image created by the image creation unit 36 shown in FIG.
- FIG. 6D is a fourth schematic diagram showing a GUI image created by the image creation unit 36 shown in FIG.
- FIG. 7 is a flowchart showing a function execution process of the computer apparatus 3 shown in FIG.
- FIG. 8 is a schematic diagram showing a schematic configuration of a user interface system 11 according to a modified example of the present invention.
- FIG. 9 is a schematic diagram showing a functional block configuration of the computer apparatus 14 shown in FIG. 8.
- FIG. 10 is a flowchart showing display processing of the computer device 14 shown in FIG.
- FIG. 11A is a first schematic diagram showing a GUI image created by the image creation unit 146 shown in FIG.
- FIG. 11B is a second schematic diagram showing a GUI image created by the image creation unit 146 shown in FIG.
- FIG. 12 is a flowchart showing the function execution processing of the computer device 14 shown in FIG.
- FIG. 13 is a schematic diagram showing a first example of a multi-function assignment method in the user interface system 1 or 11 shown in FIG. 1 or FIG.
- FIG. 14 is a schematic diagram showing a second example of the multifunctional assignment method in the user interface system 1 or 11 shown in FIG. 1 or FIG.
- FIG. 15 is a schematic diagram showing a third example of the multi-function assignment method in the user interface system 1 or 11 shown in FIG. 1 or FIG.
- FIG. 16 is a schematic diagram showing a fourth example of the multi-function assignment method in the user interface system 1 or 11 shown in FIG. 1 or FIG.
- FIG. 17 is a schematic diagram showing an example of a GUI component assignment changing method in the user interface system 1 or 11 shown in FIG. 1 or FIG. 8.
- FIG. 18A is an example of a display when a GUI component of the main menu in the user interface system 1 or 11 shown in FIG. 1 or FIG. 8 is assigned.
- FIG. 18B is a sub-menu display example displayed by selecting a function from the main menu of FIG. 18A.
- FIG. 19A is an example of a display when GUI parts are assigned in the user interface system 1 or 11 shown in FIG. 1 or FIG.
- FIG. 19B is an example in which the assignment of GUI parts in the user interface system 1 or 11 shown in FIG. 1 or FIG. 8 is changed and a button follows a specific finger.
- FIG. 20 is an example of a display in which a button is followed by a specific finger when operating the GUI.
- FIG. 21 shows an example of the display when the GUI parts are assigned in the user interface system 1 or 11 shown in FIG. 1 or FIG. 8 and the display when the finger is not separated from the contact position detection unit force. It is an example.
- FIG. 22 is an example of erasing the display of the hand shape when the GUI part is displayed.
- FIG. 23 is an example of a display when a function is added to a finger and a GUI is operated.
- FIG. 24A is an example when a finger is associated with a GUI and a GUI component is operated.
- FIG. 24B is an example of a finger movement when operating the GUI component in FIG. 24A.
- FIG. 25A is an example of a display when functions in the user interface system 1 or 11 shown in FIG. 1 or FIG. 8 are assigned.
- FIG. 25B is an example of a display when the remaining finger functions in FIG. 25A are executed.
- FIG. 1 is a schematic diagram showing a schematic configuration of a user interface system 1 according to an embodiment of the present invention.
- FIG. 1 shows a two-dimensional coordinate system including an X axis and a y axis on the display device 21 side.
- a user interface system 1 includes a touch screen 2, a computer device 3, a storage device 4, and an imaging device 5 as a schematic configuration.
- the touch screen 2 is displayed on the touch screen 2 when the user touches the display screen 21 having a predetermined size with each of, for example, five fingers F to F.
- the computer device 3 typically includes at least a processor and various memories, and operates according to a computer program stored in a memory such as a ROM.
- FIG. 2 is a schematic diagram showing a functional block configuration of such a computer apparatus 3.
- the computer device 3 executes the above-described computer program, whereby an image receiving unit 31, a finger position detecting unit 32, a finger identifying unit 33, a component assigning unit 34, an image creating unit 35, The function of the display control unit 36 is realized.
- the storage device 4 stores GUI component data necessary for generating a GUI (Graphical User Interface) image to be displayed on the display screen 21.
- GUI Graphic User Interface
- a predetermined function is assigned to the GUI part data
- the GUI part data is image data representing a GUI part such as an icon or a button.
- FIG. 1 shows two GUI parts C and C as an example.
- the imaging device 5 is installed so that at least the display area (that is, the display screen) 21 of the touch screen 2 is covered by the angle of view of itself.
- the imaging device 5 installed in this way captures the range of its angle of view at a predetermined time interval and outputs the captured image to the computer device 3.
- the user places his / her hand H (see FIG. 1) on the display screen 21.
- the imaging device 5 captures a state within its own angle of view, and outputs a captured image obtained thereby to the computer device 3 (step S31).
- the hand H is on the display screen 21, the hand H is reflected in the photographed image.
- the image reception unit 31 receives the captured image output from the imaging device 5 (step S32), and the finger position detection unit 32 determines each finger F of the user from the received captured image.
- Coordinate value indicating position on F display screen 21 hereinafter, current position coordinate value
- the finger position detection unit 32 captures the captured image data received by the image reception unit 31 into a frame memory (not shown) (step S41).
- the finger position detection unit 32 performs body region extraction processing.
- partial image data representing an area in which the user's hand H is shown is created from the captured image data in the frame memory (step S42).
- the region of the user's hand H and the region of the background (mostly the display screen 21) in the captured image in the frame memory are distinguished, and only the region of the user's hand H is extracted.
- the finger position detection unit 32 acquires in advance image data showing only the display screen 21 via the image reception unit 31 as background image data.
- the finger position detection unit 32 compares such background image data with the captured image data captured in step S41 for each pixel or block, and manually captures a region different from the background portion in the captured image. Extract as H region.
- the background of the hand H does not need to be a single color when the imaging device 5 captures an image. Therefore, the display screen 21 of the touch screen 2 is displayed on the display system and the operation system as in this embodiment. It is effective when sharing as.
- the finger position detection unit 32 performs fingertip position detection processing, and uses the partial image data as described above to indicate the current position coordinate value indicating the position of the user's fingertip.
- a group is detected (step S43).
- the finger position detection unit 32 determines one candidate point for each finger F — F.
- a typical example of a landmark is a nail
- the finger position detection unit 32 determines the brightness of the mark or the mark from the partial image data as described above. The hue is detected, and the current position coordinate value indicating the position of the tip of each finger F-F is thereby obtained.
- the finger position detection unit 32 performs calibration and tracks the tip of the finger FF placed at a predetermined initial position according to the movement of the hand H. To do. Than
- the user places the tip of the finger FF at a position (initial position) determined on the display screen 21. In this state, the image is taken by the imaging device 5.
- the finger position detection unit 32 acquires the captured image and the brightness of the initial position (that is, the tip of the finger F-F) is obtained.
- the finger position detection unit 32 Detect the degree or hue. Thereafter, each time the finger position detection unit 32 acquires a captured image from the imaging device 5, the finger position detection unit 32 has substantially the same luminance and the same as the tip of the finger FF during calibration.
- the finger identification unit 33 acquires the current position coordinate value group acquired by the finger position detection unit 32, and the current position coordinate value is the thumb F and the index finger F. , Middle finger F, ring finger F and little finger F
- the type of finger is identified using the distance between the fingertips. Specifically, first, the finger identification unit 33 sets a combination of the current position coordinate value at the end and the current position coordinate value next to it from the five current position coordinate values obtained from the finger position detection unit 32. Select two. Thereafter, the finger identification unit 33 calculates the distance between the two current position coordinate values belonging to one set and the distance between the two current position coordinate values belonging to the other set, and the longer one is the thumb F and the index finger F. Between
- finger identification The unit 33 identifies which current position coordinate value belongs to which finger.
- the finger identification unit 33 includes the thumb F and the index finger in order from the smallest or largest X coordinate value of the five current position coordinate values obtained from the finger position detection unit 32.
- the finger identification unit 33 uses the five current position coordinate values obtained from the finger position detection unit 32 as shown in FIG. 5A. Select the coordinate value P with the smallest X coordinate value and the coordinate value P with the largest X coordinate value. Also, the finger identification unit 33 is generated by the finger position detection unit 32 and max
- the thicknesses W and W of the fingers at both ends are detected as shown in FIG.
- the finger identification unit 33 One is regarded as thumb F and the thin one is regarded as little finger F. Furthermore, the finger identification unit 33
- the current position coordinate value P is identified as the thumb F.
- Positional coordinates can also be identified.
- the calculation improves the accuracy of identification at least once.
- the component allocation unit 34 identifies the identification result from the finger identification unit 33, that is, the current position coordinate value of the thumb F, the current position coordinate value of the index finger F, and the middle finger F.
- the five current position coordinate values received this time are uniquely assigned to each of the five GUI part data stored in the storage device 4 and the current assignment status is held (step S35).
- the current position coordinate value of thumb F is shown in FIG.
- the image creating unit 35 creates GUI image data representing a GUI image in a frame memory (not shown) (step S36). Specifically, in the GUI image, each of the five GUI parts represented by the GUI part data is synthesized at coordinate positions that do not overlap each other. Furthermore, in the GUI image, the component assignment unit 34 is assigned to the current position coordinate value of thumb F held by 34 and the component assignment unit 34.
- FIGS. 6A to 6D four GUI images rendered by the image creation unit 35 will be exemplified.
- the image creating unit 35 calculates the current position coordinate value P of the thumb F and
- Lines L — L are of different types (ie typically line types, line colors, line thicknesses)
- the line type may be changed according to the number of times the user specified to execute the function specified by the GUI component! /.
- each GUI component C C is synthesized toward the end of the display screen 21.
- each GUI component C-C and line L-L are translucent by the user.
- Line L — L is displayed on display screen 21.
- Each finger F — is preferably drawn to bypass the area where F is placed.
- the image creation unit 35 displays the current position coordinate value of the thumb F.
- the target GUI part C is synthesized at the position specified by P. Similarly, other fingers F — F
- the target GUI component C-C is aligned with the position specified by the current position coordinate value P-P of 1 1 2 5
- each GUI component C C 1S is a description that represents a predetermined function.
- the image creating unit 35 performs a process around the current position coordinate value P of the thumb F.
- UI component c 2 -c is synthesized.
- the finger identification unit 33 can identify whether the user H is currently using the left or right hand H, so the component assignment unit 34 can It is also possible to assign part data belonging to a predetermined first group to each finger and assign part data belonging to a second group different from the first group to each finger of the right hand. .
- the image creating unit 35 also changes the usage history or usage frequency of the GUI component C C according to the usage history or the usage frequency.
- GUI parts For GUI parts that are infrequent, they can be combined with a GUI image or combined with a GUI image to flash.
- step S36 the display control unit 36 displays the GUI image (FIGS. 1 and 6A) on the display screen 21 of the touch screen 2 according to the GUI image data created this time in the frame memory. —See Figure 6D) (step S37).
- the current position coordinate value P indicating the position of the thumb F and the GU
- the current position coordinate value P indicating the position of the index finger F, which is connected to the part I by the line L
- GUI part C is connected with line L, but for the convenience of illustration, it is attached to the other finger F-F.
- the processing described with reference to FIGS. 1 and 2 is executed at a predetermined time interval.
- the imaging device 5 captures the captured image at the above time interval.
- Output to computer device 3 periodically.
- a series of processes of the image receiving unit 31, the finger position detecting unit 32, the finger identifying unit 33, the component assigning unit 34, the image creating unit 35, and the display control unit 36 are periodically performed at the above time intervals. To be done. Therefore, while each finger F — F is moving on the display screen 21, the current position coordinate value of each finger F — F is
- the computer device 3 operates according to a computer program stored in an internal memory, and as shown in FIG. 2, a contact position detection unit 37, a contact finger determination unit 38, The function execution unit 39 is further realized.
- the user touches the display screen 21 with one of the fingers to which a specific function is assigned by the above-described display process, and increases the force. Then, the touch screen 2 gives a signal (hereinafter referred to as a position signal) that can specify the current contact position to the computer device 3.
- a position signal a signal that can specify the current contact position to the computer device 3.
- the contact position detection unit 37 is in contact with the user on the display screen 21.
- a coordinate value indicating the position (hereinafter referred to as a contact coordinate value) is detected (step S72).
- the contact finger determination unit 38 checks the allocation status held in the component allocation unit 34, and selects a current position coordinate value that substantially matches the contact coordinate value detected by the contact position detection unit 37. (Step S73). Further, the contact finger determination unit 38 specifies a GUI component that is assigned to the selected current position coordinate value (step S74).
- the function execution unit 39 executes a function assigned to the GUI component recognized by the contact finger determination unit 38 (step S75).
- the component allocation unit 34 holds the current position of each finger and the GUI component allocated to each finger. Further, the contact finger determination unit 38 determines the finger currently used by the user based on the position signal sent from the touch screen 2 and the allocation status held in the component allocation unit 34, and further, The function that the user wants to execute this time is specified. According to the user interface system 1 as described above, the user does not need to bring a finger to a GUI component arranged at a predetermined location as in the prior art. Thus, it is possible to provide the user interface system 1 that can be operated quickly by the user.
- the user interface system 1 can be realized by combining a touch panel, a touch pad, or a touch tracer with a projector.
- the projector projects a GUI image on the touch panel, touch pad, or touch tracer.
- the user interface system 1 includes the touch screen 2, and as a result, the display screen 21 is a user operation area and a GUI image display area.
- the user operation area and the GUI image display area may be located away from each other.
- the user interface system 11 in which the user operation area and the GUI image display area are provided at independent positions will be described with reference to FIG.
- a user interface system 11 includes, as a rough configuration, a pointing device 12, an imaging device 13, a computer device 14, a storage device 15, and a display device. 16 and.
- the pointing device 12 defines an area (hereinafter referred to as an operation area) 121 in which the user can move his / her finger in order to instruct the computer device 3 to execute a predetermined function.
- the contact position signal indicating the position where the user has touched with the finger is output.
- Examples of such pointing devices 12 include pressure sensitive or optical devices such as touch panels, touch pads or touch tracers.
- Another example of the pointing device 12 is an ultrasonic device.
- the imaging device 13 is installed so that at least the operation area 121 is covered by its own angle of view.
- the imaging device 13 installed in this way captures the range of its field angle at predetermined time intervals, and outputs the captured image to the computer device 14.
- the computer device 14 typically includes at least a processor and various memories, and operates according to a computer program stored in a memory such as a ROM.
- FIG. 9 is a schematic diagram showing a functional block configuration of such a computer device 14.
- the computer device 14 executes the above-described computer program, whereby an image receiving unit 141, a finger position detecting unit 142, a finger identifying unit 143, a component allocating unit 144, and an image creating unit 145
- the display control unit 146, the contact position detection unit 147, the contact finger determination unit 148, and the function execution unit 149 are realized.
- the storage device 15 stores the same GUI part data as the storage device 4 described above.
- the display device 16 is created by at least the computer device 14 that does not allow the user to specify a GUI part using a finger on the display screen as in the touch screen as described above. It would be nice if it had a function to display GUI images.
- Typical examples of such a display device 16 include a liquid crystal display, a CRT display, an organic EL display, a far-focus display, or a projector.
- the user places his / her hand H in the operation area 121.
- the imaging device 13 outputs a captured image in which the hand H is reflected to the computer device 14 (step S 101).
- the image receiving unit 141 receives a captured image (step S). 102), the finger position detection unit 142 operates the user's fingers FF from the received captured image.
- a coordinate value indicating the position on the work area 121 (hereinafter referred to as the current position coordinate value) is detected, and a coordinate value group (hereinafter referred to as the contour coordinate value) that defines the shape (contour) of the hand H. (Referred to as a group) (step S103).
- the finger position detection unit 142 creates partial image data similar to that of the above-described embodiment by the brightness threshold method as the first method. That is, the finger position detection unit 142 distinguishes the background (operation area 121) and the user's hand H in the captured image based on the luminance value of each pixel. Therefore, it is preferable that the color of the operation area 121 is black or a color close to it, and that the surface of the operation area 121 is matted. Further, the threshold value is set in advance to be a value larger than any luminance value of the pixel in which the background is reflected in the captured image.
- the finger position detection unit 142 can create partial image data representing the region of the hand H by comparing the luminance of each pixel of the captured image with the threshold value. Further, in order to stably generate partial image data even in a dark environment, the imaging device 13 irradiates, for example, visible light or near-infrared light in the range of its field angle, and the hand H Make it appear brighter than the threshold.
- the finger position detection unit 142 creates partial image data by a chromak processing method. Chromaki processing is often used for television broadcasting, and it is possible to extract parts other than blue against the background of blue, which is the most distinguishable from human skin color. Create data.
- the operation region 121 is colored in blue, and the imaging device 13 irradiates visible light, for example, in the range of its field angle so that the hand H and the background are clearly reflected.
- the finger identification unit 143 identifies which finger of each current position coordinate value group is the same as in the above-described embodiment. (Step S 10 4).
- the component assigning unit 144 receives the identification result from the finger identifying unit 143, and adds the five current position coordinate values received this time to each of the five GUI component data stored in the storage device 15. Uniquely assigned and the current assignment status is maintained (step S105).
- the image creating unit 155 performs a frame memory (not shown).
- GUI image data representing the GUI image is created in the memory (step S106).
- an object I representing the shape of the hand is synthesized in the region defined by the above-described contour coordinate value group, as compared with the above-described embodiment. It is different in point.
- the display control unit 146 displays a GUI image on the display screen 161 of the display device 16 according to the GUI image data created this time in the frame memory (step S107).
- this user interface system 11 also operates in the same manner as in the above-described embodiment, and as a result, it is easy to recognize which finger the GUI component is assigned to, so that there is a technical effect. become.
- the user touches the operation area 121 with one of the fingers to which a specific function is assigned by the display process described above, and applies force. Then, the pointing device 12 gives a position signal that can identify the current contact position to the computer device 14.
- the contact position detection unit 147 is a coordinate value indicating a position where the user is currently in contact! Hereinafter, this is referred to as a contact coordinate value) (step S122).
- the contact finger determination unit 148 checks the allocation status held in the component allocation unit 144, and the current position that substantially matches the contact coordinate value converted by the contact position detection unit 147. A position coordinate value is selected (step S123). Further, the contact finger determination unit 148 recognizes the GUI component assigned to the selected current position coordinate value (step S124).
- the function execution unit 149 executes the function assigned to the GUI component recognized by the contact finger determination unit 148 (step S125).
- the user interface system 11 operates in the same manner as in the above-described embodiment, and as a result, the user interface system that can be operated quickly by the user. System 11 can be provided.
- GUI parts are uniquely assigned to each finger, so that the user interface systems 1 and 11 can use only a maximum of five GUI parts. Therefore, as shown in FIG. 13, a state transition function is assigned to a predetermined finger (thumb F in the illustrated example), and the user force S is used to display the display screen 21 or the operation area.
- the component allocation unit 34 or 144 changes the GUI component allocated to the remaining fingers. This makes it possible to assign up to 8 GUI parts to each finger before and after specifying the state transition function.
- GUI parts can be assigned to fingers as follows.
- each of the display screen 21 and the operation area 121 is divided into a plurality of partial areas by a predetermined method.
- the display screen 21 or the operation area 121 is divided into four rectangular areas.
- a plurality of target devices are uniquely assigned to each partial area.
- One device may be assigned to a plurality of areas.
- the component allocation unit 34 or 144 changes the allocation of fingers and GUI components for each area (each device).
- a certain area is allocated to the navigation system, and the parts allocation unit 34 or 144 has a memory for the thumb F in this area.
- the other area is allocated to the audio device, and the component allocation unit 34 or 144 has a GUI component of mode change for the thumb F in this area.
- the contact finger determination unit 38 or 148 refers to the allocation status held by the component allocation unit 34 or 144, and based on the current position coordinate value, the display screen 21 or the operation area. Determine which area of 121 is operated with which finger, and the determination result power Identify the device to be controlled this time and the function to be executed by that device.
- GUI parts can be assigned to fingers as follows.
- the parts allocating unit 34 or 144 as shown in FIG.
- the thumb F is not assigned a GUI part
- the GUI part is assigned to the remaining fingers for each device.
- audio device is played on index finger F.
- GUI part C for the feed function and the little finger F are assigned GUI part C for the volume control function.
- GUI parts are assigned to each finger other than the thumb.
- a GUI component SC-SC for specifying a device is synthesized at a predetermined position in the GUI image in order to select a device to be controlled.
- GUI image creation unit 35 or 145 is configured so that a predetermined finger (thumb F) contact position is determined.
- the contact finger determination unit 38 or 148 refers to the allocation status held by the component allocation unit 34 or 144, and displays it based on the current position coordinate value. It is determined with which finger the screen 21 or the operation area 121 is operated, and the function to be executed by the device to be controlled this time is specified.
- GUI parts can be assigned to fingers as follows.
- the component assigning unit 34 or 144 has a predetermined number of fingers (three in the illustrated example) GUI components C on a predetermined finger (middle finger F in the illustrated example).
- GUI image creation unit 35 or 145 displays the current position coordinate value of the middle finger F.
- GUI image creation unit 35 or 145 also includes a mark indicating the reference position (that is, the position indicated by the coordinate value P) on the current position coordinate value P.
- the user After such a GUI image is displayed, the user first touches the reference position P on the display screen 21 or the operation area 121. After that, the user can specify what he wants to specify this time.
- the first position signal that can specify the reference position P is sent from the mapping device 12.
- the contact finger determination unit 38 or 148 holds the reception time of the first position signal (hereinafter referred to as the reference time) and also determines the type of finger (in this case, the middle finger F) in response to the reception of the first position signal. ). In addition, the finger contact determination unit 3
- 8 or 148 When the second position signal is sent, 8 or 148 indicates the continuity between the first contact of the finger and the second contact based on the time difference between the reception time of the second position signal and the reference time.
- each GUI component C — C is the reference point P
- It is preferably composed of an arrow indicating the direction with respect to 3 1 3 3.
- GUI parts C-C are assigned to one finger F.
- GUI parts may be assigned to each of a plurality of fingers.
- the user specified two points.
- the present invention is not limited to this, and the user may specify the GUI component displayed on the display screen 21 or the operation area 121 by tracing.
- Such a designation method is suitable for a case where an optimum value within a certain range is set, for example, when the volume is adjusted.
- the component allocation unit 34 or 144 may change the GUI component assigned to each finger to another GUI component. It is assumed that the GUI parts to be assigned to each finger in the parts assignment part 34 or 144 are determined by default. Further, the user interface system 1 or 11 has an assignment change mode, and in the assignment change mode, as shown in FIG. 17, the GUI image creation unit 35 or 145 is illustratively a GUI component that can be changed. C — C overlaps each other
- a main menu as shown in Fig. 18A is displayed by the first position signal indicating the contact position of the finger, and a function to be executed is determined from the displayed main menu.
- the user interface system displays a submenu of functions determined by the first position signal as shown in FIG. 18B. At this time, each menu may be displayed around the finger contact position 1 801 represented by the first position signal. In this way, many GUI parts can be assigned to fingers by the hierarchical structure.
- 18A and 18B show a state where the FM button is selected by the first position signal.
- GUI parts displayed after being selected by the first position signal are not limited to icons, buttons, and arrows, and GUI parts such as a slider bar may be displayed.
- the display when the TV function is executed from the first position signal in the state where the screen of FIG. 19A is displayed is shown.
- the channel button is displayed as shown in Fig. 19B.
- the contact position represented by the first position signal is the position 1901.
- the finger 1902 farthest from the fixed display menu is detected without losing the relevance of all GUIs to the fingertip, and only the fingertip is detected.
- the GUI parts that can escape the already executed functions such as cancel and return are displayed in association with the fingertips. In this way, even if the hand is moved, a specific GUI component always follows the movement of the finger, so the user can operate with peace of mind.
- FIG. 20 shows a scene where the user is inputting characters using the software keyboard.
- the correction button is associated with only the finger 2001, when the user operates the GUI with a finger other than the finger 2001, the software keyboard interface responds to the finger other than the finger 2001.
- the correction button follows the finger 2001, so that the user can execute the correction function regardless of the position at which the finger 2001 touches the pointing device 12.
- the function associated with a specific finger is not limited to the correction function, and it is desirable to add a function frequently required by the user.
- the timing of function execution during GUI operation is
- the function may be executed when the user's finger is separated from the pointing device 12.
- the cancel button 2101 can be displayed as shown in FIG. 21, so when the user notices an error after pressing the pointing device 1 2 to execute the function, You can cancel the function you tried to execute.
- FIG. 21 also shows a case where the user keeps pressing the TV button 2102. At this time, the function of the pressed button is executed, and the next button to be displayed is not limited to the cancel button, and GUI parts such as the setting button 2103 that can make detailed settings before executing the function But you can.
- the user interface system displays the hand shape.
- the image creation unit 145 draws a GUI part associated with the fingertip, it displays the hand shape.
- the visibility of the screen can be improved by performing the processing without performing the above.
- Fig. 22 shows an example of GUI component display when the hand shape is not displayed.
- the hand shape 2 201 is displayed. /.
- the finger names displayed on the GUI component are not limited to this.
- a method of displaying the corresponding number such as 1 for the thumb and 2 for the index finger, can be considered.
- the corresponding number is preferably displayed on the display device 16.
- GUI component to be allocated to the component allocation units 34 and 144 does not issue a single control command.
- a character string re-conversion function as shown in FIG. 23 can be raised. This is because the computer side cannot determine what should be reconverted by only the reconversion control command that is a GUI component alone attached to the fingertip, and an object to be reconverted is required.
- the finger 2301 is assigned to the re-transformation, and the text existing at the position indicated by the finger 2301 can be re-transformed.
- a GUI component such as a slider bar that allows continuous values to be input may be displayed. In this case, the contact position detection units 37 and 146 detect the movement of the finger position from the state in which the contact is detected, thereby detecting the drag operation and operating the slider bar.
- a GUI component that can input continuous values such as a slider bar is displayed according to the range to be adjusted.
- the hand-shaped movement on the contact position detection unit when the GUI part is displayed as shown in FIG. 24A performs a pulling force movement as shown in FIG. 24B.
- the slider bar can be moved up and down by moving the position while keeping the finger corresponding to the GUI part of each finger displayed in FIG. 24A in contact.
- this device includes finger identification units 33 and 143, and determines which function is assigned to which finger.
- the timing at which the function is executed is when the finger is touched in the contact position detection units 37 and 147.
- the function is assigned to each finger as shown in Fig. 25A
- the finger is bent and only one finger is extended as shown in Fig. 25B, that is, the fingertip can be detected only. It is possible to perform the function assigned to that finger when it is in a state.
- FIG. 25B is an example when the scale changing function assigned to the finger 2501 is executed. As shown in FIG. 25B, the finger 2501 remains at the end, the function is executed, and the scale is changed with the slider bar.
- the user interface system according to the present invention is suitable for a computer-related product or the like that requires good usability.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP06730193A EP1865404A4 (en) | 2005-03-28 | 2006-03-28 | USER INTERFACE SYSTEM |
US11/884,955 US7810050B2 (en) | 2005-03-28 | 2006-03-28 | User interface system |
JP2007510518A JP4702959B2 (ja) | 2005-03-28 | 2006-03-28 | ユーザインタフェイスシステム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2005092434 | 2005-03-28 | ||
JP2005-092434 | 2005-03-28 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006104132A1 true WO2006104132A1 (ja) | 2006-10-05 |
Family
ID=37053388
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2006/306245 WO2006104132A1 (ja) | 2005-03-28 | 2006-03-28 | ユーザインタフェイスシステム |
Country Status (5)
Country | Link |
---|---|
US (1) | US7810050B2 (ja) |
EP (1) | EP1865404A4 (ja) |
JP (1) | JP4702959B2 (ja) |
CN (1) | CN101180599A (ja) |
WO (1) | WO2006104132A1 (ja) |
Cited By (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008078603A1 (ja) | 2006-12-22 | 2008-07-03 | Panasonic Corporation | ユーザインターフェイス装置 |
JP2008234594A (ja) * | 2007-03-23 | 2008-10-02 | Denso Corp | 操作入力装置 |
WO2009031457A1 (en) * | 2007-09-04 | 2009-03-12 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
WO2009031633A1 (en) * | 2007-09-04 | 2009-03-12 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
JP2009075685A (ja) * | 2007-09-19 | 2009-04-09 | Sony Corp | 画像処理装置および画像処理方法、並びにプログラム |
JP2009276926A (ja) * | 2008-05-13 | 2009-11-26 | Ntt Docomo Inc | 情報処理装置及び情報処理装置の表示情報編集方法 |
JP2009301094A (ja) * | 2008-06-10 | 2009-12-24 | Sharp Corp | 入力装置及び入力装置の制御方法 |
JP2010009311A (ja) * | 2008-06-26 | 2010-01-14 | Panasonic Corp | ユーザインタフェイス装置 |
JP2010140458A (ja) * | 2008-03-19 | 2010-06-24 | Denso Corp | 表示方法、表示装置、及び操作入力装置 |
US20110206239A1 (en) * | 2008-03-12 | 2011-08-25 | Denso Corporation | Input apparatus, remote controller and operating device for vehicle |
JP2011180843A (ja) * | 2010-03-01 | 2011-09-15 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
JP2012512453A (ja) * | 2008-12-15 | 2012-05-31 | ソニー エリクソン モバイル コミュニケーションズ, エービー | 光学マーカを有するユーザ入力デバイスを検出する電子デバイス、システム、方法およびコンピュータプログラム |
JP2012527657A (ja) * | 2009-05-18 | 2012-11-08 | 日本電気株式会社 | タッチスクリーン、関連した操作方法及びシステム |
JP2013093031A (ja) * | 2012-11-16 | 2013-05-16 | Sony Corp | 情報処理装置、記憶媒体、情報処理システム、および情報処理方法、並びにプログラム |
WO2013136776A1 (ja) * | 2012-03-15 | 2013-09-19 | パナソニック株式会社 | ジェスチャ入力操作処理装置 |
JP2013257904A (ja) * | 2008-03-12 | 2013-12-26 | Atlab Inc | 接触センサ装置 |
WO2014045953A1 (ja) * | 2012-09-20 | 2014-03-27 | ソニー株式会社 | 情報処理装置および方法、並びにプログラム |
JP2014096107A (ja) * | 2012-11-12 | 2014-05-22 | Toshiba Alpine Automotive Technology Corp | 車室内カメラのスキャン方法、車室内カメラにおける対象物の探索方法及び車室内カメラ |
US8766910B2 (en) * | 2007-07-04 | 2014-07-01 | Cypress Semiconductor Corporation | Capacitive sensing control knob |
JP2014179072A (ja) * | 2013-03-14 | 2014-09-25 | Honda Motor Co Ltd | 三次元指先トラッキング |
JP2014179097A (ja) * | 2013-03-13 | 2014-09-25 | Honda Motor Co Ltd | ポインティングによる情報クエリ |
WO2015105044A1 (ja) * | 2014-01-10 | 2015-07-16 | 日本電気株式会社 | インターフェース装置、可搬装置、制御装置、モジュール、制御方法およびプログラム記憶媒体 |
JP2015534304A (ja) * | 2012-08-20 | 2015-11-26 | ウォルジエン エルエルシー | ビデオ通話センター |
JP2016507112A (ja) * | 2013-03-05 | 2016-03-07 | インテル コーポレイション | 複合的な知覚感知入力の対話 |
JP2017073090A (ja) * | 2015-10-09 | 2017-04-13 | 献平 陳 | データ入力装置 |
JP2017529024A (ja) * | 2014-06-30 | 2017-09-28 | ザ ビデオ コール センター,エルエルシー | ビデオ通話センター |
JP2020530631A (ja) * | 2017-08-14 | 2020-10-22 | シェンジェン ピーアールテック カンパニー リミテッド | インタラクション位置決定方法、システム、記憶媒体、およびスマートデバイス |
WO2023140340A1 (ja) | 2022-01-19 | 2023-07-27 | メタマティクス プライベート リミテッド | 指識別に基づくユーザー・インターフェースを実現するためのシステム、方法及びそのためのプログラム |
Families Citing this family (221)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US7958456B2 (en) | 2005-12-23 | 2011-06-07 | Apple Inc. | Scrolling list with floating adjacent index symbols |
US7996792B2 (en) * | 2006-09-06 | 2011-08-09 | Apple Inc. | Voicemail manager for portable multifunction device |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US20080168395A1 (en) * | 2007-01-07 | 2008-07-10 | Bas Ording | Positioning a Slider Icon on a Portable Multifunction Device |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
CN101369249B (zh) * | 2007-08-14 | 2011-08-17 | 国际商业机器公司 | 标识软件的gui部件的方法和装置 |
KR102559017B1 (ko) | 2007-09-24 | 2023-07-25 | 애플 인크. | 전자 장치 내의 내장형 인증 시스템들 |
KR20090058073A (ko) * | 2007-12-04 | 2009-06-09 | 삼성전자주식회사 | 단말 및 그의 기능 수행 방법 |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8405621B2 (en) * | 2008-01-06 | 2013-03-26 | Apple Inc. | Variable rate media playback methods for electronic devices with touch interfaces |
KR101297144B1 (ko) * | 2008-03-19 | 2013-08-21 | 가부시키가이샤 덴소 | 차량용 조작 입력 장치 |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
JP4982430B2 (ja) * | 2008-05-27 | 2012-07-25 | 株式会社エヌ・ティ・ティ・ドコモ | 文字入力装置及び文字入力方法 |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US9477396B2 (en) | 2008-11-25 | 2016-10-25 | Samsung Electronics Co., Ltd. | Device and method for providing a user interface |
EP3232315B1 (en) * | 2008-11-25 | 2021-08-25 | Samsung Electronics Co., Ltd. | Device and method for providing a user interface |
WO2010067118A1 (en) | 2008-12-11 | 2010-06-17 | Novauris Technologies Limited | Speech recognition involving a mobile device |
JP4771183B2 (ja) * | 2009-01-30 | 2011-09-14 | 株式会社デンソー | 操作装置 |
US8839155B2 (en) * | 2009-03-16 | 2014-09-16 | Apple Inc. | Accelerated scrolling for a multifunction device |
US8572513B2 (en) | 2009-03-16 | 2013-10-29 | Apple Inc. | Device, method, and graphical user interface for moving a current position in content at a variable scrubbing rate |
JP4840620B2 (ja) * | 2009-04-30 | 2011-12-21 | 株式会社デンソー | 車載用電子機器操作装置 |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US10255566B2 (en) | 2011-06-03 | 2019-04-09 | Apple Inc. | Generating and processing task items that represent tasks to perform |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US20100310136A1 (en) * | 2009-06-09 | 2010-12-09 | Sony Ericsson Mobile Communications Ab | Distinguishing right-hand input and left-hand input based on finger recognition |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
DE102009033882A1 (de) * | 2009-07-20 | 2011-01-27 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren zum Betrieb eines Dialogsystems mit einem Touchscreen |
JP2011028366A (ja) * | 2009-07-22 | 2011-02-10 | Sony Corp | 操作制御装置および操作制御方法 |
US8624933B2 (en) | 2009-09-25 | 2014-01-07 | Apple Inc. | Device, method, and graphical user interface for scrolling a multi-section document |
US20110102333A1 (en) * | 2009-10-30 | 2011-05-05 | Wayne Carl Westerman | Detection of Gesture Orientation on Repositionable Touch Surface |
US20110113329A1 (en) * | 2009-11-09 | 2011-05-12 | Michael Pusateri | Multi-touch sensing device for use with radiological workstations and associated methods of use |
DE102009059866A1 (de) * | 2009-12-21 | 2011-06-22 | Volkswagen AG, 38440 | Verfahren zum Betreiben einer Bedienvorrichtung und Bedienvorrichtung dafür, insbesondere in einem Fahrzeug |
JP4719296B1 (ja) * | 2009-12-25 | 2011-07-06 | 株式会社東芝 | 情報処理装置及び情報処理方法 |
US20110163967A1 (en) * | 2010-01-06 | 2011-07-07 | Imran Chaudhri | Device, Method, and Graphical User Interface for Changing Pages in an Electronic Document |
US8510677B2 (en) | 2010-01-06 | 2013-08-13 | Apple Inc. | Device, method, and graphical user interface for navigating through a range of values |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
JP2011210188A (ja) * | 2010-03-30 | 2011-10-20 | Sony Corp | 画像処理装置、画像表示方法、画像表示プログラム、および画像表示プログラムを記録した記録媒体 |
EP2395418A3 (en) | 2010-06-14 | 2015-10-28 | Sony Computer Entertainment Inc. | Information processor, device, and information processing system |
CN107665089B (zh) * | 2010-08-12 | 2021-01-22 | 谷歌有限责任公司 | 触摸屏上的手指识别 |
CN102375714A (zh) * | 2010-08-16 | 2012-03-14 | 慧荣科技股份有限公司 | 计算机系统及相关的影像绘图装置和显示装置 |
JP5434997B2 (ja) * | 2010-10-07 | 2014-03-05 | 株式会社ニコン | 画像表示装置 |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US8553001B2 (en) * | 2011-03-22 | 2013-10-08 | Adobe Systems Incorporated | Methods and apparatus for determining local coordinate frames for a human hand |
US20130132903A1 (en) * | 2011-03-22 | 2013-05-23 | Aravind Krishnaswamy | Local Coordinate Frame User Interface for Multitouch-Enabled Applications |
US8593421B2 (en) * | 2011-03-22 | 2013-11-26 | Adobe Systems Incorporated | Local coordinate frame user interface for multitouch-enabled devices |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US9043042B2 (en) * | 2011-07-19 | 2015-05-26 | GM Global Technology Operations LLC | Method to map gaze position to information display in vehicle |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
US9772759B2 (en) | 2011-10-14 | 2017-09-26 | Apple Inc. | Device, method, and graphical user interface for data input using virtual sliders |
US10477191B2 (en) * | 2011-11-21 | 2019-11-12 | Nikon Corporation | Display device, and display control program |
JP5830644B2 (ja) * | 2011-12-06 | 2015-12-09 | パナソニックIpマネジメント株式会社 | 下受けピン配置判定装置および下受けピン配置判定方法 |
JP2013120434A (ja) * | 2011-12-06 | 2013-06-17 | Denso It Laboratory Inc | 操作者識別装置及び方法並びに車載用ナビゲーション装置 |
WO2013095678A1 (en) | 2011-12-23 | 2013-06-27 | Intel Corporation | Mechanism to provide feedback regarding computing system command gestures |
TW201335833A (zh) * | 2012-02-29 | 2013-09-01 | Hon Hai Prec Ind Co Ltd | 電子設備編輯工具變更方法及系統 |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
FR2989207A1 (fr) * | 2012-04-06 | 2013-10-11 | Bic Soc | Orientation d'une tablette |
JP5594910B2 (ja) * | 2012-04-09 | 2014-09-24 | 京セラドキュメントソリューションズ株式会社 | 表示入力装置及びこれを備えた画像形成装置 |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
EP2850510A2 (en) * | 2012-05-18 | 2015-03-25 | Apple Inc. | Device, method, and graphical user interface for manipulating user interfaces based on fingerprint sensor inputs |
DE102012011177A1 (de) * | 2012-06-06 | 2013-12-24 | GM Global Technology Operations LLC (n. d. Ges. d. Staates Delaware) | Verfahren zur Bedienung von Funktionen einesFahrzeuges sowie entsprechende Vorrichtung |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US10048779B2 (en) * | 2012-06-30 | 2018-08-14 | Hewlett-Packard Development Company, L.P. | Virtual hand based on combined data |
US9305229B2 (en) | 2012-07-30 | 2016-04-05 | Bruno Delean | Method and system for vision based interfacing with a computer |
US9041689B1 (en) * | 2012-08-02 | 2015-05-26 | Amazon Technologies, Inc. | Estimating fingertip position using image analysis |
KR102026882B1 (ko) * | 2012-09-10 | 2019-09-30 | 삼성전자주식회사 | 터치 스크린을 포함한 전자 장치에서 다섯 손가락을 구별하기 위한 방법 및 장치 |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
US9904414B2 (en) * | 2012-12-10 | 2018-02-27 | Seiko Epson Corporation | Display device, and method of controlling display device |
CN103926999B (zh) * | 2013-01-16 | 2017-03-01 | 株式会社理光 | 手掌开合手势识别方法和装置、人机交互方法和设备 |
KR102516577B1 (ko) | 2013-02-07 | 2023-04-03 | 애플 인크. | 디지털 어시스턴트를 위한 음성 트리거 |
WO2014121522A1 (en) * | 2013-02-08 | 2014-08-14 | Motorola Solutions, Inc. | Method and apparatus for managing user interface elements on a touch-screen device |
US9081418B1 (en) | 2013-03-11 | 2015-07-14 | Rawles Llc | Obtaining input from a virtual user interface |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
AU2014233517B2 (en) | 2013-03-15 | 2017-05-25 | Apple Inc. | Training an at least partial voice command system |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101959188B1 (ko) | 2013-06-09 | 2019-07-02 | 애플 인크. | 디지털 어시스턴트의 둘 이상의 인스턴스들에 걸친 대화 지속성을 가능하게 하기 위한 디바이스, 방법 및 그래픽 사용자 인터페이스 |
KR101809808B1 (ko) | 2013-06-13 | 2017-12-15 | 애플 인크. | 음성 명령에 의해 개시되는 긴급 전화를 걸기 위한 시스템 및 방법 |
CN105453026A (zh) | 2013-08-06 | 2016-03-30 | 苹果公司 | 基于来自远程设备的活动自动激活智能响应 |
US9857971B2 (en) * | 2013-12-02 | 2018-01-02 | Industrial Technology Research Institute | System and method for receiving user input and program storage medium thereof |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
KR20150087666A (ko) * | 2014-01-22 | 2015-07-30 | 삼성전자주식회사 | 휴대 단말의 입력 인터페이스 제공 방법 및 그 장치 |
KR102184402B1 (ko) * | 2014-03-06 | 2020-11-30 | 엘지전자 주식회사 | 글래스 타입의 이동 단말기 |
JP2015170102A (ja) * | 2014-03-06 | 2015-09-28 | トヨタ自動車株式会社 | 情報処理装置 |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
TWI566107B (zh) | 2014-05-30 | 2017-01-11 | 蘋果公司 | 用於處理多部分語音命令之方法、非暫時性電腦可讀儲存媒體及電子裝置 |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
DE102014012550A1 (de) * | 2014-08-20 | 2016-02-25 | e.solutions GmbH | Kraftfahrzeug-Bedienungsvorrichtung |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US9606986B2 (en) | 2014-09-29 | 2017-03-28 | Apple Inc. | Integrated word N-gram and class M-gram language models |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US20160179323A1 (en) * | 2014-12-18 | 2016-06-23 | Avaya Inc. | In-place web communications related to content of a webpage |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
WO2017034425A1 (en) * | 2015-08-21 | 2017-03-02 | Motorola Solutions, Inc. | System and method for disambiguating touch interactions |
JP2017047781A (ja) * | 2015-09-01 | 2017-03-09 | 三菱自動車工業株式会社 | 車載情報処理装置 |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
KR102324716B1 (ko) * | 2015-11-02 | 2021-11-10 | 에스케이텔레콤 주식회사 | 멀티 터치 패턴을 이용한 유저 인터페이스 메뉴 디스플레이 장치 및 방법 |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
DE102016003072A1 (de) | 2016-03-12 | 2017-09-14 | Audi Ag | Bedienvorrichtung und Verfahren zum Erfassen einer Benutzerauswahl zumindest einer Bedienfuktion der Bedienvorrichtung |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
CN107203320A (zh) * | 2016-03-18 | 2017-09-26 | 大陆汽车投资(上海)有限公司 | 基于多点触摸的用户界面控制方法 |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179309B1 (en) | 2016-06-09 | 2018-04-23 | Apple Inc | Intelligent automated assistant in a home environment |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
EP3355178B1 (en) * | 2017-01-25 | 2023-05-24 | Canon Medical Systems Corporation | Ultrasound diagnosis apparatus |
JP7175605B2 (ja) * | 2017-01-25 | 2022-11-21 | キヤノンメディカルシステムズ株式会社 | 超音波診断装置 |
DE102017101669A1 (de) * | 2017-01-27 | 2018-08-02 | Trw Automotive Electronics & Components Gmbh | Verfahren zum Betreiben einer Mensch-Maschinen-Schnittstelle sowie Mensch-Maschinen-Schnittstelle |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | USER INTERFACE FOR CORRECTING RECOGNITION ERRORS |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | LOW-LATENCY INTELLIGENT AUTOMATED ASSISTANT |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | FAR-FIELD EXTENSION FOR DIGITAL ASSISTANT SERVICES |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
DE102017217914A1 (de) * | 2017-10-09 | 2019-04-11 | Bayerische Motoren Werke Aktiengesellschaft | Fortbewegungsmittel, Anwenderschnittstelle und Verfahren zum Bedienen einer Anwenderschnittstelle |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
DE102018100196A1 (de) * | 2018-01-05 | 2019-07-11 | Bcs Automotive Interface Solutions Gmbh | Verfahren zum Betreiben einer Mensch-Maschinen-Schnittstelle sowie Mensch-Maschinen-Schnittstelle |
DE102018100197A1 (de) * | 2018-01-05 | 2019-07-11 | Bcs Automotive Interface Solutions Gmbh | Verfahren zum Betreiben einer Mensch-Maschinen-Schnittstelle sowie Mensch-Maschinen-Schnittstelle |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
DK179822B1 (da) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | VIRTUAL ASSISTANT OPERATION IN MULTI-DEVICE ENVIRONMENTS |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
US11076039B2 (en) | 2018-06-03 | 2021-07-27 | Apple Inc. | Accelerated task performance |
US10838544B1 (en) * | 2019-08-21 | 2020-11-17 | Raytheon Company | Determination of a user orientation with respect to a touchscreen device |
US11409410B2 (en) | 2020-09-14 | 2022-08-09 | Apple Inc. | User input interfaces |
US20230136028A1 (en) * | 2021-11-03 | 2023-05-04 | Jin Alexander Yu | Ergonomic eyes-off-hand multi-touch input |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02132510A (ja) * | 1988-11-12 | 1990-05-22 | Sony Corp | 入力装置 |
JPH0683523A (ja) * | 1991-12-13 | 1994-03-25 | American Teleph & Telegr Co <Att> | ワークサーフェスシステム及びその実現方法、及びセンサ利用方法及び人物識別方法 |
JPH1040014A (ja) * | 1996-06-13 | 1998-02-13 | Internatl Business Mach Corp <Ibm> | 仮想ポインティング・デバイスの生成指示方法及び装置 |
JP2001070293A (ja) * | 1999-09-06 | 2001-03-21 | Toshiba Corp | X線診断装置 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE69204045T2 (de) * | 1992-02-07 | 1996-04-18 | Ibm | Verfahren und Vorrichtung zum optischen Eingang von Befehlen oder Daten. |
US6600480B2 (en) * | 1998-12-31 | 2003-07-29 | Anthony James Francis Natoli | Virtual reality keyboard system and method |
EP2256605B1 (en) * | 1998-01-26 | 2017-12-06 | Apple Inc. | Method and apparatus for integrating manual input |
AU780674B2 (en) | 1999-10-27 | 2005-04-07 | Keyless Systems Ltd. | Integrated keypad system |
US6519607B1 (en) * | 1999-10-28 | 2003-02-11 | Hewlett-Packard Company | Image driven operating system |
US6611253B1 (en) * | 2000-09-19 | 2003-08-26 | Harel Cohen | Virtual input environment |
US20020075334A1 (en) * | 2000-10-06 | 2002-06-20 | Yfantis Evangelos A. | Hand gestures and hand motion for replacing computer mouse events |
FI113094B (fi) * | 2000-12-15 | 2004-02-27 | Nokia Corp | Parannettu menetelmä ja järjestely toiminnon aikaansaamiseksi elektronisessa laitteessa ja elektroninen laite |
WO2002059868A1 (en) * | 2001-01-24 | 2002-08-01 | Interlink Electronics, Inc. | Game and home entertainment device remote control |
CN100350363C (zh) * | 2001-12-21 | 2007-11-21 | 拉尔夫·特拉赫特 | 灵活的计算机输入系统 |
US7461356B2 (en) * | 2002-06-03 | 2008-12-02 | Fuji Xerox Co., Ltd. | Function control unit and method thereof |
CN1570960A (zh) | 2004-05-09 | 2005-01-26 | 万发良 | 一种以手指动作识别为基础的输入装置 |
-
2006
- 2006-03-28 WO PCT/JP2006/306245 patent/WO2006104132A1/ja active Application Filing
- 2006-03-28 CN CN200680010552.4A patent/CN101180599A/zh active Pending
- 2006-03-28 JP JP2007510518A patent/JP4702959B2/ja not_active Expired - Fee Related
- 2006-03-28 EP EP06730193A patent/EP1865404A4/en not_active Withdrawn
- 2006-03-28 US US11/884,955 patent/US7810050B2/en not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02132510A (ja) * | 1988-11-12 | 1990-05-22 | Sony Corp | 入力装置 |
JPH0683523A (ja) * | 1991-12-13 | 1994-03-25 | American Teleph & Telegr Co <Att> | ワークサーフェスシステム及びその実現方法、及びセンサ利用方法及び人物識別方法 |
JPH1040014A (ja) * | 1996-06-13 | 1998-02-13 | Internatl Business Mach Corp <Ibm> | 仮想ポインティング・デバイスの生成指示方法及び装置 |
JP2001070293A (ja) * | 1999-09-06 | 2001-03-21 | Toshiba Corp | X線診断装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP1865404A4 * |
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008078603A1 (ja) | 2006-12-22 | 2008-07-03 | Panasonic Corporation | ユーザインターフェイス装置 |
JP5179378B2 (ja) * | 2006-12-22 | 2013-04-10 | パナソニック株式会社 | ユーザインターフェイス装置 |
EP2124139A4 (en) * | 2006-12-22 | 2013-01-09 | Panasonic Corp | User interface device |
EP2124139A1 (en) * | 2006-12-22 | 2009-11-25 | Panasonic Corporation | User interface device |
US8284168B2 (en) | 2006-12-22 | 2012-10-09 | Panasonic Corporation | User interface device |
JP2008234594A (ja) * | 2007-03-23 | 2008-10-02 | Denso Corp | 操作入力装置 |
US8766910B2 (en) * | 2007-07-04 | 2014-07-01 | Cypress Semiconductor Corporation | Capacitive sensing control knob |
US8405626B2 (en) | 2007-09-04 | 2013-03-26 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
KR101198727B1 (ko) | 2007-09-04 | 2012-11-12 | 캐논 가부시끼가이샤 | 화상투영장치 및 그 제어 방법 |
EP2191352A4 (en) * | 2007-09-04 | 2016-12-28 | Canon Kk | IMAGE PROJECTION DEVICE AND CONTROL METHOD THEREFOR |
WO2009031457A1 (en) * | 2007-09-04 | 2009-03-12 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
WO2009031633A1 (en) * | 2007-09-04 | 2009-03-12 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
US8118433B2 (en) | 2007-09-04 | 2012-02-21 | Canon Kabushiki Kaisha | Image projection apparatus and control method for same |
JP2009075685A (ja) * | 2007-09-19 | 2009-04-09 | Sony Corp | 画像処理装置および画像処理方法、並びにプログラム |
US8896535B2 (en) | 2007-09-19 | 2014-11-25 | Sony Corporation | Image processing apparatus and method, and program therefor |
US8643598B2 (en) | 2007-09-19 | 2014-02-04 | Sony Corporation | Image processing apparatus and method, and program therefor |
US20110206239A1 (en) * | 2008-03-12 | 2011-08-25 | Denso Corporation | Input apparatus, remote controller and operating device for vehicle |
JP2013257904A (ja) * | 2008-03-12 | 2013-12-26 | Atlab Inc | 接触センサ装置 |
JP2010140458A (ja) * | 2008-03-19 | 2010-06-24 | Denso Corp | 表示方法、表示装置、及び操作入力装置 |
JP2009276926A (ja) * | 2008-05-13 | 2009-11-26 | Ntt Docomo Inc | 情報処理装置及び情報処理装置の表示情報編集方法 |
JP2009301094A (ja) * | 2008-06-10 | 2009-12-24 | Sharp Corp | 入力装置及び入力装置の制御方法 |
JP2010009311A (ja) * | 2008-06-26 | 2010-01-14 | Panasonic Corp | ユーザインタフェイス装置 |
JP2012512453A (ja) * | 2008-12-15 | 2012-05-31 | ソニー エリクソン モバイル コミュニケーションズ, エービー | 光学マーカを有するユーザ入力デバイスを検出する電子デバイス、システム、方法およびコンピュータプログラム |
JP2012527657A (ja) * | 2009-05-18 | 2012-11-08 | 日本電気株式会社 | タッチスクリーン、関連した操作方法及びシステム |
JP2011180843A (ja) * | 2010-03-01 | 2011-09-15 | Sony Corp | 情報処理装置、情報処理方法、およびプログラム |
WO2013136776A1 (ja) * | 2012-03-15 | 2013-09-19 | パナソニック株式会社 | ジェスチャ入力操作処理装置 |
JPWO2013136776A1 (ja) * | 2012-03-15 | 2015-08-03 | パナソニックIpマネジメント株式会社 | ジェスチャ入力操作処理装置 |
JP7394822B2 (ja) | 2012-08-20 | 2023-12-08 | ウォルジエン エルエルシー | ビデオ通話センター |
JP2015534304A (ja) * | 2012-08-20 | 2015-11-26 | ウォルジエン エルエルシー | ビデオ通話センター |
JP2022046513A (ja) * | 2012-08-20 | 2022-03-23 | ウォルジエン エルエルシー | ビデオ通話センター |
JP2019033503A (ja) * | 2012-08-20 | 2019-02-28 | ウォルジエン エルエルシー | ビデオ通話センター |
WO2014045953A1 (ja) * | 2012-09-20 | 2014-03-27 | ソニー株式会社 | 情報処理装置および方法、並びにプログラム |
US10168784B2 (en) | 2012-09-20 | 2019-01-01 | Sony Corporation | Information processing apparatus and method, and program |
US10754435B2 (en) | 2012-09-20 | 2020-08-25 | Sony Corporation | Information processing apparatus and method, and program |
JP2014096107A (ja) * | 2012-11-12 | 2014-05-22 | Toshiba Alpine Automotive Technology Corp | 車室内カメラのスキャン方法、車室内カメラにおける対象物の探索方法及び車室内カメラ |
JP2013093031A (ja) * | 2012-11-16 | 2013-05-16 | Sony Corp | 情報処理装置、記憶媒体、情報処理システム、および情報処理方法、並びにプログラム |
JP2016507112A (ja) * | 2013-03-05 | 2016-03-07 | インテル コーポレイション | 複合的な知覚感知入力の対話 |
JP2014179097A (ja) * | 2013-03-13 | 2014-09-25 | Honda Motor Co Ltd | ポインティングによる情報クエリ |
JP2014179072A (ja) * | 2013-03-14 | 2014-09-25 | Honda Motor Co Ltd | 三次元指先トラッキング |
WO2015105044A1 (ja) * | 2014-01-10 | 2015-07-16 | 日本電気株式会社 | インターフェース装置、可搬装置、制御装置、モジュール、制御方法およびプログラム記憶媒体 |
JPWO2015105044A1 (ja) * | 2014-01-10 | 2017-03-23 | 日本電気株式会社 | インターフェース装置、可搬装置、制御装置、モジュール、制御方法およびコンピュータプログラム |
JP2017529024A (ja) * | 2014-06-30 | 2017-09-28 | ザ ビデオ コール センター,エルエルシー | ビデオ通話センター |
JP2017073090A (ja) * | 2015-10-09 | 2017-04-13 | 献平 陳 | データ入力装置 |
JP2020530631A (ja) * | 2017-08-14 | 2020-10-22 | シェンジェン ピーアールテック カンパニー リミテッド | インタラクション位置決定方法、システム、記憶媒体、およびスマートデバイス |
WO2023140340A1 (ja) | 2022-01-19 | 2023-07-27 | メタマティクス プライベート リミテッド | 指識別に基づくユーザー・インターフェースを実現するためのシステム、方法及びそのためのプログラム |
Also Published As
Publication number | Publication date |
---|---|
US7810050B2 (en) | 2010-10-05 |
JP4702959B2 (ja) | 2011-06-15 |
US20080163131A1 (en) | 2008-07-03 |
CN101180599A (zh) | 2008-05-14 |
JPWO2006104132A1 (ja) | 2008-09-11 |
EP1865404A1 (en) | 2007-12-12 |
EP1865404A4 (en) | 2012-09-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4702959B2 (ja) | ユーザインタフェイスシステム | |
US11048333B2 (en) | System and method for close-range movement tracking | |
JP6393341B2 (ja) | 投射型映像表示装置 | |
JP4666808B2 (ja) | 画像表示システム、画像表示方法、及び記憶媒体、プログラム | |
CN102906671B (zh) | 手势输入装置及手势输入方法 | |
US9389779B2 (en) | Depth-based user interface gesture control | |
KR101136153B1 (ko) | 지문 인식 또는 멀티 터치가 가능한 센서 그리드 방식의 투명 패널 사용자 입력 장치, 사용자 지문 인식 방법, 및 사용자 터치 인식 방법 | |
US8666115B2 (en) | Computer vision gesture based control of a device | |
US8085243B2 (en) | Input device and its method | |
US9448714B2 (en) | Touch and non touch based interaction of a user with a device | |
US20120327125A1 (en) | System and method for close-range movement tracking | |
WO2007061057A1 (ja) | ジェスチャー入力装置、及び方法 | |
TWI463355B (zh) | 多點觸控介面之訊號處理裝置、訊號處理方法及使用者介面圖像選取方法 | |
US20150186037A1 (en) | Information processing device, information processing device control method, control program, and computer-readable recording medium | |
GB2470654A (en) | Data input on a virtual device using a set of objects. | |
CN101375235A (zh) | 信息处理装置 | |
CN104123031B (zh) | 图样互换方法与相关多点触控装置 | |
JP2011203830A (ja) | 投写システム及びその制御方法 | |
JP2011192081A (ja) | 情報処理装置及びその制御方法 | |
JP2005301668A (ja) | 情報処理装置および情報処理プログラム | |
KR102118421B1 (ko) | 카메라 커서 시스템 | |
JP2009282637A (ja) | 表示方法および表示装置 | |
JP5062898B2 (ja) | ユーザインタフェイス装置 | |
KR20120070133A (ko) | 카메라를 이용한 가상 터치 인터페이스 제공 장치 및 그 방법 | |
KR20160115022A (ko) | 비접촉식 마우스 장치 및 이 장치를 채용한 디지털 응용 시스템 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200680010552.4 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2007510518 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11884955 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2006730193 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
NENP | Non-entry into the national phase |
Ref country code: RU |
|
WWP | Wipo information: published in national office |
Ref document number: 2006730193 Country of ref document: EP |