[go: up one dir, main page]

GB2345183A - Monitoring speech presentation - Google Patents

Monitoring speech presentation Download PDF

Info

Publication number
GB2345183A
GB2345183A GB9828545A GB9828545A GB2345183A GB 2345183 A GB2345183 A GB 2345183A GB 9828545 A GB9828545 A GB 9828545A GB 9828545 A GB9828545 A GB 9828545A GB 2345183 A GB2345183 A GB 2345183A
Authority
GB
United Kingdom
Prior art keywords
speech
user
signals
speaker
received
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB9828545A
Other versions
GB2345183B (en
GB9828545D0 (en
Inventor
Robert Alexander Keiller
Richard Anthony Kirk
De Veen Evelyn Van
Gerhardt Paul Otto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Technology Europe Ltd
Original Assignee
Canon Research Centre Europe Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Research Centre Europe Ltd filed Critical Canon Research Centre Europe Ltd
Priority to GB0320871A priority Critical patent/GB2389220B/en
Priority to GB9828545A priority patent/GB2345183B/en
Publication of GB9828545D0 publication Critical patent/GB9828545D0/en
Publication of GB2345183A publication Critical patent/GB2345183A/en
Application granted granted Critical
Publication of GB2345183B publication Critical patent/GB2345183B/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B19/00Teaching not covered by other main groups of this subclass
    • G09B19/04Speaking
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Educational Administration (AREA)
  • Educational Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • User Interface Of Digital Computer (AREA)
  • Debugging And Monitoring (AREA)

Abstract

A system for monitoring speech, eg during a presentation to an audience, provides a feedback signal to the speaker and/or enables subsequent analysis. An audible, visible or vibrating alarm 11 may respond to characteristics such as talking too fast, leaving gaps or pauses, speaking in a monotone, or repeatedly using certain words or phrases; a datalog 15 feeds a printer 17 to provide statistics for study afterwards. In other embodiments, a monitoring system (31) can assess the rate of speaking (or possibly pitch and style) against pre-set parameters and control the speed at which an interacting computer (35) operates, Fig 2 (not shown); and the monitor (51) can match the measured delivery with an autocue display (57) by comparing with a known text file (55), Fig 3 (not shown).

Description

SPEECH MONITORING SYSTEM The present invention relates to an apparatus for and a method of monitoring speech. The invention has particular, although not exclusive, relevance to the monitoring of various characteristics of a user's speech signal in order to provide control signals for controlling the way in which the user gives a presentation to an audience and/or for controlling an interaction between the user and a computer system.
Another aspect of the present invention concerns the monitoring of certain characteristics of a speech signal representative of a known text for controlling an autocue or the like.
An audience's interest in a speech or presentation made by a speaker often depends upon the presentation and communication skills which the speaker has. For example, if the speaker speaks too quickly, then the audience will not be able to keep up with the information which is being presented to them and consequently they will lose interest in the remainder of the speech. Similarly, speakers who speak in a monotone are liable to send the audience to sleep, even though the content of the speech may be very interesting to the audience. Similarly, speakers who tend to leave large gaps or pauses within their presentations or who repeatedly use certain words or phrases, such as"basically"or"to be honest", are likely to annoy or bore the audience.
According to a first aspect, the present invention provides a system for monitoring a user's speech and for providing a feedback signal to the user for controlling the user's presentation. This system can be used, for example, to monitor for predetermined characteristics, such as speaking too fast, speaking in a monotone, leaving large gaps or pauses etc. If a speech recognition unit is used as well, then the system can also monitor for the occurrence of preselected words within the user's speech.
In existing computer-user interactive systems, the rate of interaction is usually the same each time the user interacts with the computer. According to second aspect, the present invention provides a system for monitoring a user's speech and for varying the interaction between the user and a computer system which the user is using in dependence upon the monitored speech. The system can, for example, try to infer the user's mood from the input speech and vary the interaction in dependence upon the inferred mood. In particular, if the user is talking fast and sounds bright and alert, then the computer system can increase its rate of interaction with the user by, for example, decreasing the response time of the software or by increasing the speaking rate of a speech synthesiser forming part of the computer system.
According to a third aspect, the present invention provides a system for monitoring the progress of a speaker as he/she delivers a known speech. The system can be used for identifying the approximate position within the known speech so as to control the automatic advancement of an autocue system.
Exemplary embodiments of the present invention will now be described with reference to the accompanying drawings in which: Figure 1 is a schematic block diagram illustrating a speech monitoring system according to a first aspect of the present invention; Figure 2 is a schematic block diagram illustrating a speech monitoring system according to a second aspect of the present invention; and Figure 3 is a schematic block diagram illustrating a speech monitoring system according to a third aspect of the present invention.
Figure 1, is a schematic block diagram illustrating a speech monitoring system according to a first aspect of the present invention. The purpose of the speech monitoring system shown in Figure 1 is to warn the user when he/she is making a speech that they are talking too fast or doing something else that impairs their communication. As shown, the monitoring system comprises a microphone 1, a speech processor and analysis unit 3, a speech recognition unit 5 and associated language and word models 7, a control unit 9, an alarm 11 a display 13, a data log 15 and a printer 17.
In operation, the microphone 1 converts an acoustic speech signal from the user into an equivalent electrical signal which is passed, via connector 21, to the speech processor and analysis unit 3. In this embodiment, the speech processor and analysis unit 3 is operable to convert the input speech signal from the microphone 1 into a sequence of parameter frames, each parameter frame representing a corresponding time frame of the input speech signal. The parameters in each parameter frame typically include cepstral coefficients and power/energy coefficients, which provide important information characteristic of the input speech signal. The sequence of parameter frames generated by the speech processor and analysis unit 3 are supplied, via conductor 23, to the speech recognition unit 5 which is operable to try to identify the presence of predetermined words and/or phrases within the user's speech for which there is a model in the language and word models 7. In this embodiment, the language and word models 7 are generated in advance by the user identifying words and/or phases which he/she tends to repeat and which may cause annoyance to the audience. In the event that the speech recognition unit 5 identifies one of the words and/or phases in the input speech, it outputs a corresponding signal to the control unit 9, via conductor 27.
In this embodiment, the speech processor and analysis unit 3 is also arranged to process the input speech to derive (i) a value indicative of the rate at which the user is speaking; (ii) a value indicative of whether or not the user is speaking in a monotone; and (iii) a value indicative of the gaps or pauses within the input speech, which values are passed to the control unit 9 via conductor 25. The control unit 9 is operable for receiving the output from the speech recognition unit 5 via conductor 27 and the above values output by the speech processor and analysis unit 3 via conductor 25 and to generate control signals for controlling the alarm 11 and the display 13. The alarm 11 may be, for example, an audible, visible or vibrating alarm. More specifically, the control unit 9 is operable to monitor the output from the speech recognition unit 5 and the speech processing and analysis unit 3 and to generate, if appropriate, a warning to the user either by activating the alarm 11 or by displaying appropriate information on the display 13 in order to inform the user that, for example, he is speaking too quickly, speaking in a monotone, is leaving large gaps in the speech or is repeatedly saying one of the words and/or phrases which the speech recognition unit is designed to identify. In this embodiment, the speech monitoring system operates in real time so that the speaker is given instantaneous feedback, so that he/she and can modify their presentation accordingly.
In addition to being able to be used in real time during a presentation, the speech monitoring system shown in Figure 1, and described above can also be used to monitor the entire speech given by a speaker and to provide an analysis of the speech after it has ended. The analysis might include the number of repetitions of selected words and/or phases, the average speaking rate, the variation of the speaking rate throughout the speech, the number of gaps or pauses within the speech, etc. This analysis is generated by the control unit and logged in the data log 15 which can be printed out to the printer 17 or displayed on the display 13, so that the speaker is given the appropriate feedback so that they can improve their presentation skills.
The speech monitoring system illustrated in Figure 1, might be built into a separate portable computer device or it might be implemented in computer software 1 on a personal computer which may also be assisting the user in their presentation by, for example, generating slides for display on an overhead projector (not shown).
Figure 2 is a schematic block diagram of a speech monitoring system according to a second aspect of the present invention. The purpose of the speech monitoring system shown in Figure 2 is to monitor a user's speech and to vary an interaction between the user and a computer system in dependence upon the mood of the user which is inferred from the user's speech. As shown, the speech monitoring system comprises a microphone 1, a speech monitoring unit 31, a control unit 33 and a computer application 35, all of which, in this embodiment, are operated in a common computer system 37.
In operation, the microphone 1 converts an acoustic speech signal from the user 39 into an equivalent electrical signal which is passed, via connector 41, to the speech monitoring unit 31. In this embodiment, the speech monitoring unit identifies the rate at which the user 39 is speaking (and optionally the user's pitch and style of speaking) and passes this information to the control unit 33 via the conductor 43. In this embodiment, the rate at which the user is speaking is identified by monitoring the beginning of known words in the input speech and their duration. The rate at which the user is speaking is then determined by comparing the durations with prestored durations loaded in memory.
From this information, the control unit 33 infers the speakers'mood and outputs a control signal on conductor 45 for controlling the computer application 35 accordingly. In particular, the control signal output by the control unit 33 changes the way in which the computer application 35 interacts (as represented by the doubleheaded arrow 47) with the user 39. For example, if the user 39 is talking quickly, the control unit 33 infers that the user is bright and alert, and accordingly causes the computer application 35 to react more quickly to the user 39. For example, where the computer application 35 includes a speech synthesiser, in response to the control unit 33 inferring that the user is bright and alert the computer application 35 increases the speed of speech synthesis. Alternatively, the computer application 35 might control the time resolution of the double click detection of a mouse button (not shown).
Conversely, if the speaker is speaking slowly, then the control unit 33 infers that the user is not alert and therefore causes the computer application to react more slowly to the user 39.
Figure 3, is a schematic block diagram of a speech monitoring system according to a third aspect of the present invention. The purpose of the speech monitoring system shown in Figure 3 is to track the position of a speaker as he/she delivers a known speech. As shown, the speech monitoring system comprises a microphone 1, a speech monitoring unit 51, a control unit 53, a text file 55 and a display 57, all of which, in this embodiment, form part of a single computer system 59.
In operation, the microphone 1 converts an acoustic speech signal of the user 39 into an equivalent electrical signal which is supplied, via conductor 61, to the speech monitoring unit 51. In this embodiment, the speech monitoring unit 51 identifies the words and/or syllables in the input speech signal and outputs a signal to the control unit via conductor 63 whenever a word and/or syllable is identified. In response, the control unit 53 counts the words and/or syllables and outputs a control signal to the text file 55 for identifying a part of the text file corresponding to the speech which the user 39 is about the speak. The identified part of the text file is then passed to the display 57, so that the user 59 can read the next part of the speech from the display 57.
In addition to being able to be used in an autocue system, the speech monitoring system illustrated in Figure 3 can be used in similar applications, such as in a play or in an opera, where the input speech will come from the stage whilst the displayed text will be displayed to a stage manager who can provide oral prompts whenever necessary. The advantage of such an automatic autocue system is that even if the operator/stage manager is temporarily distracted, the autocue will not loose its place within the known text.
In an alternative embodiment, the input speech could be passed to a speech recognition unit which is operable to compare the input speech signal with the text file in order to identify the next part of the text file to be displayed on the display 57.
As those skilled in the art will appreciate, the above embodiments could be combined to provide, for example, a system which can be used to promote the presentational skills of a speaker and which interacts with the speaker in dependence upon the mood of the speaker which is inferred from the speaker's speech. The system may also track the speaker's progress through his speech so as to control an automatic autocue.
The present invention is not limited to the exemplary embodiments described above, and various other modifications and embodiments will be apparent to those skilled in the art.

Claims (24)

  1. CLAIMS: 1. A speech monitoring system for use in promoting the presentational skills of a user, the apparatus comprising: means for receiving speech signals from the user; means for processing the received speech signals and for generating signals indicative of the occurrence of one or more predetermined events within the received speech signals; and control means responsive to said generated signals for outputting a feedback signal to said user for warning the user of the occurrence of the predetermined events within the received speech signals.
  2. 2. A speech monitoring system according to claim 1, further comprising a speech recognition unit operable for receiving the processed speech and for identifying one or more predetermined words and/or phases within the received speech signals by comparing the processed speech with stored reference models, and wherein said control means is operable to generate said feedback signal in dependence upon said recognition result.
  3. 3. A speech monitor according to claim 1 or 2, wherein said feedback signal is supplied to said user via at least one of an audible, visible or vibrating alarm.
  4. 4. A speech monitoring system according to any of claims 1 to 3, which is operable to provide said feedback signal in real time, so that said user can try to reduce the occurrence of said predetermined event or events within subsequent speech signals output by the user.
  5. 5. A speech monitor according to any preceding claim, further comprising means for generating a data log indicative of the occurrences of the predetermined event or events which are identified within the user's speech.
  6. 6. A speech monitor according to any preceding claim, wherein said predetermined event or events comprise at least one of speaking too fast, leaving too many gaps or pauses within the speech, speaking in a monotone or the like.
  7. 7. A computer system comprising: a computer application operable for interacting with a user; means for receiving speech signals from the user; means for processing the received speech signals and for deriving from the processed speech signals an indication of the mood of the user; and control means for controlling the interaction between the computer application and the user in dependence upon the derived indication of the mood of the user.
  8. 8. A system according to claim 7, wherein said received speech signals are processed in order to extract at least one of the speed at which the user is speaking, the user's pitch and the style of speech employed by the user.
  9. 9. A system according to claim 7 or 8, wherein said computer application comprises a speech synthesiser, and wherein said computer application is arranged to vary the speaking rate of the speech synthesiser in dependence upon the indication of the user's mood.
  10. 10. A system according to any of claims 7 to 9, wherein said control means is operable to vary the response rate of the computer application in dependence upon the indication of the user's mood.
  11. 11. A speech processing system comprising: means for storing signals representative of a known speech to be spoken by a speaker; means for receiving speech signals from the speaker as the speaker delivers the known speech; and means for determining from said stored signals and said received signals the position of the speaker within the known speech.
  12. 12. An autocue for prompting a user with the next part of a known speech comprising: means for storing signals representative of the known speech to be spoken by a speaker; means for receiving speech signals from the speaker as the speaker delivers the known speech; means for determining from said stored signals and said received signals the position of the speaker within the known speech; and means for informing the user of the next part of the known speech to be spoken by said speaker.
  13. 13. A system according to claim 12, wherein said user and said speaker are different people.
  14. 14. A system according to claim 12 or 13, wherein said next part of the known speech is displayed to said user on a display.
  15. 15. A system according to any of claims 11 to 14, wherein said determining means comprising a counter for counting words and/or syllables within the received speech signals.
  16. 16. A system according to any of claims 11 to 14, wherein said determining means comprises speech recognition means for comparing the received speech signals with the known speech signals.
  17. 17. A computer system for use by a user during the presentation of a speech, the computer system comprising: a speech monitoring system according to any of claims 1 to 6 for promoting the presentational skills of the user ; and a computer system according to any of claims 7 to 10 for varying an interaction between the computer system and the user.
  18. 18. A computer system according to claim 17, further comprising a speech processing system according to claim 11 or an autocue according to claim 12.
  19. 19. A method of promoting the presentational skills of a user, comprising steps of: receiving speech signals from the user; processing the received speech signals and generating signals indicative of the occurrence of one or more predetermined events within the received speech; and in response to the generated signals, outputting a feedback signal to the user for warning the user of the occurrence of the predetermined events within the received speech signals.
  20. 20. A method of varying the interaction between a computer application and a user, the method comprising the steps of: receiving speech signals from the user; processing the received speech signals and deriving from the processed speech signals an indication of the mood of the user; and controlling the interaction between the computer application and the user in dependence upon the derived mood of the user.
  21. 21. A speech processing method comprising the steps of: storing signals representative of a known speech to be spoken by a speaker: receiving speech signals from the speaker as the speaker delivers the known speech; and determining from the stored signals and the received signals the position of the speaker within the known speech.
  22. 22. A method of operating an autocue for prompting a user with the next part of a known speech, the method comprising steps of: storing signals representative of the known speech to be spoken by a speaker; receiving speech signals from the speaker as the speaker delivers the known speech; determining from the stored signals and the received signals the position of the speaker within the known speech; and informing the user of the next part of the known speech to be spoken by the speaker.
  23. 23. A data carrier programmed with instructions for carrying out the method according to any of claims 19 to 22 or for implementing the apparatus of any of claims 1 to 18.
  24. 24. A speech monitoring system or method substantially as hereinbefore described with reference to or as shown in any of Figures 1 to 3.
GB9828545A 1998-12-23 1998-12-23 Speech monitoring system Expired - Fee Related GB2345183B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
GB0320871A GB2389220B (en) 1998-12-23 1998-12-23 Speech monitoring system
GB9828545A GB2345183B (en) 1998-12-23 1998-12-23 Speech monitoring system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB9828545A GB2345183B (en) 1998-12-23 1998-12-23 Speech monitoring system

Publications (3)

Publication Number Publication Date
GB9828545D0 GB9828545D0 (en) 1999-02-17
GB2345183A true GB2345183A (en) 2000-06-28
GB2345183B GB2345183B (en) 2003-11-05

Family

ID=10844974

Family Applications (2)

Application Number Title Priority Date Filing Date
GB0320871A Expired - Fee Related GB2389220B (en) 1998-12-23 1998-12-23 Speech monitoring system
GB9828545A Expired - Fee Related GB2345183B (en) 1998-12-23 1998-12-23 Speech monitoring system

Family Applications Before (1)

Application Number Title Priority Date Filing Date
GB0320871A Expired - Fee Related GB2389220B (en) 1998-12-23 1998-12-23 Speech monitoring system

Country Status (1)

Country Link
GB (2) GB2389220B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1089246A2 (en) * 1999-10-01 2001-04-04 Siemens Aktiengesellschaft Method and apparatus for speech impediment therapy
WO2003052742A1 (en) * 2001-12-18 2003-06-26 Intel Corporation Voice-bearing light
FR2847706A1 (en) * 2002-11-27 2004-05-28 Vocebella Sa Voice transformation/speech recognition system having modules transforming input/providing representative characteristic and module processing set providing quality level selected signal
US7894849B2 (en) 2006-07-10 2011-02-22 Accenture Global Services Limited Mobile personal services platform for providing feedback
US20150269929A1 (en) * 2014-03-21 2015-09-24 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
CN109639935A (en) * 2019-01-25 2019-04-16 合肥学院 The automatic word extractor system and method for video record
GB2597975A (en) * 2020-08-13 2022-02-16 Vitec Group Plc Voice controlled studio apparatus

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE102020102468B3 (en) 2020-01-31 2021-08-05 Robidia GmbH Method for controlling a display device and display device for dynamic display of a predefined text

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2049190A (en) * 1979-04-27 1980-12-17 Friedman E Voice fluency monitor
GB2102171A (en) * 1981-06-24 1983-01-26 John Graham Parkhouse Speech aiding apparatus and method
EP0360909A1 (en) * 1988-09-30 1990-04-04 Siemens Audiologische Technik GmbH Speech practising apparatus
US5015179A (en) * 1986-07-29 1991-05-14 Resnick Joseph A Speech monitor
WO1995012189A1 (en) * 1993-10-27 1995-05-04 Freeman, Jacqueline, Carol Language analysis instrument
JPH07334075A (en) * 1994-06-03 1995-12-22 Hitachi Ltd Presentation support device
GB2317341A (en) * 1995-06-30 1998-03-25 Ron Samuel Speech-based biofeedback method and system

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB903888A (en) * 1960-03-22 1962-08-22 Autocue Great Britain Ltd Improvements in or relating to prompting or cueing equipment
GB1578054A (en) * 1978-05-26 1980-10-29 Autocue Holdings Ltd Transmitter unit of a display system
EP0692128B1 (en) * 1993-04-02 1997-06-18 Pinewood Associates Limited Information display apparatus
US6272461B1 (en) * 1999-03-22 2001-08-07 Siemens Information And Communication Networks, Inc. Method and apparatus for an enhanced presentation aid

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2049190A (en) * 1979-04-27 1980-12-17 Friedman E Voice fluency monitor
GB2102171A (en) * 1981-06-24 1983-01-26 John Graham Parkhouse Speech aiding apparatus and method
US5015179A (en) * 1986-07-29 1991-05-14 Resnick Joseph A Speech monitor
EP0360909A1 (en) * 1988-09-30 1990-04-04 Siemens Audiologische Technik GmbH Speech practising apparatus
WO1995012189A1 (en) * 1993-10-27 1995-05-04 Freeman, Jacqueline, Carol Language analysis instrument
JPH07334075A (en) * 1994-06-03 1995-12-22 Hitachi Ltd Presentation support device
GB2317341A (en) * 1995-06-30 1998-03-25 Ron Samuel Speech-based biofeedback method and system

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1089246A2 (en) * 1999-10-01 2001-04-04 Siemens Aktiengesellschaft Method and apparatus for speech impediment therapy
EP1089246A3 (en) * 1999-10-01 2002-07-17 Siemens Aktiengesellschaft Method and apparatus for speech impediment therapy
WO2003052742A1 (en) * 2001-12-18 2003-06-26 Intel Corporation Voice-bearing light
GB2399979A (en) * 2001-12-18 2004-09-29 Intel Corp Voice-bearing light
GB2399979B (en) * 2001-12-18 2005-10-26 Intel Corp Apparatus and system for guiding a speaker to the sensitivity region of a microphone
FR2847706A1 (en) * 2002-11-27 2004-05-28 Vocebella Sa Voice transformation/speech recognition system having modules transforming input/providing representative characteristic and module processing set providing quality level selected signal
WO2004049303A1 (en) * 2002-11-27 2004-06-10 Vocebella Sa Analysis of the vocal signal quality according to quality criteria
EP2037798B1 (en) * 2006-07-10 2012-10-31 Accenture Global Services Limited Mobile personal services platform for providing feedback
US7894849B2 (en) 2006-07-10 2011-02-22 Accenture Global Services Limited Mobile personal services platform for providing feedback
US8442578B2 (en) 2006-07-10 2013-05-14 Accenture Global Services Limited Mobile personal services platform for providing feedback
US20150269929A1 (en) * 2014-03-21 2015-09-24 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
US9344821B2 (en) * 2014-03-21 2016-05-17 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
US9779761B2 (en) 2014-03-21 2017-10-03 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
US10395671B2 (en) 2014-03-21 2019-08-27 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
US11189301B2 (en) 2014-03-21 2021-11-30 International Business Machines Corporation Dynamically providing to a person feedback pertaining to utterances spoken or sung by the person
CN109639935A (en) * 2019-01-25 2019-04-16 合肥学院 The automatic word extractor system and method for video record
GB2597975A (en) * 2020-08-13 2022-02-16 Vitec Group Plc Voice controlled studio apparatus
WO2022034335A1 (en) 2020-08-13 2022-02-17 The Vitec Group Plc Voice controlled studio apparatus
GB2597975B (en) * 2020-08-13 2023-04-26 Videndum Plc Voice controlled studio apparatus

Also Published As

Publication number Publication date
GB2389220B (en) 2004-02-25
GB0320871D0 (en) 2003-10-08
GB2345183B (en) 2003-11-05
GB2389220A (en) 2003-12-03
GB9828545D0 (en) 1999-02-17

Similar Documents

Publication Publication Date Title
US6358054B1 (en) Method and apparatus for teaching prosodic features of speech
JP4972645B2 (en) System and method for synchronizing sound and manually transcribed text
US20020086269A1 (en) Spoken language teaching system based on language unit segmentation
US7050978B2 (en) System and method of providing evaluation feedback to a speaker while giving a real-time oral presentation
US7818179B2 (en) Devices and methods providing automated assistance for verbal communication
US20150161908A1 (en) Method and apparatus for providing sensory information related to music
JP2002503353A (en) Reading aloud and pronunciation guidance device
Astolfi et al. Duration of voicing and silence periods of continuous speech in different acoustic environments
US20110144993A1 (en) Disfluent-utterance tracking system and method
JP2003186379A (en) Program for voice visualization processing, program for voice visualization figure display and for voice and motion image reproduction processing, program for training result display, voice-speech training apparatus and computer system
JP2003177784A (en) Method and device for extracting sound turning point, method and device for sound reproducing, sound reproducing system, sound delivery system, information providing device, sound signal editing device, recording medium for sound turning point extraction method program, recording medium for sound reproducing method program, recording medium for sound signal editing method program, sound turning point extraction method program, sound reproducing method program, and sound signal editing method program
GB2345183A (en) Monitoring speech presentation
JPH08286693A (en) Information processing device
KR101877559B1 (en) Method for allowing user self-studying language by using mobile terminal, mobile terminal for executing the said method and record medium for storing application executing the said method
Möller et al. Evaluating the speech output component of a smart-home system
JP3588596B2 (en) Karaoke device with singing special training function
US10896689B2 (en) Voice tonal control system to change perceived cognitive state
JP2844817B2 (en) Speech synthesis method for utterance practice
WO2024079605A1 (en) Assisting a speaker during training or actual performance of a speech
CN111182409B (en) Screen control method based on intelligent sound box, intelligent sound box and storage medium
CN113257246B (en) Prompting method, device, equipment, system and storage medium
JP5092311B2 (en) Voice evaluation device
Nelson et al. Effects of facial paralysis and audiovisual information on stop place identification
WO2002047067A2 (en) Improved speech transformation system and apparatus
KR20020087709A (en) A language training system

Legal Events

Date Code Title Description
PCNP Patent ceased through non-payment of renewal fee

Effective date: 20061223