[go: up one dir, main page]

WO1997037344A1 - Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal - Google Patents

Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal Download PDF

Info

Publication number
WO1997037344A1
WO1997037344A1 PCT/JP1996/000853 JP9600853W WO9737344A1 WO 1997037344 A1 WO1997037344 A1 WO 1997037344A1 JP 9600853 W JP9600853 W JP 9600853W WO 9737344 A1 WO9737344 A1 WO 9737344A1
Authority
WO
WIPO (PCT)
Prior art keywords
text
information
function
reading
display
Prior art date
Application number
PCT/JP1996/000853
Other languages
English (en)
Japanese (ja)
Inventor
Katsuki Ikuta
Tooru Yokozawa
Tadashi Kuwabara
Original Assignee
Hitachi, Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi, Ltd. filed Critical Hitachi, Ltd.
Priority to PCT/JP1996/000853 priority Critical patent/WO1997037344A1/fr
Publication of WO1997037344A1 publication Critical patent/WO1997037344A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M11/00Telephonic communication systems specially adapted for combination with other electrical systems
    • H04M11/06Simultaneous speech and data transmission, e.g. telegraphic transmission over the same conductors
    • H04M11/066Telephone sets adapted for data transmision
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems

Definitions

  • the present invention relates to an information providing system that provides voice information and character information to a user, and more particularly to display of text information accompanied by voice information.
  • the present invention has been made in view of the above-described problem in the case of performing voice output in displaying text information.
  • the present invention provides voice information and its contents. The purpose of this is to synchronize display information with characters.
  • a terminal receiving distribution of text information
  • Display means for displaying text information on the screen, voice output means for reading out the displayed contents, input means for detecting the position on the screen designated by the user, and text separation before and after the designated position.
  • the apparatus is configured to search for a good position, make the voice output means read out the text information displayed from the position of the delimiter, and provide control means for displaying the read out position on the display means.
  • the invention includes a data conversion means for associating the text information with the voice information, and for deciphering the sentence to identify a head of a segment or a sentence or a paragraph from the text information, and to generate synchronization information.
  • FIG. 1 is a functional block diagram in the first embodiment of the present invention
  • FIG. 2 is a diagram showing an example of the structure of article information as text information in the first embodiment of the present invention.
  • FIG. 3 is a diagram showing an example of a text segment in the first embodiment of the present invention.
  • FIG. 4 is a diagram showing an example of article information as text information in the first embodiment of the present invention.
  • FIG. 5 is a diagram showing a display example of a menu on a terminal in the first embodiment of the present invention.
  • FIG. 6 is a diagram showing a display example of an article on a terminal according to the first embodiment of the present invention.
  • FIG. 7 is a diagram showing the order of processing at the start of audio output in the first embodiment of the present invention.
  • FIG. 8 is a diagram showing the order of processing at the time of switching the reading position during audio output in the first embodiment of the present invention.
  • FIG. 9 is a diagram of an example in which the reading position during voice output is switched in the first embodiment of the present invention.
  • FIG. 10 is a diagram showing a processing order when a character is selected in the first embodiment of the present invention when audio is not output,
  • FIG. 11 is a functional block diagram according to a second embodiment of the present invention
  • FIG. 12 is a functional block diagram according to a third embodiment of the present invention
  • FIG. It is a diagram showing a display example of an article on a terminal in the third embodiment
  • FIG. 14 is a diagram showing a display example of an article on a terminal according to the fourth embodiment of the present invention.
  • FIG. 15 is a diagram showing a display example of an article on a terminal according to the fifth embodiment of the present invention.
  • FIG. 16 is a diagram showing a configuration example of article information as text information in the sixth embodiment of the present invention.
  • FIG. 17 is a diagram showing a display example of an article on a terminal according to the sixth embodiment of the present invention.
  • FIG. 18 is a diagram showing an example of article information as sentence information in the sixth embodiment of the present invention.
  • FIG. 19 is a functional block diagram according to the seventh embodiment of the present invention.
  • FIG. 20 is a configuration diagram of menu information according to the first embodiment of the present invention.
  • FIG. 1 is a functional block diagram of a text information-added sentence information providing system according to a first embodiment of the present invention.
  • This embodiment shows an example of a method in which a user retrieves a digitized newspaper article with audio information to a terminal via a telephone line 3, and an example of audio reproduction control.
  • reference numeral 1 denotes an article information transmission system, which stores various kinds of digitized newspaper article information with audio information and distributes the information to terminals.
  • Reference numeral 101 denotes a data converter, which has a function of outputting article information 1001 by performing grammatical analysis on text information 9001 of an electronic newspaper input by an external system or manually.
  • Article information 1001 is text information 9001, reading information 9002, synchronization information 9003, and identification for identifying these information as one set.
  • the reading information 9002 is obtained by performing a grammatical analysis on the text information 9001 to obtain a phonetic symbol string including information such as reading, accent, pause, intonation, and speed. It provides the information necessary for reading aloud.
  • the synchronizing information 9003 is also subjected to grammatical analysis of the text information 9001 to provide an index indicating that the sentence is well-separated, such as a clause or the beginning of a sentence or paragraph, and a text. This provides information that associates the character string in the information 9001 with the phonetic symbol string in the reading information 9002. Also, identification ID 90
  • each article information for example, as K J 0 1, K J 0 2, and the like.
  • An article information storage unit 102 holds and manages the article information 1001 output from the data conversion unit 101. Furthermore, menu information 1002 for the user to select desired article information 1001 in a menu format is created. Redundant. It also has a management function.
  • the menu information 1002 is composed of text information 9001, an identification ID 9004, and reference information 9005 as shown in FIG.
  • the reference information 9004 is composed of an identification ID 9004 for further menu information or an identification ID 9004 for another article information. The display of each item is performed using the text information 9001 added to the menu information.
  • the identification ID 900 4 is information for identifying the menu information 1002 itself.
  • Reference numeral 103 denotes a receiving unit, which has a function of receiving an identification ID 904 transmitted from a user terminal described later via the telephone line 3 and outputting the received identification ID 904 to a control unit 104 described later.
  • Reference numeral 104 denotes a control unit, which searches the corresponding article information 1001 or menu information 1002 from the article information storage unit 102 based on the identification ID output from the receiving unit 103, and It has a function to output.
  • Reference numeral 105 denotes a transmission unit, which has a function of transmitting the article information 1001 or the menu information 1002 output from the control unit 104 to the user terminal via the telephone line 3.
  • reference numeral 2 denotes an article information receiving system serving as each user's terminal.
  • the article information 100 distributed from the article information transmitting system 1 is displayed, raised, and new article information 100 is displayed. Requests delivery of 1 or menu information 1002.
  • Reference numeral 201 denotes a receiving unit, which receives the article information 1001 or menu information 1002 sent from the article information transmitting system 1 via the telephone line 3 and stores it in the article information storage unit 202. Output.
  • Reference numeral 202 denotes an article information storage unit, which holds article information 1001 or menu information 1002.
  • Reference numeral 203 denotes an article information acquisition unit, which analyzes the article information 1001 and the menu information 1002 obtained from the article information storage unit 202 and obtains the article information 1001 and the menu information 1
  • the above-mentioned various information constituting 02 is stored in each storage unit.
  • the synchronization information 9003 The readout information 9002 is output to the synchronization information storage section 204, the readout information 9002 is output to the readout information storage section 205, and the text information 9001 is output to the text information storage section 206.
  • the reference information 9005 consisting of the identification ID 9004 corresponding to each item is stored in the reference information storage unit 207, and the text information 9001 is textualized. Output to the information storage unit 206.
  • Reference numeral 208 denotes an information display unit, which displays the text information 9001 in the text information storage unit 206 together with an icon for controlling the reading of article text by a liquid crystal display device or the like. Furthermore, in the reverse display processing of the displayed text information 9001, a function of reversely displaying up to the character corresponding to the synchronization information obtained from the reading position storage unit 212, which will be described later, starting from the beginning of the sentence. Prepare. In addition, when the information input unit 209 described later notifies the user of the selected coordinates 201, the type of the icon selected by the user based on the coordinates, or the text information of the selected character. It also has a function of notifying the information input unit 209 of the information 901.
  • Reference numeral 209 denotes an information input unit, which is arranged on the upper surface of the information display unit 208, detects which position on the screen the user has designated, and selects the information displayed on the information display unit 208. It has a function of inquiring the information display unit 208 about which item or text information 9001 has been selected, and outputting it to the operation control unit 210 described later.
  • Reference numeral 210 denotes an operation control unit
  • 211 denotes a speech synthesis unit
  • 212 denotes a reading position storage unit.
  • the reading position storage unit 212 stores synchronization information 9003, and outputs it as reading position synchronization information 2121 when it is referenced from another processing unit.
  • the user specifies the reading position using the information input unit 209
  • the specified character is converted into the selected text information 209 by the inquiry to the information display unit 209, and is input to the information input unit 209. And then output to the operation control unit 210.
  • the operation control unit 210 selects the selected text information 200
  • the synchronization information corresponding to 82 is read from the synchronization information storage unit 204.
  • the synchronization information storage unit 204 searches for the corresponding synchronization information, and outputs it as search synchronization information 2041 to the operation control unit 210.
  • the operation control unit 210 sets the search synchronization information 2041 in the reading position storage unit 212, and sets it as the reading position synchronization information 2121.
  • the speech synthesis section 211 is outputting the speech of the article
  • the synchronization information corresponding to the character being read out by the speech synthesis section 211 is read from the synchronization information storage section 204.
  • the synchronization information storage unit 204 searches for the corresponding synchronization information, and then outputs it as search synchronization information 2041 to the speech synthesis unit 211.
  • the speech synthesizing unit 211 sets the search synchronization information 2041 in the reading position storage unit 212, and sets the same as the reading position synchronization information 2121.
  • the reading position synchronization information 2 1 2 1 in the reading position memory 2 1 2 is read out from the information display section 2 08 when the information value changes due to reading, and the speech synthesis processing is started.
  • the operation control unit 210 controls the speech synthesis unit 211 to start and stop speech synthesis processing in response to an input from the information input unit 209 by a user's operation (operation command 2101). .
  • the voice synthesis unit 211 inquires the reading position storage unit 212 to obtain the reading position synchronization information 2121.
  • read-aloud information starting from the read-aloud position synchronization information 2 1 2 1 is sequentially obtained as read-aloud information 2 0 5 1 from the read-aloud information storage unit 2 05, and a voice synthesizing process is performed. 2 1 3 Output audio.
  • the operation control unit 210 instructs to stop the voice output, the voice synthesis processing is interrupted (operation command 211).
  • the operation control unit 210 recognizes the identification corresponding to the menu item selected by the user from the reference information 905 in the reference information storage unit 207. Get ID 9 0 4 It can be output to the transmission unit 214 described later as the selection identification ID 207 1.
  • Reference numeral 214 denotes a transmission unit, which has a function of transmitting the selection identification ID 207 1 input from the operation control unit 210 to the article information transmission system 1 via the telephone line 3 as a control signal.
  • FIG. 3 and FIG. 4 are diagrams for explaining the correspondence between the text information 9001 and the read-out information 9002 based on the synchronization information 9003 in the article information 1001.
  • Figure 3 shows the original example sentence, the result of separating this example sentence at the beginning of a phrase by grammatical analysis, and an example of a generated phonetic symbol string.
  • FIG. 4 shows article information generated based on the example sentence of FIG.
  • the delimiter flag in the synchronization information in Fig. 4 is 1 only for the block with the first synchronization number. If the text break is at the beginning of the paragraph, only the first block of the first sentence of the paragraph will be 1.
  • the text information 9001 is a text string displayed as shown in Fig. 4 and arranged in blocks from the beginning of the text as a block of characters.
  • the reading information 9002 is the text information 900.
  • a phonetic symbol string obtained by performing a grammatical analysis of 01 is stored in the block, separated by the same character unit as the text information 9001.
  • "policy" is stored in the first block of the text information 9001
  • "policy" is stored in the second block.
  • the first block of reading information 9002 stores "Say”
  • the second block stores "Saku".
  • the synchronization information 9003 is a synchronization number that is a number for each block of the text information 9001 and the corresponding reading information 9002 to refer to each other, and the block is a text delimiter. A flag indicating whether or not it corresponds to the eye is stored in the block. In the example of Fig. 4, the fourth It can be seen from the delimiter flag that the block is at the beginning of the text break. Also, since the synchronization number is "4", the fourth "middle” can be referenced from the corresponding text information, and the fourth "naka '" can be referenced from the corresponding reading information.
  • the voice synthesis unit 211 reads out the first block of the read-out information 9002 and simultaneously reads the first block of the synchronization information in the synchronization information storage unit 204 , Is output to the reading position storage section 2 1 2 and is set as reading position synchronization information 2 1 2 1.
  • the information display section 208 refers to the synchronization number of the reading position synchronization information 2121 in the reading position storage section 212, and reads the characters up to the first block of the text information 9001. Highlight it.
  • the operation control section 210 traces the synchronization information in the synchronization information storage section 204 back from the 11th block, The ninth block having the flag of the break of the near sentence is acquired, output to the reading position storage unit 2 12, and set as the reading position synchronization information 2 1 2 1.
  • the information display section 208 refers to the reading position synchronization information 2121 in the reading position storage section 212 and displays the text information 9001 up to the ninth character in reverse video with reference to the synchronization number. .
  • the voice synthesis unit 211 refers to the synchronization number of the synchronization information and starts the voice synthesis processing from the ninth reading information in the reading information storage unit 205. .
  • Fig. 1 and Fig. 5 a mechanism is described in which the user activates the electronic newspaper providing system with sound and then obtains desired article information by selecting menu items.
  • the operation control unit 210 displays the selection identification ID 207 1 in order to display the initial menu on the information display unit 208.
  • the information is output to the transmission section 214, and the transmission section 214 transmits the selected identification ID 2071 to the article information transmission system 1.
  • the receiving unit 103 receives the selection identification ID 2 07 1 and outputs it to the control unit 104, and the control unit 10
  • the 4 obtains the menu information 1002 of the initial menu from the information storage unit 102.
  • the initial menu is, for example, "1: Top News,
  • each item is an identification ID for obtaining a submenu
  • the control unit 104 transmits the acquired initial menu menu information 1002 to the article information receiving system 2 through the transmission unit 105.
  • the receiving unit 201 receives the menu information 1002 and writes it to the article information storage unit 202.
  • the article information acquisition unit 203 acquires menu information from the article information storage unit 202 and stores text information 9001 for displaying items in the text information storage unit 206.
  • the reference information 900 consisting of the identification ID 900 corresponding to each item is written to the reference information storage unit 207.
  • the information display unit 208 displays initial menu items based on the text information 9001 in the text information storage unit 206.
  • the information input unit 209 notifies the selected coordinates 209 1 to the information display unit 209 and what is selected. And notifies the operation control unit 210 of the obtained selection icon 2081. For example, assuming that “Sports” of item number 4 is selected, the operation control unit 210 queries the reference information storage unit 207 and searches for the reference information 905 to find the selected icon 2 0 8 1 ( Acquisition of the identification ID 9004 corresponding to “spot” is obtained as the selection identification ID 2071, and output to the transmission unit 214. In the following, the user repeats the same procedure, and completes the submenu.
  • the article information transmission system 1 distributes the information 1001 corresponding to the selected identification ID 2 071.
  • the article information receiving system 2 receives the article information 1001
  • the article information acquiring unit 203 sends the synchronization information 9003 in the article information 1001 to the synchronization information storage unit 204.
  • the reading information 9002 is output to the reading information storage unit 205
  • the text information 9001 is output to the text information storage unit 206.
  • the operation control section 210 When the synchronization information 9003 is written to the synchronization information storage section 204, the operation control section 210 writes the synchronization information 9003 corresponding to the beginning of the sentence to the reading position storage section 212, Speaking position synchronization information shall be 2 1 2 1. Further, the information display unit 208 acquires the text information 9001 in the text information storage unit 206, displays the text of the article, and displays an icon for controlling the audio output. I do. At this time, the reading position synchronization information 2 1 2 1 set in the reading information storage unit 205 is the synchronization information corresponding to the beginning of the sentence, and there is no character to be highlighted. Not displayed.
  • FIG. 6 is an example of an electronic newspaper realized by the present system and immediately after an article is displayed on the information display unit 208 by the above-described procedure.
  • the start and stop of reading are performed by selecting an icon displayed in the information display unit 208, and the reading position is specified by direct instruction to the displayed text.
  • Reference numeral 401 denotes the text of the article based on the text information 9001, and icon 420.403 denotes the user's selection for controlling the output of the voice information.
  • the icon 402 has a function of interrupting audio output, and the icon 4003 has a function of starting and restarting audio output.
  • FIG. 7 is an excerpt from the functional block diagram of the article information receiving system 2 in Fig. 1, where S1-S10 in the figure indicate the order of data flow from processing unit to processing unit. Is shown.
  • the information input unit 209 notifies the selected coordinate 209 1 to the information display unit 208 (S1), and the icon 403 is selected. Is received (S2), and this is notified to the operation control unit 210 (S3).
  • the operation control unit 210 outputs an operation command 211 to the voice synthesis unit 211 to start the voice synthesis process (S4). ).
  • the voice synthesizing unit 2 1 1 acquires the reading position synchronization information 2 1 2 1 specifying the reading start position from the reading position storage unit 2 1 2 (S 5), and the reading information storage unit 205 reads the reading position synchronization information.
  • the read-out information 9002 is sequentially obtained as the reference read-out information 205 from the position of the readout information corresponding to the position synchronization information 2 1 2 1 (S6), and the voice output unit 2 1 Voice output is performed by 3 (S7).
  • the synchronization information corresponding to the head of the sentence is written in the reading position synchronization information 2 1 2 1, and thus the voice output is started from the head of the sentence.
  • the voice synthesizing unit 211 acquires search synchronization information 2041 corresponding to the voice being output from the synchronization information storage unit 204 (S8), and stores it in the reading position storage unit 212.
  • the write / readout position synchronization information is set to 2 1 2 1 (S 9).
  • the information display section 208 refers to the reading position synchronization information 2 121 (S 10), and the text information 9 0 0 from the beginning of the sentence to the position corresponding to the reading position synchronization information 2 1 2 1. Change the display of 1 to reverse, for example, like karaoke O o
  • FIG. Fig. 8 is a partial excerpt from the functional block diagram of the article information receiving system 2 in Fig. 1, where S1 to S8 indicate the order of data flow from processing unit to processing unit. .
  • the information input section 209 notifies the selected display section 209 to the information display section 208.
  • the selected text information 2082 corresponding to the selected character is received (S 2), and notified to the operation control unit 210 (S 3).
  • the operation control unit 210 Upon receiving the notification, the operation control unit 210 outputs an operation command 211 to the voice synthesis unit 211 to stop the voice synthesis process (S4).
  • the synchronization information 9003 in the synchronization information storage unit 204 is traced back from the synchronization information corresponding to the selected text information 2082, and the synchronization information having the index of the break of the sentence at the closest position.
  • the information is obtained as search synchronization information 2041 (S5) c.
  • the operation control unit 210 writes the search synchronization information 2041 into the reading position storage unit 212 and reads the reading position.
  • the synchronization information is set to 2 1 2 1 (S 6), and the operation command 2 101 is output to the voice synthesis section 2 1 1 to start the voice output again (S 8).
  • Fig. 9 shows a scene in which the user switches the reading position by voice output by actually selecting the character displayed on the information display section 208 of the terminal during voice output.
  • the characters highlighted in STEP 1 to STEP 3 are the parts that have already been read out, and as the reading progresses, the characters read out are highlighted.
  • STEP 1 is a scene where voice output is being performed.
  • the information input unit 209 inquires the information display unit 208 and-icon 402 is selected. Is notified to the operation control unit 210, and the operation control unit 210 instructs the voice synthesis unit 211 to stop the voice synthesis process (operation command 211). Then, the synchronization information 9003 corresponding to the beginning of the sentence is written into the reading position storage section 212, and is used as the reading position synchronization information 2121.
  • the read-out position synchronization information 2121 is the synchronization information corresponding to the beginning of the sentence, the display of the text information 9001 such as a power mark is canceled.
  • Fig. 10 is a partial excerpt from the function block diagram of the article information receiving system 2 in Fig. 1, and S1 to S6 in the figure indicate the order of data flow from processing unit to processing unit. Location.
  • the information input section 209 notifies the selected display section 209 to the information display section 208. Then, an inquiry is made as to what the selected character is (S2), and after obtaining as selected text information 2082, the operation control unit 210 is notified (S3).
  • the synchronization information 9003 in the synchronization information storage unit 204 is traced back to the synchronization information corresponding to the selected text information 2082, and the index of the break of the text at the closest position is obtained. Then, the synchronization information is retrieved (S4), and is written to the reading position storage unit 212 (S5).
  • the information display section 208 changes the display of text characters by referring to the reading position synchronization information 2 1 2 1 in the reading position storage section 2 1 2. Is displayed in reverse video like karaoke (S6).
  • the audio output is started by selecting the icon 4003, the audio output starts from the character specified by the user.
  • the user can select any character in the text of the article to specify a position where the sentence is well separated as the restart position of the audio output, and furthermore, the position can be easily adjusted. Can be identified.
  • the information display section 208 changes the display of text characters with reference to the reading position synchronization information 2 121, the display from the beginning of the sentence to the character specified by the user is emphasized like karaoke. Or, an underline may be added.
  • a remote control with a button corresponding to the icon function on the user's terminal, it is possible to operate the terminal only by operating the icon even in situations where it is difficult to operate the terminal such as in a crowded train or walking. Functions can be used.
  • the audio output when performing the audio output of an article in an electronic newspaper, by specifying an arbitrary character in the article, the audio output is immediately started from a segment of a sentence that is meaningful as a word. Can be confirmed by the change of the character display.
  • the current reading position can be easily checked based on the change in character display, so the user can specify the position at which to start skip-reading while referring to the current reading position. can do.
  • the reading position is corrected, so that the operability of the user can be improved.
  • FIG. 11 is a diagram for explaining the second embodiment.
  • This embodiment is different from the first embodiment in that the data pre-processing unit 106 is provided in the article information transmission system 3 and the article information reception system 4 is provided with the data conversion unit 215.
  • the preprocessing unit 106 outputs article information 1003 composed of text information 9001 and identification ID 9004.
  • the data converter 2 15 converts the text information 9 0 0 1 in the article information 1 0 3 received from the receiver 2 0 1 into the read information 9 0 0 2 and the synchronization information 9 0 3. It is generated and output to the article information storage unit 202 as article information 1001 together with the text information 9001 and the identification ID 9004.
  • the data transmitted on the telephone line 3 is text information. It was article information 1001, consisting of 9001, reading information 9002, synchronization information 9003, and identification ID9004. According to the present embodiment, the data of the article information 1003 transmitted on the telephone line 3 is the text information 9001 and the identification ID 9004, and the amount of data transmitted on the telephone line 3 is reduced. can do.
  • FIGS. 12 and 13 are diagrams for explaining the third embodiment. This embodiment is different from the first embodiment in that the operation control unit 210 in FIG. 12 can acquire the reading position synchronization information 2 1 2 1 from the reading position storage unit 2 1 2. In addition, as shown in FIG. 13, an icon 404 for temporarily stopping audio output has been added.
  • the operation control section 210 goes back to the synchronization information 9003 in the synchronization information storage section 204 starting from the speech information synchronization information 2121 acquired from the speech information storage section 212, and is the closest position. Is searched for a break in the sentence, and is set as the search synchronization information 2 0 4 1.
  • the search synchronization information 2041 is written into the reading position storage section 212, and is set as the reading position synchronization information 2121. With this setting, the updating of the reading position synchronization information 2 1 2 1 is stopped, so that the reading position indicated by the information display section 208 does not change.
  • the speech synthesizing unit 204 reads the index of the next sentence break from the synchronization information storage unit 204. It is also possible to continue the speech synthesis process until the synchronization information is acquired and then stop. In this case, it is not necessary for the operation control unit 211 to write the synchronization information into the reading position storage unit 212, but after the user selects the icon 404, the text is well separated. Reading will continue to this point.
  • FIG. 14 is a view for explaining the fourth embodiment.
  • This embodiment is different from the first embodiment in that, like the third embodiment, the operation control unit 210 in FIG. 12 stores the reading position synchronization information 2 1 2 1 from the reading position storage unit 2 12 in FIG. This means that icons can be obtained, and icons 405 and 406 for forward and backward audio output have been added as shown in Fig. 14.
  • the operation control unit 210 acquires the reading position synchronization information 2 121 from the reading position storage unit 212, and stores the synchronization information 9003 after the reading position synchronization information 2 121 into the synchronization information storage unit.
  • the search synchronization information 2 0 4 1 is acquired sequentially from 204 and written to the reading position storage unit 2 1 2 to update the reading position synchronization information 2 1 2 1 You.
  • the information display section 208 changes the display of text characters by referring to the reading position synchronization information 2 1 2 1, so that the display from the beginning of the sentence to the character specified by the user is displayed in reverse video like karaoke Let it.
  • the information input unit 209 inquires the information display unit 208 to operate the selection icon 209 which indicates that the selection of the icon 406 has been released. is notified to the control unit 2 1 0 is the operation control unit 2 1 0 written from c further reading position storage unit 2 1 2 to stop writing the synchronization information to the position storage section 2 1 2 reading the last synchronization information
  • the reading position synchronization information 2 1 2 1 is acquired, and the start position is used to search the synchronization information 9 0 3 in the synchronization information storage unit 204 to search for the nearest sentence segment (search synchronization information). 2 0 4 1), and this is written into the reading position storage section 2 12, thereby updating the reading position synchronization information 2 1 2 1.
  • the voice synthesis unit 211 is instructed to start the voice synthesis process (operation command 211). Since the reading position synchronizing information 2 1 2 1 is the synchronizing information of the position where the user has finished the forward feeding, the voice synthesizing unit 2 11 1 starts outputting the sound again from that position.
  • FIG. 15 is a diagram for explaining the fifth embodiment. This embodiment is different from the first embodiment. The difference is that, as in the third embodiment, in FIG. 12 the operation control section 210 can acquire the reading position period information 2 121 from the reading position storage section 2 12 in FIG. As shown in Fig. 15, an icon 407 * 408 for skipping forward and backward audio output was added.
  • the operation control unit 210 sequentially reads only the synchronization information having the index of the segmentation of the sentence from the synchronization information storage unit 204. Obtain and write it to the reading position storage unit 2 1 2. As a result, the reading position synchronization information 2 1 2 1 is updated for each synchronization information having a text segmentation index.
  • the reverse display of the text characters in the information display section 208 is performed in the fourth embodiment. Compared to the sequential example, it can be implemented more quickly for each sentence index. Further, when the selection of the icon 405 is released, the process of acquiring the synchronization information having the index of the segment break of the sentence by acquiring the synchronization information storage unit 212 in the backward direction is unnecessary, and The next process can be started promptly as compared with the sequential forwarding in the fourth embodiment. Even when the user wants to specify the reverse skip processing by selecting the icon 407, the operation control unit 210 is synchronized in the same processing as the forward skip processing when the icon 408 is selected. This is realized by reading the synchronization information 9003 from the information storage unit 204 in the reverse direction.
  • the user can skip forward or backward at each break of the text while visually confirming the point at which the voice output is restarted. This has the effect that it is possible to start from a place with good separation.
  • the first 6 figures and first 7 view and the first 8 figures refer to the sixth embodiment c present embodiment is a diagram illustrating an example is the difference from the first embodiment article information as in the first 6 Figure
  • the control unit 210 can acquire the reading position synchronization information 2 121 from the reading position storage unit 2 12.
  • the reference information 9005 is composed of blocks separated by the same character unit as the text information 9001, as shown in FIG.
  • each block of the reference information corresponding to the text information forming this key word character string includes: The identification ID for all relevant article information is described.
  • the reference information 9005 can be referred to from the corresponding text information 9001 and read-out information 9002 through the synchronization information 9003.
  • the reference information 9 0 5 in the article information 1 0 0 1 is written into the reference information storage 2 0 7 by the article information acquisition unit 2 3, and the identification ID in each block of the reference information is the operation control unit 2 1 0 Thus, it is obtained as the selection identification ID 207 1.
  • the information input unit 209 inquires of the information display unit 208 to notify the operation control unit 210 of a selection icon 209 indicating that the icon 409 has been selected.
  • the operation control unit 210 acquires the reading position synchronization information 2 121 from the reading position storage unit 212 and searches the reference information storage unit 207 based on the reading position synchronization information 211.
  • the corresponding reference information has the identification ID 9004, after acquiring this, it outputs it to the transmission section 214 as the selection identification ID2071.
  • the transmitting unit 2 14 sends the selection identification ID 2 07 1 to the article information transmitting system 1, so that the article information transmitting system 1 performs the procedure described in the first embodiment. Similarly, the related article information is distributed. If a voice is being output, the operation control unit 210 instructs the voice synthesis unit 211 to stop the voice synthesis process. If the retrieved reference information 9005 does not have an identification ID, the operation control unit 210 ignores the selection of the icon 409.
  • the user can receive the related article information by specifying the keyword for the keyword in which the article information that can be referred to exists.
  • the operation control unit 210 reads out the reading position information storage unit. Starting from the reading position synchronization information 2 1 2 1 obtained from 2 1 2, the synchronization information 9 0 0 3 in the synchronization information storage section 204 is traced back from the synchronization information storage section 204, and the corresponding reference information within an appropriate range is the identification ID 90. If it has 04, the identification ID 900 4 can be output to the transmission section 214 as the selected identification ID 207 1.
  • the user does not necessarily have to select the icon 409 at the moment when the user is reading the desired keyword, and selects the icon 409 with some time margin. Can be performed.
  • FIG. 19 is a view for explaining the seventh embodiment.
  • This embodiment is different from the first embodiment in FIG. 19 in that the operation control unit 210 can acquire the reading position synchronization information 2 121 from the reading position storage unit 2 12, Another point is that a repetition position storage unit 2 16 is provided.
  • the repeated position storage unit 2 16 allows the user to start character selection and the character It is possible to repeatedly execute the text-to-speech reading in the range specified by the character from which the character has been canceled. However, if the character from which the character selection is started and the character from which the selection is deselected are the same character and the range is not specified, there is no need to repeat, and the reading is not repeated.
  • the information input unit 209 queries the information display unit 208 to select the character corresponding to the character from which the user started selecting characters and the character corresponding to the character from which the user deselected the characters. After acquiring the text information 2082, it outputs it to the operation control unit 210.
  • the operation control unit 210 acquires synchronization information corresponding to the selected text information 2082 from the synchronization information storage unit 204 as search synchronization information 2041, respectively, and repeats the position storage unit 21 Write to 6. At this time, the synchronization information indicating the start position of the repetition is written as start synchronization information 216 1, and the synchronization information indicating the end position of the repetition is written as end synchronization information 216 2.
  • the start synchronization information 2 161 is used by the operation control unit 210 to convert the synchronization information 900 3 in the synchronization information storage unit 204 into text information of the character that the user has started to select.
  • the end synchronization information 2 162 corresponds to the synchronization information 900 3 in the synchronization information storage unit 204 by the operation control unit 210 corresponding to the text information of the character that the user has deselected.
  • the synchronization information is searched in the direction toward the end of the sentence starting from the synchronization information, and is the synchronization information that has the index of the delimitation of the text at the closest position.
  • the two pieces of synchronization information in the repetition position storage unit 2 16 are cleared by the operation control unit 210, and the repeated reading process is stopped.
  • the operation control unit 210 specifies Clear The reading process is not performed repeatedly.
  • the information input unit 209 queries the information display unit 208 to notify the operation control unit 210 that the icon 403 has been selected, and the operation control unit 210 notifies the speech synthesis unit 211. Command the start of voice synthesis processing. If voice output is in progress, stop the voice synthesis process once and then instruct it to restart.
  • the operation control unit 210 starts sound output, if there are two pieces of synchronization information in the repetition position storage unit 216, it writes the start synchronization information 216 in the reading position storage unit 212.
  • the speech synthesis unit 211 is instructed to start speech synthesis processing. Further, during voice output, the read-out position synchronization information 2 1 2 1 in the read-out position storage unit 2 12 is referred to, and if the end synchronization information 2 16 2 in the repetition position storage unit 2 16 is the same, the sound is output.
  • the synthesizing unit 211 stops the speech synthesizing process, writes the start synchronization information 216 1 again in the reading position storage unit 212, and then starts the speech synthesizing process.
  • the present invention when sound output of document information (for example, electronic newspaper), by designating any character in the text directly or by forward or backward, the text closest to the desired position can be specified. It is possible to start sound output from the break, and to check the location by changing the display of characters.
  • document information for example, electronic newspaper
  • the position of the current reading Since the position can be easily checked, the position E at which to start skipping again can be specified directly or by forward or backward while referring to the current reading position. It is possible to quickly switch the reading position to voice output from the break point of the sentence closest to the specified position.
  • the embodiment of the present invention is not limited to electronic newspapers, but obtains document information and data from a device that stores various text information, a database, or the like, and outputs a voice while displaying the document information and data. Even in such a case, the above-described various effects of the present invention can be enjoyed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Information Transfer Between Computers (AREA)
  • Digital Computer Display Output (AREA)

Abstract

Dans un terminal pourvu d'une fonction d'affichage et de lecture d'un texte, lorsque la position de début de lecture est désignée, les données de synthèse vocales sont synchronisées avec les données d'affichage, de manière à convertir la position de début de reproduction du texte en pause dans le texte, au commencement d'un membre de phrase, d'une phrase ou d'un paragraphe, par exemple, et ce pour faciliter l'audition. Un système de réception de données de texte (2) d'un terminal, chargé de la réception d'un texte distribué, comporte un dispositif d'affichage (208) affichant le texte sur un écran, un dispositif de synthèse vocale destiné à la lecture à voix haute du texte affiché, un dispositif d'entrée (209) servant à la détection d'une position sur l'écran indiquée par l'utilisateur et un dispositif de commande (210) ayant pour fonction de rechercher à quel endroit le texte est doté de la meilleure ponctuation, avant et après la position indiquée, ce qui amène le dispositif de synthèse vocale à débuter la lecture à l'emplacement défini par cette position et le dispositif d'affichage à afficher la position du mot en cours de lecture.
PCT/JP1996/000853 1996-03-29 1996-03-29 Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal WO1997037344A1 (fr)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP1996/000853 WO1997037344A1 (fr) 1996-03-29 1996-03-29 Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP1996/000853 WO1997037344A1 (fr) 1996-03-29 1996-03-29 Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal

Publications (1)

Publication Number Publication Date
WO1997037344A1 true WO1997037344A1 (fr) 1997-10-09

Family

ID=14153131

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP1996/000853 WO1997037344A1 (fr) 1996-03-29 1996-03-29 Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal

Country Status (1)

Country Link
WO (1) WO1997037344A1 (fr)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000267687A (ja) * 1999-03-19 2000-09-29 Mitsubishi Electric Corp 音声応答装置
GB2357943A (en) * 1999-12-30 2001-07-04 Nokia Mobile Phones Ltd User interface for text to speech conversion
WO2002027710A1 (fr) * 2000-09-27 2002-04-04 International Business Machines Corporation Procede et systeme destines a synchroniser une presentation audio et visuelle dans un restituteur de contenu multimode
GB2369219A (en) * 2000-07-31 2002-05-22 John Christian Doughty Nissen System for synchronous display of text and audio data
JP2003108171A (ja) * 2001-09-27 2003-04-11 Clarion Co Ltd 文書読み上げ装置
JP2003521725A (ja) * 1999-06-25 2003-07-15 ディスカバリー・コミニュケーションズ・インコーポレーテッド 音声エミュレーション機能を備えた電子書籍
US8285547B2 (en) 2005-04-18 2012-10-09 Ricoh Company, Ltd. Audio font output device, font database, and language input front end processor
US9053640B1 (en) 1993-12-02 2015-06-09 Adrea, LLC Interactive electronic book

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5612700A (en) * 1979-07-12 1981-02-07 Tokyo Shibaura Electric Co Audio response system
JPS6238716B2 (fr) * 1982-01-29 1987-08-19 Intaanashonaru Bijinesu Mashiinzu Corp
JPS62290966A (ja) * 1986-06-10 1987-12-17 Nec Corp 文章編集装置
JPS63163925A (ja) * 1986-12-26 1988-07-07 Fujitsu Ltd 文章読み上げ装置
JPH02137899A (ja) * 1988-11-18 1990-05-28 Sanyo Electric Co Ltd 規則合成装置の文字表示方法
JPH07152532A (ja) * 1993-11-29 1995-06-16 Toshiba Corp 文章読み上げ装置
JPH07175909A (ja) * 1993-12-20 1995-07-14 Canon Inc データ処理装置

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5612700A (en) * 1979-07-12 1981-02-07 Tokyo Shibaura Electric Co Audio response system
JPS6238716B2 (fr) * 1982-01-29 1987-08-19 Intaanashonaru Bijinesu Mashiinzu Corp
JPS62290966A (ja) * 1986-06-10 1987-12-17 Nec Corp 文章編集装置
JPS63163925A (ja) * 1986-12-26 1988-07-07 Fujitsu Ltd 文章読み上げ装置
JPH02137899A (ja) * 1988-11-18 1990-05-28 Sanyo Electric Co Ltd 規則合成装置の文字表示方法
JPH07152532A (ja) * 1993-11-29 1995-06-16 Toshiba Corp 文章読み上げ装置
JPH07175909A (ja) * 1993-12-20 1995-07-14 Canon Inc データ処理装置

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9053640B1 (en) 1993-12-02 2015-06-09 Adrea, LLC Interactive electronic book
JP2000267687A (ja) * 1999-03-19 2000-09-29 Mitsubishi Electric Corp 音声応答装置
JP2003521725A (ja) * 1999-06-25 2003-07-15 ディスカバリー・コミニュケーションズ・インコーポレーテッド 音声エミュレーション機能を備えた電子書籍
US9099097B2 (en) 1999-06-25 2015-08-04 Adrea, LLC Electronic book with voice emulation features
GB2357943B (en) * 1999-12-30 2004-12-08 Nokia Mobile Phones Ltd User interface for text to speech conversion
US6708152B2 (en) 1999-12-30 2004-03-16 Nokia Mobile Phones Limited User interface for text to speech conversion
EP1113416A3 (fr) * 1999-12-30 2001-08-16 Nokia Mobile Phones Ltd. Interface utilisateur pour la conversion texte-parole
EP1113416A2 (fr) * 1999-12-30 2001-07-04 Nokia Mobile Phones Ltd. Interface utilisateur pour la conversion texte-parole
GB2357943A (en) * 1999-12-30 2001-07-04 Nokia Mobile Phones Ltd User interface for text to speech conversion
GB2369219A (en) * 2000-07-31 2002-05-22 John Christian Doughty Nissen System for synchronous display of text and audio data
WO2002027710A1 (fr) * 2000-09-27 2002-04-04 International Business Machines Corporation Procede et systeme destines a synchroniser une presentation audio et visuelle dans un restituteur de contenu multimode
JP2003108171A (ja) * 2001-09-27 2003-04-11 Clarion Co Ltd 文書読み上げ装置
US8285547B2 (en) 2005-04-18 2012-10-09 Ricoh Company, Ltd. Audio font output device, font database, and language input front end processor

Similar Documents

Publication Publication Date Title
US7159174B2 (en) Data preparation for media browsing
US8457611B2 (en) Audio file edit method and apparatus for mobile terminal
US20080300012A1 (en) Mobile phone and method for executing functions thereof
WO1997037344A1 (fr) Terminal possedant une fonction de synthese vocale et systeme dispensant des donnees caracteres utilisant ce terminal
JP2741833B2 (ja) マルチメデイア提示内で音声的サーチ・パターンを使用するためのシステム及びその方法
EP1073036A2 (fr) Analyse de documents téléchargés pour un navigateur équipé d'un dispositif de synthèse de la parole
JP2007219218A (ja) 語学学習用電子機器および訳文再生方法
JP2007199315A (ja) コンテンツ提供装置
JP6273560B1 (ja) 拡声装置およびその制御方法
JP2008171173A (ja) 文書検索方法および装置、プログラム
JP2000278624A (ja) 情報表示装置
JP4080965B2 (ja) 情報提示装置及び情報提示方法
JP2022051500A (ja) 関連情報提供方法及びシステム
JP3088399B2 (ja) 文書データ音声合成方式
JP6650636B1 (ja) 翻訳装置、その制御方法およびプログラム
JP4895759B2 (ja) 音声メッセージ出力装置
JP2007189534A (ja) 移動通信端末装置
JP2020074124A (ja) 翻訳装置、その制御方法およびプログラム
JP2019179081A (ja) 会議支援装置、会議支援制御方法およびプログラム
JP6628157B2 (ja) 翻訳装置、その制御方法およびプログラム
JP7335115B2 (ja) カラオケ用入力装置
JP6269771B1 (ja) 翻訳装置、拡声装置、およびその制御方法
JP4520375B2 (ja) 音声操作支援装置、電子機器、画像形成装置およびプログラム
JP2006195900A (ja) マルチメディアコンテンツ生成装置及び方法
KR100705901B1 (ko) 이동통신 단말기 및 텔레비젼을 위한 문자 및 음성정보서비스 시스템

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CN JP KR SG US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE

DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase