WO2000054249A1 - Dispositif de reproduction de donnees, procede de reproduction de donnees et terminal d'informations - Google Patents
Dispositif de reproduction de donnees, procede de reproduction de donnees et terminal d'informations Download PDFInfo
- Publication number
- WO2000054249A1 WO2000054249A1 PCT/JP2000/000602 JP0000602W WO0054249A1 WO 2000054249 A1 WO2000054249 A1 WO 2000054249A1 JP 0000602 W JP0000602 W JP 0000602W WO 0054249 A1 WO0054249 A1 WO 0054249A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- unit
- event
- information
- time
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 71
- 239000000872 buffer Substances 0.000 claims description 147
- 238000012545 processing Methods 0.000 claims description 78
- 238000009826 distribution Methods 0.000 claims description 40
- 238000003860 storage Methods 0.000 claims description 29
- 230000000694 effects Effects 0.000 claims description 20
- 230000006870 function Effects 0.000 claims description 18
- 238000012546 transfer Methods 0.000 claims description 13
- 230000003252 repetitive effect Effects 0.000 claims description 6
- 230000000630 rising effect Effects 0.000 claims description 2
- 238000005065 mining Methods 0.000 claims 1
- 230000001360 synchronised effect Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 19
- 238000004891 communication Methods 0.000 description 10
- 230000000007 visual effect Effects 0.000 description 7
- 241001342895 Chorus Species 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 5
- HAORKNGNJCEJBX-UHFFFAOYSA-N cyprodinil Chemical compound N=1C(C)=CC(C2CC2)=NC=1NC1=CC=CC=C1 HAORKNGNJCEJBX-UHFFFAOYSA-N 0.000 description 5
- 238000004519 manufacturing process Methods 0.000 description 4
- 230000001755 vocal effect Effects 0.000 description 4
- 238000004590 computer program Methods 0.000 description 2
- 239000012141 concentrate Substances 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000033764 rhythmic process Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- ZYXYTGQFPZEUFX-UHFFFAOYSA-N benzpyrimoxan Chemical compound O1C(OCCC1)C=1C(=NC=NC=1)OCC1=CC=C(C=C1)C(F)(F)F ZYXYTGQFPZEUFX-UHFFFAOYSA-N 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10K—SOUND-PRODUCING DEVICES; METHODS OR DEVICES FOR PROTECTING AGAINST, OR FOR DAMPING, NOISE OR OTHER ACOUSTIC WAVES IN GENERAL; ACOUSTICS NOT OTHERWISE PROVIDED FOR
- G10K15/00—Acoustics not otherwise provided for
- G10K15/04—Sound-producing devices
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0033—Recording/reproducing or transmission of music for electrophonic musical instruments
- G10H1/0041—Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
- G10H1/0058—Transmission between separate instruments or between individual components of a musical system
- G10H1/0066—Transmission between separate instruments or between individual components of a musical system using a MIDI interface
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/36—Accompaniment arrangements
- G10H1/361—Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
- G10H1/368—Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems displaying animated or moving pictures synchronized with the music or audio part
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/021—Background music, e.g. for video sequences or elevator music
- G10H2210/026—Background music, e.g. for video sequences or elevator music for games, e.g. videogames
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/011—Files or data streams containing coded musical information, e.g. for transmission
- G10H2240/031—File merging MIDI, i.e. merging or mixing a MIDI-like file or stream with a non-MIDI file or stream, e.g. audio or video
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/171—Transmission of musical instrument data, control or status information; Transmission, remote access or control of music data for electrophonic musical instruments
- G10H2240/201—Physical layer or hardware aspects of transmission to or from an electrophonic musical instrument, e.g. voltage levels, bit streams, code words or symbols over a physical link connecting network nodes or instruments
- G10H2240/241—Telephone transmission, i.e. using twisted pair telephone lines or any type of telephone network
- G10H2240/251—Mobile telephone transmission, i.e. transmitting, accessing or controlling music data wirelessly via a wireless or mobile telephone receiver, analogue or digital, e.g. DECT, GSM, UMTS
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/325—Synchronizing two or more audio tracks or files according to musical features or musical timings
Definitions
- Another object of the present invention is to provide a data reproducing apparatus capable of obtaining a music performance full of a sense of reality.
- Another object of the present invention is to provide a data reproducing method capable of further reducing the amount of reproduced data.
- Another object of the present invention is to achieve stable operation irrespective of fluctuations in transmission line capacity. Another object of the present invention is to provide a data reproducing apparatus capable of reproducing data by using the data reproducing apparatus.
- FIG. 1 is a block diagram showing an example of a data reproducing apparatus according to the present invention.
- FIG. 2 is a diagram showing the format of received data in the SMF format.
- Fig. 5 shows an example of data format for voice, text, and images.
- FIG. 7 shows another example of the format of data relating to voice, text, and images.
- FIG. 9 is a flowchart showing an example of a data reproducing method according to the present invention.
- FIG. 12 is a flowchart of the repetitive reproduction process.
- FIG. 14 is a diagram showing an example of insertion of divided data.
- FIG. 19 is a block diagram showing another example of the data reproducing apparatus of the present invention.
- FIG. 20 is a flowchart showing another example of the data reproducing method of the present invention.
- FIG. 21 is a diagram for explaining the principle of time calculation in data distribution.
- Fig. 26 is a time chart of data reception.
- Figure 27 is a time chart illustrating the caching of data.
- FIG. 28 is a block diagram showing another example of the data reproducing apparatus of the present invention.
- FIG. 29 is a time chart showing the operation of the apparatus shown in FIG.
- FIG. 30 is a block diagram showing another example of the data reproducing apparatus of the present invention.
- FIG. 31 is a time chart showing the operation of the apparatus shown in FIG.
- FIG. 32 is a flowchart in the case of performing a charging discount process using the data reproducing apparatus of the present invention.
- FIG. 34 is an example of a tag added to character data.
- FIG. 35 is a flowchart in the case of providing a service with an expiration date using the data reproducing apparatus of the present invention.
- FIG. 37 is a diagram showing a mobile phone equipped with the data reproducing apparatus of the present invention.
- FIG. 39 is a diagram showing a system using a mobile phone.
- FIG. 1 shows an example of a data reproducing apparatus according to the present invention.
- l a and 1 b are files in which data is recorded
- l a is, for example, a file on a server on the Internet
- 1 b is, for example, a file on a hard disk in the device.
- Reference numeral 2 denotes a CPU that controls the entire data reproducing apparatus, and includes a data receiving unit 3 and a data distribution unit 4. Although the CPU 2 includes other blocks having various functions, they are not shown in the figure because they are not directly related to the present invention.
- the data receiving unit 3 accesses the files 1a and 1b to receive the data stored therein. File la data is received via wire or wirelessly Is done. These received data are temporarily stored in the buffer 3a.
- the data distribution unit 4 distributes the data received by the data reception unit 3 to the data reproduction unit 6 by type.
- the data reproduction unit 6 includes a MIDI reproduction unit 11 for reproducing data relating to MIDI, an audio reproduction unit 12 for reproducing data relating to audio, a character reproduction unit 13 for reproducing data relating to characters, and a data reproduction unit relating to images. And an image reproducing unit 14 for reproducing data.
- the MIDI playback unit 11 has a sound source ROM 11a that stores sound source data of various musical instruments used for music to be reproduced. This sound source ROM 11a can be replaced with a RAM so that the built-in data can be exchanged.
- the image reproducing unit 14 has a function of reproducing a still image and a moving image.
- Reference numeral 19 denotes a speaker as a sound generator for outputting a sound based on the data in the output buffer 17
- reference numeral 20 denotes a display for displaying visible information such as characters and pictures based on the data in the output buffer 18. It is.
- Fig. 3 (a) is the same as Fig. 2 (a).
- the MIDI event is composed of status information and data as shown in FIGS. 3 (b) and (c).
- Fig. 3 (b) shows the event of the sounding start command.
- the status information records the type of instrument, the data 1 contains the scale, and the data 2 contains the dynamics of the sound.
- FIG. 3 (c) shows the event of the sound generation stop command.
- the status information records the type of instrument, the data 3 stores the scale, and the data 4 stores the strength of the sound.
- the MIDI event is an event in which performance information is stored, and a single event constitutes a command such as, for example, "produce the sound of the piano with the piano sound at this strength.” You.
- FIG. 4 shows an example of a simplified MDI format in which the data amount is reduced by simplifying the format of FIG.
- the sounding start instruction and the sounding stop instruction are configured separately, but in Fig. 4, sounding and stopping are integrated into one event by adding the sounding time to the data. ing. Also, the data of the sound intensity is omitted, and the scale data is included in the status information.
- the format in FIG. 4 is not a standard format such as SMF, but the data handled in the present invention includes such a format other than SMF.
- Fig. 5 (a) is the same as Fig. 2 (b).
- the META event can control the transfer of data and control the start and stop of playback, etc., and can extend the force format, which is an event. — —
- Seed data can be embedded.
- 5 (b) to (e) show examples of the format of the extended META event, (b) shows the format in which audio data is embedded, and (c) shows the format in which character data is embedded.
- (D) shows a format in which image data is embedded, and (e) shows a format in which character data and image data are embedded.
- Images include still images such as pictures and photographs, as well as moving images.
- the first F Fh is a header indicating that this event is a META event.
- the following 3Oh, 31h, ... are identifiers indicating that the format of the META event is an extended format.
- 1 en represents the data length of the META event
- typ represents the format of the data to be transferred
- id represents the data number.
- e v e n t indicates the content of the event to be executed, and is represented by, for example, an instruction such as “start transfer of audio data” or “end transfer of image data”. The end position of these data can be known from the value of 1 en representing the data length.
- the Sys.Ex event is called a system exclusive event, and is an event related to configuration information, for example, when setting up a system suitable for orchestra. — —
- the data in the SMF format is configured as described above, and a number of these data are combined to form a series of data strings.
- FIG. 8 shows an example of such a data string.
- M is the data related to MIDI and has the format shown in Fig. 3.
- A is the audio data, which has the format shown in Fig. 5 (b).
- T is data relating to characters, and has the format shown in Fig. 5 (c).
- P is data related to the image, and has the format shown in Fig. 5 (d).
- the arrangement order of each data is not limited to that shown in FIG. 8, and various patterns may exist.
- audio, text, and image data are recorded in the META event, but they can also be recorded in the Sys. Ex event.
- Each piece of data M, A, P, and P is configured as a packet, and these are chained to form a series of data strings. This data sequence is received by the data receiving unit 3 in FIG. 1 and stored in the buffer 3a.
- the received data is distributed by the data distribution unit based on each delta time ⁇ T, and the data reproduction unit 6 executes an event to reproduce the data.
- the timing at which the event is executed is determined by the delta time ⁇ T. That is, the event is executed when the relationship between the elapsed time ⁇ from the event executed immediately before and the delta time ⁇ T of the event executed this time is ⁇ t ⁇ . That is, when an event is executed, the elapsed time since the start of the event is counted, and this elapsed time is the delta time of the next event.
- the delta time is information indicating how much time has elapsed since the previous event to execute the current event, and does not indicate the absolute time, but the delta time. It is possible to calculate the time from the start of playback by accumulating the time.
- the data distribution unit 4 of the CPU 2 sequentially reads the received data from the buffer 3a according to a program stored in a ROM (not shown). If the read data is the data M related to MDI (FIG. 3), the event information is given to the MIDI reproducing unit 11. Assuming that the content of the event is, for example, "produce the sound of M with the sound of a piano", the MIDI playback unit 11 decodes this instruction and sends it from the sound source ROM 11a.
- the piano sound is read, a synthesizer sound is generated by the software synthesizer, and the sound is started at the pitch of mi. From this point on, CPU 2 counts the elapsed time and this elapsed time may be equal to or exceed the delta time attached to the next event “Stop sounding my”. Then, this command is given to the MIDI playback unit 11, and the MIDI playback unit 11 decodes this command and stops generating the mi sound. In this way, the sound of mi is played as a piano sound only from the start of sound production to the stop of sound production.
- CPU 2 counts the elapsed time from the stop of the sound of mi, and the elapsed time is attached to the next event, for example, "produce the sound of la with the sound of the piano". If this time is equal to or exceeds the delta 'time, the command is given to the — —
- the MID I playback unit 11 decodes this instruction, reads the piano sound from the sound source ROM 11a, generates a synthesizer sound, and starts sounding at the pitch of LA. Then, from this time, the CPU 2 counts the elapsed time, and the elapsed time may be equal to or equal to the delta time attached to the next event “stop sounding la”. If it exceeds, this command is given to the MIDI playback unit 11, and the MIDI playback unit 11 decodes this command and stops the sound of the sound. In this way, the sound of La is played as a piano sound for the time from the start of sound production to the stop of sound production. By repeating such an operation, the 101 reproducing section 1t reproduces the sound by MDI.
- the event information of the read data is distributed to the image playback unit 14 according to the delta time.
- the image playback unit 14 decodes the contents of the event. To execute the event and play back the image.
- the audio reproducing unit 12 is transmitted from the data distribution unit 4 by, for example, When receiving the event "pronounce sound B", the sound reproducing unit 12 decodes and reproduces the data of the sound B added to the event. From this point on, CPU 2 counts the elapsed time, and if this elapsed time is equal to or exceeds the delta time associated with the next event, for example, "show character C", the statement The character reproducing unit 13 decodes and reproduces data of the character C added to the event. Next, CPU 2 counts the elapsed time since the reproduction of character C, and this elapsed time may be equal to the delta time attached to the next event, for example, “Display picture D”. If the number exceeds that, the image reproducing unit 14 decodes and reproduces the picture D data added to the event. This point is basically the same as the principle of reproducing the MDI data described above.
- the playback operation by the MIDI playback unit 11 and the playback operation by the playback units 12 to 14 other than MIDI are described separately, but in reality, as shown in FIG.
- data M having MIDI events and data A, T, and ⁇ having non-MIDI events are mixed in time series and input.
- MIDI (1 ⁇ 1) ⁇ picture (?) ⁇ character (cho) ⁇ ⁇ 11 0 1 (1 ⁇ ) ⁇ voice ()-trajectory (P)- Entered.
- the data distribution unit 4 distributes these data to the respective reproduction units 11 to 14 according to the type according to the delta time, and each of the reproduction units 11 to 14 performs the reproduction processing of the corresponding data. .
- the data played back by the MIDI playback unit 11 and the data played back by the audio playback unit 12 were mixed in the mixer 15 and subjected to echo processing and the like in the sound effect unit 15a. Thereafter, the data is temporarily stored in the output buffer 17 and output from the speaker 19 as a sound.
- the data played back by the character playback unit 13 and the data played back by the image playback unit 14 The mixed data is mixed by the mixer 16, subjected to special video processing and the like by the visual effect section 15 a, temporarily stored in the output buffer 18, and displayed on the display 20 with visual information. Is displayed. Then, when the data distribution unit 4 receives the META event of the reproduction stop shown in FIG. 6 (b), the data reproduction ends.
- the data reproducing apparatus shown in FIG. 1 can sort and reproduce each type of data from a data string in which MID I, voice, characters, and images are mixed.
- the delta time is referenced in the same manner as in the reproduction of MDI, and data is reproduced at a timing according to the delta time. Therefore, it is possible to easily synchronize between different types of data such as sound images simply by describing the delta time, and to incorporate a clock in a program that processes each data as in the past. Since there is no need, there is no problem of synchronization loss due to mismatch between clocks.
- FIG. 9 is a flowchart showing a data reproducing method in the reproducing apparatus of FIG. 1, and shows a procedure executed by the CPU 2. The operation will be described below by taking as an example the case where the playback device is a playback device for communication karaoke. In the following, the steps of the flowchart are abbreviated as “S”.
- the data receiving unit 3 receives data from the server file 1a on the network via the communication line (S101), the received data is stored in the buffer 3a (S102). ).
- the data distribution unit 4 reads the data of the buffer 3a and counts the elapsed time since the immediately preceding event was executed (S103). Then, it is determined whether the elapsed time coincides with the time indicated by the delta time (or the force exceeded the time (S104), and if the force does not exceed the delta time (S104NO). Return to S103 and continue counting elapsed time. If the elapsed time matches or exceeds the delta time (S104 YES), processing of the data is started.
- the type of received data is first determined. That is, it is determined whether or not the received data is MIDI data M (S105). If the received data is MIDI data (S105 YES), this is distributed to the MIDI playback unit 11 and the MIDI playback is performed. A synthesizer sound is generated in the section 11 (S111). Since the detailed principle has already been described, the description is omitted here. The karaoke accompaniment is output from speaker 19 by the sound reproduction by the synthesizer.
- the received data is not MIDI data M (S105 NO)
- the received data is not voice data A (S106 NO)
- this is distributed to the character reproducing unit 13, and the character is processed by the character reproducing unit 13 to reproduce the character (S113).
- the title of the karaoke song and the lyrics are displayed on the display unit 20.
- the received data is not character data T (S107NO)
- the still image and the moving image are processed by the unit 14 and the image is reproduced (S114).
- the background image such as animation / movie is displayed on the display unit 20.
- the received data is not image data (S108N)
- the data is, for example, data relating to setting and control, and predetermined processing is performed according to the content (S109).
- it is determined whether or not to stop the reproduction that is, whether or not the META event of FIG. 6 (b) has been received (S11 •). If the playback is not to be stopped (S110N ⁇ ), the process returns to S101 and waits for reception of the next data. If the playback is to be stopped (S110YES), the operation ends.
- the data reproducing apparatus shown in FIG. 1 has a sound reproducing section composed of the MIDI reproducing section 11 and the audio reproducing section 12, and a visual information reproducing section composed of the character reproducing section 13 and the image reproducing section 14.
- the unit is suitable for communication karaoke.
- the audio playback unit 12 is not always necessary and may be omitted.
- the audio playback unit 12 is provided, and the parts of the musical instrument are played back by the MIDI playback unit 11. By playing back the vocal portion with the audio playback portion 12, the vocal portion can be played back with the original sound, and an extremely realistic performance can be realized.
- the SMF format data received by the data receiving unit 3 is stored in the file 1a of the server on the network as described above, and the data of the new song is periodically uploaded to the file 1a. And the contents of Huainole 1a are updated.
- FIG. 10 is a flowchart showing a reproducing method when the data reproducing apparatus of FIG. 1 is used for broadcasting a commercial (CM) on a television, and shows a procedure executed by the CPU 2.
- S121 to S124 correspond to S101 to L104 in FIG. 9, respectively.
- This announcement data is composed of audio data. If the data is analysis data (S126 YES), the data is distributed to the audio reproduction unit 12 to perform audio processing and reproduce the audio (S133). The speaker 19 outputs an announcer's commentary and the like when the sound is reproduced. If the received data is not “announcement data” (S 1 2 6 NO), then it is determined whether or not the data is character data representing a product name or the like (S 1 2 7). YES), this is distributed to the character reproducing unit 13 and the character is reproduced by the character reproducing unit 13 and displayed on the display 20 (S134).
- the received data is not character data (S127 NO)
- This is distributed to the image playback unit 14, the still image is processed by the image playback unit 14, and the picture is played back, and displayed on the display 20 (S 1
- the received data is not picture data (S 1 28 NO)
- the moving image is processed, and the moving image is reproduced and displayed on the display 20 (S136).
- S130 and S131 correspond to S109 and S110 in FIG. 9, respectively, and their operations are the same as those in FIG. 9, and therefore description thereof is omitted.
- the same data when reproducing voice, character and image data embedded in the data of the SMF format, the same data may be reproduced several times. For example, a karaoke back chorus may be repeated three times, or the same character may be displayed twice at the beginning and end of a commercial.
- the method shown in Fig. 11 can be considered as a solution.
- the transmitting side receives the data as shown in (b).
- the bucket with the embedded R is sent only once at the beginning.
- the receiving side data reproducing device
- the transmitting side does not send the data R, but sends only the message "Denoleter ⁇ Play the data R after the time indicated by the time has elapsed".
- the receiving side reads the data R from the memory at a predetermined time according to the delta time according to the message and reproduces it. By performing this operation three times, t l, t 2, and t 3, the amount of data to be transmitted can be reduced to one third.
- the transmission data is temporarily stored in the memory and then played back is described as an example.However, the method of Fig. Smell of reception Can be applied.
- the transmitted data R is stored in the memory at the first reproduction time t1.
- FIG. 12 is a flowchart showing the above-described repetitive reproduction processing.
- FIG. 9 shows a detailed procedure in S112, S113 or S114 of FIG. 9, or FIG. This is the detailed procedure in S133, S134, S135 or S136 of FIG.
- S141 data R to be repeatedly reproduced
- S141NO repeated data
- S144 YES set the number of times of reproduction to the counter N inside the CPU (S144), read data R from memory (S144), and output this (S144).
- the counter N is decremented by one and updated to N-1 (S145).
- S146 it is determined whether or not the counter N has become 0 (S146). Shift to S131. When the counter N reaches 0 (S146YES), the recorded data R is erased and the memory is released (S147).
- FIG. 13 is a diagram showing the principle of data advance in the stream system.
- (a) when data such as voice and image is transmitted following MIDI data, the data amount is small in the MIDI part, but suddenly becomes large in the data X part such as voice and image. Data volume increases. (The reason that the amount of MIDI data is small is that MIDI is not a sound itself, but a command for controlling sound generation, and is composed of binary data.) If data X is sent as it is, a large-capacity communication line is required.
- data X is appropriately divided, and the divided data is given IDs XI, X2, and X3.
- IDs XI, X2, and X3 By inserting the divided data into the preceding MIDI data question and sending it forward, the amount of data to be transmitted is leveled and the line capacity can be reduced.
- the data X may be divided over the entire section.
- each divided data of data X and data Y is provided with an ID for each group of X and Y such as XI, X2,... And Yl, Y2,.
- FIG. 14 (b) shows an example in which the divided data is inserted between the preceding MDI data.
- the received divided data is separated from the MIDI data and stored in the memory in chronological order from the first data in FIG. 14 (b).
- the contents of this memory are shown in Figure 15.
- the starting address of the subsequent divided data linked to the divided data is recorded for each of the X and Y groups. For example, at the end of data X1, the start address of data X2 is recorded, and at the end of data X2, the start address of data X3 is recorded. At the end of data Y1, the start address of data Y2 is recorded, and at the end of data Y2, the start address of data Y3 is recorded.
- FIG. 16 is a flowchart showing the operation of extracting the divided data and storing it in the memory when the data receiving unit 3 receives the data group of FIG. 14 (b).
- the first data XI is read (s 15 1), and the read data X 1 is written in the memory (S 15 2).
- the start address of the area where data X 2 is stored is written to the end of data XI (S 15 4), and then data X 2 is written to memory (S155).
- the data Y1 is read (S157), and the read data Y1 is written to the memory (S15.8).
- data X3 is read (S159), and at this time, the start address of the area where data X3 is stored is written to the end of data X2 (S160), and then data X3 is read. Is written to the memory (S1661). Then, the data Y 2 is read (S 16 2), and at this time, the start address of the area where the data Y 2 is stored is written at the end of the data Y 1 (S 16 3), and then the data Y 2 is written. Write to the memory (S164). Hereinafter, similarly, data X4 to data X6 are written to the memory.
- the divided data can be easily synthesized and restored. That is, for the data X, since the divided data XI, X2, ' ⁇ ' ⁇ 6 are connected in a chain via the starting address, as shown in FIG. Even if the divided data is mixed and stored, the original data X can be easily restored by reading the data of XI, X 2, —X 6 with reference to the start address and combining them. Can be. The same applies to data ⁇ .
- FIG. 17 is a diagram for explaining processing of audio data having a silent section.
- the voice of an announcer is recorded as a voice signal and embedded in the SMF format shown in FIG. 5 (b) or 7 (b).
- the voice of the announcer may be interrupted on the way, and the data in this interrupted section (silent section) is essentially unnecessary data. Therefore, cut the data of this silent section and — —
- the amount of data can be reduced.
- the section of T is a silent section.
- the silent section T is a section in which the signal level is originally 0, but the level is not always 0 due to noise and the like. Therefore, a level value L within a certain range is determined, and when a section in which the signal level does not exceed L continues for a certain section, this section is defined as a silent section T. Then, voice data in which the silent section T is cut is created, embedded in the SMF format shown in FIG. 5 (b) or FIG. 7 (b), and played according to the above-described playback method. By doing so, the amount of data to be transmitted can be reduced, and the memory capacity of the receiving side can be saved.
- W1 to W4 are portions to be subjected to window processing.
- Fig. 18 shows a chart for recording data by cutting out silent sections.
- Data is read sequentially from the beginning (S 171), and it is determined whether or not the level of the read data exceeds a certain value (S 172). If it does not exceed the certain value (S172N ⁇ ), return to S171 and read the data continuously, and if it exceeds the certain value (S1772YES), the above described near the rise of the data Window processing is performed, and the processed data is written to the memory (S173).
- the window processing here is the window processing at W1 in FIG. 17, and is a fade-in processing in which the signal rises slowly.
- the data is read again (S174), and it is determined whether or not the level of the read data exceeds a certain value (S175). If the value exceeds a certain value (S175YES), the data is written to the memory (S176), and the process returns to S174 to read the next data. If it does not exceed a certain value (S175NO), it is determined whether the section is continuous for a certain section (S177), and if it is not continuous (S177NO). Then, write the data to the memory (S176), and return to S174 to read the next data.
- a section that does not exceed a certain level is continuous for a certain section (S177 YES)
- the section is regarded as a silent section and window processing is performed on the portion W2 in Fig. 1-7, and processing is performed.
- the subsequent data is written to the memory (S178).
- the window processing here is processing of the feed-out where the signal gradually falls.
- S 178 of the data written in S 176, a process of deleting unnecessary data in a silent section is also performed.
- audio, characters, and images are taken as information to be embedded in the SMF extended format.
- the information to be embedded may be anything, for example, a computer program.
- the computer program is played following the MIDI data, the MIDI music will be played first, and the program will start up automatically after this. Can be used.
- the file 1a of the server on the network is used.
- data is received via a communication line from a personal computer.However, data in SMF format is created on a personal computer, stored in the file 1b on the hard disk, and the data is downloaded from here. You can do it.
- FIG. 19 shows another example of the data reproducing apparatus according to the present invention.
- 1a and 1b are files in which data is recorded
- 1a is a file on a server on the Internet, for example
- 1b is a file on a node disk inside the device, for example.
- Reference numeral 2 denotes a CPU which controls the entire data reproducing apparatus, and includes a data receiving unit 3 and a data distribution unit 4.
- CPU 2 includes blocks having various other functions, but is not shown in the drawing because it is not directly related to the present invention.
- the data receiving unit 3 accesses the files 1a and 1b and receives the data stored therein.
- the data of the file la is received via a wire or wirelessly.
- the format of the received data is the same as in FIGS. 2 to 8. These received data are temporarily stored in the buffer 3a.
- the data sorting unit 4 sorts the data received by the data receiving unit 3 by type and stores the sorted data in the buffers 7 to 10 constituting the storage unit 5.
- Reference numeral 6 denotes a data reproducing unit which includes a MIDI reproducing unit 11 for processing data relating to MIDI, an audio reproducing unit 12 for processing data relating to voice, a character reproducing unit 13 for processing data relating to characters, and And an image reproducing unit 14 for processing related data.
- the MIDI playback unit 11 includes the sound source ROM 11a shown in FIG.
- the image reproducing unit 14 has a function of reproducing a still image and a moving image.
- a mixer 15 mixes the outputs of the MIDI playback unit 11 and the audio playback unit 12
- a mixer 16 mixes the outputs of the character reproducing unit 13 and the image reproducing unit 14.
- the mixer 15 has the sound effect unit 15a shown in FIG. 1, and the mixer 16 has the visual effect unit 16a shown in FIG.
- Reference numeral 17 denotes an output buffer in which the output of the mixer 15 is temporarily stored
- reference numeral 18 denotes an output buffer in which the output of the mixer 16 is temporarily stored.
- Reference numeral 19 denotes a speaker as a sound generator for outputting a sound based on the data in the output buffer 17, and reference numeral 20 denotes a display for displaying visual information such as characters and pictures based on the data in the output buffer 18. is there.
- Reference numeral 21 denotes a timing control unit that generates a system clock serving as a reference time of the system and controls the timing of each unit
- 22 denotes an external storage device externally attached to the data reproducing device.
- the storage unit 5, the data reproduction unit 6, the mixers 15 and 16, the output buffers 17 and 18, and the timing control unit 21 are composed of a DSP (Digital Signal Processor).
- DSP Digital Signal Processor
- buffers 7 to 10 are provided between the data distribution unit 4 and the data reproducing unit 6.
- a storage unit 5 is provided, and a timing control unit 21 is provided. Further, an external storage device 22 is also added.
- FIG. 20 is a flowchart showing the entire operation of the data reproducing apparatus shown in FIG.
- the data receiving unit 3 receives data from the file la or the file 1b (S181). This received data is stored in buffer 3a.
- the CPU 2 sorts the data based on the system clock from the timing control unit 21 and the delta time of each data received by the data receiving unit 3. —
- the data sorting unit 4 sorts the data to be processed according to the result of the time calculation according to the type, and stores the data in the corresponding buffers 7 to 10 (S183). The details of S183 will be described later.
- the data stored in the buffers 7 to 10 are read out by the data reproducing units 11 to 14 corresponding to the buffers, respectively, and are converted into data in the data reproducing units 11 to 14.
- the recorded event is executed and the data is reproduced (S184). Details of S184 will also be described later.
- MID I and audio data are mixed by mixer 15 and text and image data are mixed by mixer 16 (S 18).
- FIG. 21 is a view for explaining the principle of time calculation in S182.
- t is the time axis
- events 0 to 4 indicate the playback timing of the events included in the received data string (however, this playback timing is the Note that the timing is assumed to be played according to their delta-time, and is not the timing actually played on the time axis t).
- event 0 is a picture event
- event 1 is a MIDI event
- event 2 is a voice event
- event 3 is a text event
- event 4 is a picture event.
- ⁇ 1 to T4 is the delta time
- ⁇ 1 is the delta time for event 1
- ⁇ 2 is the delta time for event 2
- ⁇ 3 is the delta time for event 3.
- ⁇ Time, ⁇ ⁇ 4 is the delta time of event 4.
- the delta time is the time from when the previous event was executed to when this event was executed.
- Event 2 is executed when ⁇ T2 elapses from the time when event 1 is executed, and event 3 is executed when ⁇ 3 elapses from the time when event 2 is executed.
- t 1 represents the time when the previous data was processed
- t 2 represents the current time
- the difference t 2 —t 1 corresponds to one frame, which is a unit interval.
- This one-frame section has a time width of, for example, 15 ms.
- the first and last timings of one frame are determined by the timing control unit 21 (see FIG. 19). Determined by the clock.
- Q is a data processing section, and is defined as the difference between the current time t2 and the execution time t0 of the last event (event 0) in the immediately preceding frame.
- FIG. 22 is a flowchart showing the procedure of data distribution by the data distribution unit 4.
- a procedure for distributing data will be described with reference to FIGS. 21 and 22.
- the system enters the WAKE state (S191), and CPU2 Calculate the time width of the processing section Q. (S192) This Q is, as described above,
- event 0 has already been processed the previous time, so judgment is made in order from event 1.
- the judgment in S194 is YES, and then it is judged whether or not the data is MIDI (S195).
- the event 1 is an MDI event (S 195 Y E S)
- the data is sent to the buffer 7 to temporarily store the data (S 196).
- the event 1 is not an MDI event (S195N ⁇ )
- it is determined whether or not the event is a voice event (S198). If the event 1 is a voice event (S 198 Y E S), the data is sent to the buffer 8 and the data is temporarily stored (S 199).
- Event 1 is not a voice event (S 198 NO), it is determined whether or not it is a character event (S 200). If the event 1 is a character event (S200YES), the data is sent to the buffer 9 and the data is temporarily stored (S201). If the event 1 is not a character event (S200NO), it is determined whether or not the event is an image event (S202). If the event 1 is an image event (S202YS), the data is sent to the buffer 10 and the data is temporarily stored (S203). If event 1 is not an event of the image (S202 NO), other processing is performed.
- S 192 to S 194 and S 197 are details of S 182 of FIG. 20, and S 195 and S 199 6, S198 to S203 are details of S189 of FIG.
- FIG. 23 is a flowchart showing the processing procedure in each data reproducing unit.
- (A) shows the processing procedure in the MIDI reproducing unit 11. ⁇ 11 0 1 When the data for one frame section distributed by the data distribution section 4 is stored in the buffer 7, the reproduction section 11 transfers this data to the next one frame section. —
- FIG. 23 (b) shows a processing procedure in the audio reproduction unit 12.
- the audio reproducing section 12 reads this data in the next one frame section (S311). Then, the audio data (see FIGS. 5 (b) and 7 (b)) recorded in the event of the read data is decoded, and the audio is reproduced (S312).
- This playback data is temporarily stored in a buffer (not shown) inside the audio playback unit 12, and is output from this buffer to the mixer 15 (S3 13).
- FIG. 23 (c) shows a processing procedure in the character reproducing unit 13.
- the character reproducing section 13 reads this data in the next one frame section (S411). Then, the character data (see FIGS. 5 (c) and 7 (c)) recorded in the event of the read data is decoded, and the characters are reproduced (S412).
- the reproduced data is temporarily stored in a buffer (not shown) inside the character reproducing unit 13 and is output from this buffer to the mixer 16 (S413).
- FIG. 23 (d) shows a processing procedure in the image reproducing unit 14.
- the data of one frame section distributed by the data distribution unit 4 is stored in the buffer 10, the data is One —
- the data is read in the next one frame interval (S511). Then, the image data recorded in the event of the read data (Fig. 5
- the reproduced data is temporarily stored in a buffer (not shown) inside the image reproducing unit 14 and is output from the buffer to the mixer 16 (S513).
- FIGS. 23 (a) to (d) described above are performed in accordance with the order defined by the program.
- the processes are performed in the order of (a) to (d). That is, the MIDI processing of (a) is performed first, and when this is completed, the processing proceeds to the audio processing of (b).
- the processing proceeds to the character processing of (c), and when the character processing is completed, (d ) Image processing.
- the reason why the processing is performed in series in this manner is that the storage unit 5 and the data reproduction unit 6 have only one DSP, and the DSP is provided for each reproduction unit. In this case, processing can be performed in parallel.
- the MIDI playback data output to the mixer 15 in S2 13 and the audio playback data output to the mixer 15 in S3 13 are mixed by the mixer 15 and output to the output buffer 17. It is stored and output from the speaker as a sound. Also, the reproduced data of the character output to the mixer 16 in S4 13 and the reproduced data of the image output to the mixer 16 in S5 13 are mixed by the mixer 16 and output buffer 18 And displayed on the display unit 20 as visible information.
- the output buffer 17 and the speaker 19 constitute a first output unit, and the output buffer 18 and the display unit 20 constitute a second output unit.
- the output buffer 17 has a function of counting the number of data to be output to the speaker 19, and sends a control signal to the timing control section 21 based on the counted value, and outputs the control signal to the timing control section 2. 1 is based on this control signal.
- a timing signal system clock
- the timing control section 21 also supplies a timing signal to the output buffer 18 in accordance with the above control signal, and controls the timing of the data output from the output buffer 18.
- Fig. 24 shows the overall operation from the above-mentioned data distribution to playback, and (a) shows the relationship between the amount of data processed by each playback unit and the frame section. (B) shows the relationship between the processing time in each playback unit and the frame section.
- F 1 to F 3 is 1 full rate-time interval, the time width of each frame interval is set to for example 1 5 m S. That is, the data distribution unit 4 receives a clock interrupt from the timing control unit 21 every 15 ms.
- t indicates a time axis
- M indicates an event of MIDI
- A indicates a voice event
- T indicates a character event
- P indicates a playback timing of an image event. Note that these playback timings show the timing when it is assumed that the received data has been played back according to the delta time, as in Fig. 21. It does not indicate the timing of the playback.
- the data processed in the section F1 is distributed to the buffers 7 to 10 and stored at the last timing of the section. Then, each of the reproducing units 11 to 14 reads out data from the buffer in the next one frame period F2 and performs a reproducing process.
- the amount of data transferred from each buffer to each playback unit is an amount of data that each playback unit can process in one frame interval. As shown in the figure, each playback unit can process all data in the next one frame section F2.
- the time chart for this processing is shown in FIG. 24 (b), and the length of the white arrow indicates the processing time.
- This processing time differs for each frame.
- the data stored in the buffer is sequentially read out by the reproducing units 11 to 14 in a predetermined order in the next one frame period F2, and In, the event recorded in the data is executed, and the data is reproduced.
- M (M I D I), A (audio), and P (image) are reproduced in this order.
- the regenerated M and A are used in mixer 1 (mixer in Fig. 19).
- the P processed and regenerated in 15) is mixed with mixer 2 (mixer in Fig. 19).
- data of A, M, and T are distributed to the buffer, and these data are read out in the order of M, A, and T in the F3 section, and the same as above in each playback unit. It is reproduced in the same way and output in the next F4 section (not shown in Fig. 24).
- the received data is sorted for each frame, stored in the buffer, and read out from the buffer in the next frame to reproduce the data. Next It is output as sound-visible information in frames. Therefore, it is possible to reproduce the data while synchronizing the data in frame units.
- the data distribution unit 4 concentrates on the work of distributing the received data to the buffers 7 to 10, and the reproduction units 11 to 14 concentrate on reading out the data stored in the buffer and reproducing the data. Therefore, the data received by the data receiving unit 3 can be pipelined and processed at high speed.
- the reproduction timing should be controlled according to the delta time.
- the individual delta times have virtually no meaning in determining the playback timing.
- one frame period is as short as 15 ms as described above, the data reproduced during this period is considered to be reproduced simultaneously, regardless of the reproduction timing of each data. No problem. In fact, it has been empirically confirmed that the deviation of the data reproduction timing within a section of about 15 ms cannot be identified by ordinary human senses. Therefore, at the time of data distribution, if the data to be processed in one frame interval is determined based on the delta time, the reproduction timing of those data follows the delta time in one frame interval. There is no problem even if it is off the playback timing.
- each playback unit reads data from the buffer according to the order M, A, and P of the received data.
- the order of the received data is Is A, M, T, but the playback unit is a buffer
- the order of reading data from them is M, A, T, and ⁇ and ⁇ are interchanged.
- the processing order in each reproducing unit is determined by the program as ⁇ , A, D, and P.
- the processing order is changed in this way, if each playback unit performs data processing within 15 ms, the data playback timing is invisible to human senses as described above, which is a problem. There is no.
- the data allocated in one frame interval is processed in the next one frame interval, but this is not always essential. That is, if the output buffer has a size that exceeds the processing amount in one frame section, the output buffer will not be processed even if there is data that could not be processed in one frame. Since the previously processed data remains in 17 and 18, the data can be output without interruption.
- FIG. 25 is a diagram for explaining the operation of the data receiving section 3 when the data reproducing apparatus shown in FIG. 1 or FIG. 19 performs reproduction while downloading data and adopts a stream system.
- the buffer 3a is composed of three buffers, a buffer A, a buffer B, and a buffer C.
- 3b is registers A, B, and C provided for buffers A, B, and C, respectively.
- the data received is shown as stream data S.
- a header H is recorded, followed by MIDI, voice, character, and image data in packets P1, P2, P3, ".
- the total data amount of the stream data S is assumed to be ⁇ .
- the reception operation will be described by taking the case of playing music as an example.
- the data receiving unit 3 starts receiving the stream data S from the file 1a, first, the stream data S From the beginning, data A 1 equivalent to the size (capacity) of buffer A is stored in buffer A. As a result, the buffer A becomes full, and a flag indicating that the buffer A is full is set in the register A. Subsequently, data B 1 corresponding to the size of the buffer B is stored in the buffer B. As a result, the buffer B is also in the full state, and the flag indicating that the buffer B is in the full state is set in the register B.
- the data distribution unit 4 starts distributing the data, and compares the data A 1 stored in the buffer A and the data B 1 stored in the buffer B with the data. Transfer to buffers 7 to 10 by type. The transferred data is played back by each of the playback units 11 to 14, and the performance of the music starts.
- buffer C stores data C 1 corresponding to the size. As a result, the buffer C becomes full, and a flag indicating that the buffer C is full is set in the register C.
- the data cache as described above is not necessary, but when the buffer becomes full, It is necessary to instruct the server from the data receiving unit 3 not to transmit the above data. In this case, when the data in the buffer is consumed and the buffer becomes empty, the data receiving unit 3 acquires the data from the server.
- the data distribution unit 4 starts distribution of data from the time when data A 1 is received.
- the data receiving unit 3 requests the server to stop data transmission when the buffer becomes full. Then, when the data is consumed and the buffer becomes available, the data receiving unit 3 requests the server to transmit the data again. That is, the data receiving unit 3 intermittently acquires data from the server.
- MIDI data is stored in an internal memory (not shown) of the mobile phone 50 so as to correspond to the incoming signal, and the notification is made in a different manner according to the incoming signal. This makes it easy to identify who is calling.
- a vibrator (not shown) for incoming call notification built in the mobile phone 50 may be vibrated based on the MDI data, for example, the vibrator may be vibrated at the same rhythm as the drum rhythm. In addition, it can be used to add BGM (back 'ground' music) by MID I to the answering machine message.
- “Y Y Y” is information indicating what the purchased music data is.
- the viewer first downloads music data from file 1a on a server on the Internet (S701).
- the music data is received by the data receiving unit 3, distributed to each unit by the data distribution unit 4, reproduced according to the above-described procedure, and output from the speakers 19 and the display 20.
- Each data is stored in the external storage device 22 (the information storage medium 70 in FIG. 37).
- the received character data is transferred to the character
- the data reproducing apparatus of the present invention can be mounted on various types of information terminals such as the above-described portable telephones and game machines, personal computers and STBs (Set Top Boxes) for Internet TV.
- information terminals such as the above-described portable telephones and game machines, personal computers and STBs (Set Top Boxes) for Internet TV.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Electrophonic Musical Instruments (AREA)
- Telephonic Communication Services (AREA)
- Circuits Of Receivers In General (AREA)
- Reverberation, Karaoke And Other Acoustics (AREA)
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP00902081.9A EP1172796B1 (en) | 1999-03-08 | 2000-02-03 | Data reproducing device, data reproducing method, and information terminal |
JP2000604397A JP4236024B2 (ja) | 1999-03-08 | 2000-02-03 | データ再生装置および情報端末機 |
AU23258/00A AU2325800A (en) | 1999-03-08 | 2000-02-03 | Data reproducing device, data reproducing method, and information terminal |
US09/936,055 US6979769B1 (en) | 1999-03-08 | 2000-02-03 | Data reproducing device, data reproducing method, and information terminal |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP11/60919 | 1999-03-08 | ||
JP6091999 | 1999-03-08 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2000054249A1 true WO2000054249A1 (fr) | 2000-09-14 |
Family
ID=13156286
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2000/000602 WO2000054249A1 (fr) | 1999-03-08 | 2000-02-03 | Dispositif de reproduction de donnees, procede de reproduction de donnees et terminal d'informations |
Country Status (7)
Country | Link |
---|---|
US (1) | US6979769B1 (ja) |
EP (1) | EP1172796B1 (ja) |
JP (1) | JP4236024B2 (ja) |
KR (1) | KR100424231B1 (ja) |
CN (1) | CN1175393C (ja) |
AU (1) | AU2325800A (ja) |
WO (1) | WO2000054249A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1291843A1 (en) * | 2001-09-04 | 2003-03-12 | Yamaha Corporation | Electronic music apparatus that enables user to purchase music related product from server |
JP2003304309A (ja) * | 2003-04-11 | 2003-10-24 | Sharp Corp | 携帯端末装置、携帯端末装置の制御プログラム、および携帯端末装置の制御プログラムを記録したコンピュータ読取可能な記録媒体 |
JP2008233557A (ja) * | 2007-03-20 | 2008-10-02 | Yamaha Corp | 電子楽器及びプログラム |
JP2008233558A (ja) * | 2007-03-20 | 2008-10-02 | Yamaha Corp | 電子楽器及びプログラム |
Families Citing this family (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6121536A (en) * | 1999-04-29 | 2000-09-19 | International Business Machines Corporation | Method and apparatus for encoding text in a MIDI datastream |
JP2002045567A (ja) * | 2000-08-02 | 2002-02-12 | Konami Co Ltd | 携帯端末装置、ゲーム実行支援装置および記録媒体 |
DE60133660T2 (de) * | 2000-09-25 | 2009-05-28 | Yamaha Corp., Hamamatsu | Mobiles endgerät |
WO2003015409A1 (en) * | 2001-08-02 | 2003-02-20 | Matsushita Electric Industrial Co., Ltd. | Point-used electronic trading system, point-used electronic trading method, broadcast reception apparatus, and broadcast reception method |
US7708642B2 (en) * | 2001-10-15 | 2010-05-04 | Igt | Gaming device having pitch-shifted sound and music |
GB0127234D0 (en) * | 2001-11-13 | 2002-01-02 | British Sky Broadcasting Ltd | Improvements in receivers for television signals |
JP2003162355A (ja) * | 2001-11-26 | 2003-06-06 | Sony Corp | タスクの表示切り替え方法及び携帯機器及び携帯通信機器 |
KR20030043299A (ko) * | 2001-11-27 | 2003-06-02 | 주식회사 엘지이아이 | 오디오 데이터와 부가 데이터간의 동기 기록 관리 및재생방법 |
KR100563680B1 (ko) * | 2001-11-27 | 2006-03-28 | 엘지전자 주식회사 | 재기록 가능 기록매체의 오디오 가사 데이터 기록 관리 및재생방법 |
US7863513B2 (en) * | 2002-08-22 | 2011-01-04 | Yamaha Corporation | Synchronous playback system for reproducing music in good ensemble and recorder and player for the ensemble |
CN100338590C (zh) * | 2002-09-27 | 2007-09-19 | 松下电器产业株式会社 | 内容发送装置及接收装置、内容发送接收系统及其方法 |
EP1435603A1 (en) * | 2002-12-06 | 2004-07-07 | Sony Ericsson Mobile Communications AB | compact media data format |
AU2003298150A1 (en) * | 2002-12-06 | 2004-06-30 | Sony Ericsson Mobile Communications Ab | Compact media data format |
EP1735848A4 (en) * | 2003-01-07 | 2010-05-19 | Medialab Solutions Llc | SYSTEMS AND METHOD FOR LOCATED AUDIOSYNTHESIS |
CN100380367C (zh) | 2003-09-28 | 2008-04-09 | 诺基亚公司 | 具有音乐数据库的电子设备和形成音乐数据库的方法 |
EP1544845A1 (en) * | 2003-12-18 | 2005-06-22 | Telefonaktiebolaget LM Ericsson (publ) | Encoding and Decoding of Multimedia Information in Midi Format |
JP4702689B2 (ja) * | 2003-12-26 | 2011-06-15 | ヤマハ株式会社 | 音楽コンテンツ利用装置及びプログラム |
JP4489442B2 (ja) | 2004-01-13 | 2010-06-23 | ヤマハ株式会社 | 鍵盤装置 |
JP4453393B2 (ja) * | 2004-02-26 | 2010-04-21 | ヤマハ株式会社 | 音楽コンテンツを再生し得る電子音楽装置及びそのプログラム |
US7179979B2 (en) * | 2004-06-02 | 2007-02-20 | Alan Steven Howarth | Frequency spectrum conversion to natural harmonic frequencies process |
JP4284620B2 (ja) * | 2004-12-27 | 2009-06-24 | ソニー株式会社 | 情報処理装置および方法、並びにプログラム |
JP4277218B2 (ja) * | 2005-02-07 | 2009-06-10 | ソニー株式会社 | 記録再生装置、その方法及びプログラム |
JP4321476B2 (ja) * | 2005-03-31 | 2009-08-26 | ヤマハ株式会社 | 電子楽器 |
ES2334514T3 (es) * | 2005-05-12 | 2010-03-11 | Ipg Electronics 504 Limited | Metodo para sincronizar al menos un periferico multimedia de un dispositivo de comunicacion portatil con un archivo de audio, y dispositivo de comunicacion portatil correspondiente. |
US20090160862A1 (en) * | 2005-10-13 | 2009-06-25 | Tae Hyeon Kim | Method and Apparatus for Encoding/Decoding |
WO2007043824A1 (en) * | 2005-10-13 | 2007-04-19 | Lg Electronics Inc. | Method and apparatus for encoding/decoding |
KR20070081368A (ko) * | 2006-02-10 | 2007-08-16 | 삼성전자주식회사 | 노래 가사의 반복 패턴을 기초로 가사 구조를 추출하는장치, 시스템, 및 그 방법 |
WO2008007905A1 (en) * | 2006-07-12 | 2008-01-17 | Lg Electronics Inc. | Method and apparatus for encoding/decoding signal |
WO2008048064A1 (en) * | 2006-10-19 | 2008-04-24 | Lg Electronics Inc. | Encoding method and apparatus and decoding method and apparatus |
US20080222685A1 (en) * | 2007-03-09 | 2008-09-11 | At&T Knowledge Ventures, L.P. | Karaoke system provided through an internet protocol television system |
US7968785B2 (en) * | 2008-06-30 | 2011-06-28 | Alan Steven Howarth | Frequency spectrum conversion to natural harmonic frequencies process |
JP4748330B2 (ja) * | 2008-07-31 | 2011-08-17 | セイコーエプソン株式会社 | 伝送装置、伝送システム、プログラムおよび情報記憶媒体 |
JP5399831B2 (ja) * | 2009-09-11 | 2014-01-29 | 株式会社コナミデジタルエンタテインメント | 音楽ゲームシステム及びそのコンピュータプログラム並びに効果音データの生成方法 |
CN101694772B (zh) * | 2009-10-21 | 2014-07-30 | 北京中星微电子有限公司 | 将文本文字转换成说唱音乐的方法及装置 |
US8649727B2 (en) * | 2010-11-01 | 2014-02-11 | Fu-Cheng PAN | Portable karaoke system, karaoke method and application program for the same |
TWI480854B (zh) * | 2010-11-04 | 2015-04-11 | Fu Cheng Pan | 攜帶型伴唱系統、伴唱方法及應用程式 |
JP5477927B2 (ja) * | 2011-01-28 | 2014-04-23 | 日本電気株式会社 | ストレージシステム |
KR101932539B1 (ko) * | 2013-02-18 | 2018-12-27 | 한화테크윈 주식회사 | 동영상 데이터를 기록하는 방법, 및 이 방법을 채용한 촬영 장치 |
JP6402878B2 (ja) * | 2013-03-14 | 2018-10-10 | カシオ計算機株式会社 | 演奏装置、演奏方法およびプログラム |
CN103310776B (zh) * | 2013-05-29 | 2015-12-09 | 亿览在线网络技术(北京)有限公司 | 一种实时混音的方法和装置 |
KR20180012800A (ko) * | 2015-05-27 | 2018-02-06 | 광저우 쿠고우 컴퓨터 테크놀로지 컴퍼니, 리미티드 | 오디오 처리 방법, 장치 및 시스템 |
JP6729699B2 (ja) * | 2016-07-22 | 2020-07-22 | ヤマハ株式会社 | 制御方法、及び、制御装置 |
WO2021049181A1 (ja) * | 2019-09-10 | 2021-03-18 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
EP4120239A4 (en) * | 2020-09-04 | 2023-06-07 | Roland Corporation | INFORMATION PROCESSING DEVICE AND INFORMATION PROCESSING METHOD |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH044473U (ja) * | 1990-04-27 | 1992-01-16 | ||
JPH05110536A (ja) * | 1991-10-11 | 1993-04-30 | Nec Corp | Dsi音声雑音切替器 |
JPH06318090A (ja) * | 1993-05-10 | 1994-11-15 | Brother Ind Ltd | カラオケ通信システム |
JPH0728462A (ja) * | 1993-07-12 | 1995-01-31 | Kawai Musical Instr Mfg Co Ltd | 自動演奏装置 |
JPH07327222A (ja) * | 1994-06-01 | 1995-12-12 | Ekushingu:Kk | データ伝送装置 |
JPH0854888A (ja) * | 1994-08-12 | 1996-02-27 | Matsushita Electric Ind Co Ltd | 音楽データ伝送装置 |
JPH0887271A (ja) * | 1994-09-17 | 1996-04-02 | Victor Co Of Japan Ltd | 演奏情報圧縮方法及び演奏情報伸張方法 |
EP0715296A2 (en) * | 1994-12-02 | 1996-06-05 | Sony Corporation | Sound source controlling device |
JPH09214371A (ja) * | 1996-02-01 | 1997-08-15 | Matsushita Electric Ind Co Ltd | 車載用音響装置 |
JPH1020877A (ja) * | 1996-07-04 | 1998-01-23 | Brother Ind Ltd | 楽音再生装置 |
JPH10105186A (ja) * | 1996-09-28 | 1998-04-24 | Brother Ind Ltd | 楽音再生装置 |
JPH10187174A (ja) * | 1996-12-19 | 1998-07-14 | Nec Corp | 通信カラオケシステム |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR940004830B1 (ko) * | 1991-03-14 | 1994-06-01 | 주식회사 금성사 | 연주용 데이타화일 기록방법 및 재생장치 |
JP3149093B2 (ja) | 1991-11-21 | 2001-03-26 | カシオ計算機株式会社 | 自動演奏装置 |
JP2521016B2 (ja) * | 1991-12-31 | 1996-07-31 | インターナショナル・ビジネス・マシーンズ・コーポレイション | マルチメディア・デ―タ処理システム |
JP3540344B2 (ja) * | 1993-07-27 | 2004-07-07 | 株式会社リコス | カラオケ装置におけるバックコーラス再生装置 |
US5768350A (en) * | 1994-09-19 | 1998-06-16 | Phylon Communications, Inc. | Real-time and non-real-time data multplexing over telephone lines |
JPH09134173A (ja) | 1995-11-10 | 1997-05-20 | Roland Corp | 自動演奏装置の表示制御方法および表示制御装置 |
US5953005A (en) * | 1996-06-28 | 1999-09-14 | Sun Microsystems, Inc. | System and method for on-line multimedia access |
US5815426A (en) * | 1996-08-13 | 1998-09-29 | Nexcom Technology, Inc. | Adapter for interfacing an insertable/removable digital memory apparatus to a host data part |
US5764965A (en) * | 1996-09-23 | 1998-06-09 | Silicon Graphics, Inc. | Synchronization infrastructure for use in a computer system |
US6283764B2 (en) * | 1996-09-30 | 2001-09-04 | Fujitsu Limited | Storage medium playback system and method |
JPH10124071A (ja) | 1996-10-16 | 1998-05-15 | Xing:Kk | カラオケ装置 |
JPH10150505A (ja) | 1996-11-19 | 1998-06-02 | Sony Corp | 情報通信処理方法及び情報通信処理装置 |
US5951646A (en) * | 1996-11-25 | 1999-09-14 | America Online, Inc. | System and method for scheduling and processing image and sound data |
JPH10173737A (ja) | 1996-12-06 | 1998-06-26 | Digital Vision Lab:Kk | 個人用機器 |
JPH10198361A (ja) | 1997-01-09 | 1998-07-31 | Yamaha Corp | 電子楽器及び記憶媒体 |
JP3405181B2 (ja) | 1997-03-11 | 2003-05-12 | ヤマハ株式会社 | 楽音発生方法 |
WO1999040566A1 (fr) * | 1998-02-09 | 1999-08-12 | Sony Corporation | Procede et appareil de traitement de signaux numeriques, procede et appareil de generation de donnees de commande et support pour programme d'enregistrement |
JP3801356B2 (ja) * | 1998-07-22 | 2006-07-26 | ヤマハ株式会社 | データ付き楽曲情報作成装置、再生装置、送受信システム及び記録媒体 |
JP2000105595A (ja) * | 1998-09-30 | 2000-04-11 | Victor Co Of Japan Ltd | 歌唱装置及び記録媒体 |
JP2000181449A (ja) * | 1998-12-15 | 2000-06-30 | Sony Corp | 情報処理装置および方法、並びに提供媒体 |
-
2000
- 2000-02-03 JP JP2000604397A patent/JP4236024B2/ja not_active Expired - Lifetime
- 2000-02-03 AU AU23258/00A patent/AU2325800A/en not_active Abandoned
- 2000-02-03 WO PCT/JP2000/000602 patent/WO2000054249A1/ja active IP Right Grant
- 2000-02-03 KR KR10-2001-7011396A patent/KR100424231B1/ko not_active Expired - Lifetime
- 2000-02-03 US US09/936,055 patent/US6979769B1/en not_active Expired - Lifetime
- 2000-02-03 CN CNB008047952A patent/CN1175393C/zh not_active Expired - Fee Related
- 2000-02-03 EP EP00902081.9A patent/EP1172796B1/en not_active Expired - Lifetime
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH044473U (ja) * | 1990-04-27 | 1992-01-16 | ||
JPH05110536A (ja) * | 1991-10-11 | 1993-04-30 | Nec Corp | Dsi音声雑音切替器 |
JPH06318090A (ja) * | 1993-05-10 | 1994-11-15 | Brother Ind Ltd | カラオケ通信システム |
JPH0728462A (ja) * | 1993-07-12 | 1995-01-31 | Kawai Musical Instr Mfg Co Ltd | 自動演奏装置 |
JPH07327222A (ja) * | 1994-06-01 | 1995-12-12 | Ekushingu:Kk | データ伝送装置 |
JPH0854888A (ja) * | 1994-08-12 | 1996-02-27 | Matsushita Electric Ind Co Ltd | 音楽データ伝送装置 |
JPH0887271A (ja) * | 1994-09-17 | 1996-04-02 | Victor Co Of Japan Ltd | 演奏情報圧縮方法及び演奏情報伸張方法 |
EP0715296A2 (en) * | 1994-12-02 | 1996-06-05 | Sony Corporation | Sound source controlling device |
JPH09214371A (ja) * | 1996-02-01 | 1997-08-15 | Matsushita Electric Ind Co Ltd | 車載用音響装置 |
JPH1020877A (ja) * | 1996-07-04 | 1998-01-23 | Brother Ind Ltd | 楽音再生装置 |
JPH10105186A (ja) * | 1996-09-28 | 1998-04-24 | Brother Ind Ltd | 楽音再生装置 |
JPH10187174A (ja) * | 1996-12-19 | 1998-07-14 | Nec Corp | 通信カラオケシステム |
Non-Patent Citations (1)
Title |
---|
"Kabushiki Kaisha Ritto Music", MIDI BIBLE II, 30 March 1998 (1998-03-30), JAPAN, pages 71,134 - 135, XP002935471 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1291843A1 (en) * | 2001-09-04 | 2003-03-12 | Yamaha Corporation | Electronic music apparatus that enables user to purchase music related product from server |
JP2003304309A (ja) * | 2003-04-11 | 2003-10-24 | Sharp Corp | 携帯端末装置、携帯端末装置の制御プログラム、および携帯端末装置の制御プログラムを記録したコンピュータ読取可能な記録媒体 |
JP2008233557A (ja) * | 2007-03-20 | 2008-10-02 | Yamaha Corp | 電子楽器及びプログラム |
JP2008233558A (ja) * | 2007-03-20 | 2008-10-02 | Yamaha Corp | 電子楽器及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
EP1172796B1 (en) | 2016-11-09 |
AU2325800A (en) | 2000-09-28 |
JP4236024B2 (ja) | 2009-03-11 |
US6979769B1 (en) | 2005-12-27 |
EP1172796A4 (en) | 2007-05-30 |
EP1172796A1 (en) | 2002-01-16 |
KR100424231B1 (ko) | 2004-03-25 |
KR20010102534A (ko) | 2001-11-15 |
CN1343348A (zh) | 2002-04-03 |
CN1175393C (zh) | 2004-11-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4236024B2 (ja) | データ再生装置および情報端末機 | |
US20080184870A1 (en) | System, method, device, and computer program product providing for a multiple-lyric karaoke system | |
KR100591378B1 (ko) | 악음 재생 장치와 방법, 및 휴대 단말 장치 | |
JP2000224269A (ja) | 電話機および電話システム | |
JP7149193B2 (ja) | カラオケシステム | |
US20060194626A1 (en) | Defined ringing tone segments in an audio source | |
CN102394860A (zh) | 信号传送系统、方法、电脑程序产品及电脑可读取储存媒体 | |
KR20010076533A (ko) | 휴대전화 단말기의 노래방 기능 구현 및 사용방법 | |
JP4229058B2 (ja) | 端末装置および記録媒体 | |
JP2002073049A (ja) | 音楽配信サーバ、音楽再生端末、及びサーバ処理プログラムを記憶した記憶媒体、端末処理プログラムを記憶した記憶媒体 | |
JP3772072B2 (ja) | カラオケ楽曲の非歌唱区間にてスポット番組を映像出力するカラオケ装置 | |
JP5088881B2 (ja) | 曲間メッセージ出力制御機能を備えたカラオケ装置 | |
JP2002182658A (ja) | 楽曲データ配信方法、楽曲データ配信システム、楽曲データ配信装置および楽曲データ配信プログラム | |
JP2008089755A (ja) | 文字データ再生装置 | |
JP2004348012A (ja) | 携帯端末におけるカラオケシステム | |
JP2002091464A (ja) | 演奏中の操作履歴を記憶して再現するカラオケ装置 | |
JP3621904B2 (ja) | カラオケ映像作品を選択できるカラオケ演奏端末 | |
JP2005215304A (ja) | カラオケ演奏装置 | |
JPWO2002023522A1 (ja) | データ配信システム、データ配信方法、データ配信装置、サーバ、媒体、およびプログラム | |
JP4114344B2 (ja) | カラオケデータ再生装置 | |
JP4444069B2 (ja) | カラオケ映像作品の投稿を受け付けてカラオケ演奏時に再生可能とする通信カラオケシステム | |
JP2002140618A (ja) | 広告配信システム、サーバ、端末、方法及び記録媒体 | |
KR20060017043A (ko) | Mp3 음악을 이용한 휴대폰의 벨소리 서비스방법 | |
JP4436449B2 (ja) | 音楽情報配信システムとカラオケシステムとの連携システム | |
JP3946669B2 (ja) | 顧客の個人情報に関連したメッセージを顧客に向けて出力する通信カラオケシステム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 00804795.2 Country of ref document: CN |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AU BR CA CN ID IL IN JP KR MX NO NZ PL RU SG US |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
ENP | Entry into the national phase |
Ref document number: 2000 604397 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 09936055 Country of ref document: US Ref document number: 1020017011396 Country of ref document: KR |
|
REEP | Request for entry into the european phase |
Ref document number: 2000902081 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2000902081 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1020017011396 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2000902081 Country of ref document: EP |
|
WWG | Wipo information: grant in national office |
Ref document number: 1020017011396 Country of ref document: KR |