US20020129692A1 - Method and system for embedding audio titles - Google Patents
Method and system for embedding audio titles Download PDFInfo
- Publication number
- US20020129692A1 US20020129692A1 US09/809,019 US80901901A US2002129692A1 US 20020129692 A1 US20020129692 A1 US 20020129692A1 US 80901901 A US80901901 A US 80901901A US 2002129692 A1 US2002129692 A1 US 2002129692A1
- Authority
- US
- United States
- Prior art keywords
- audio
- information
- program
- format
- description
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 39
- 230000005236 sound signal Effects 0.000 claims abstract description 19
- 230000001131 transforming effect Effects 0.000 claims 6
- 101150048216 tag-131 gene Proteins 0.000 description 10
- 101100327917 Caenorhabditis elegans chup-1 gene Proteins 0.000 description 7
- 238000010586 diagram Methods 0.000 description 6
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0033—Recording/reproducing or transmission of music for electrophonic musical instruments
- G10H1/0041—Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/30—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
- G11B27/3027—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/011—Files or data streams containing coded musical information, e.g. for transmission
- G10H2240/046—File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
- G10H2240/061—MP3, i.e. MPEG-1 or MPEG-2 Audio Layer III, lossy audio compression
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/091—Info, i.e. juxtaposition of unrelated auxiliary information or commercial messages with or between music files
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2545—CDs
Definitions
- This invention relates in general to the field of multimedia. Specifically, this invention relates to methods and systems for delivering multimedia information.
- Digital audio is increasingly being delivered to listeners in compressed form.
- Sound compression formats such as MPEG Layer 3 (MP3), Motion Picture Experts Group, ISO/IEC 11172-3 and ISO/IEC 13818-3, substantially compress audio data, yet maintain high sound quality.
- tags have been developed to convey information that may relate to audio files. For instance, an ID3 tag may reside at the end of an MP3 file.
- ID3v2.3.0 Within a tag conforming to the ID3v2.3.0 informal standard (see http://id3.org), a host of textual information may be included, such as artist, album, song, title, lyrics, and comments.
- the ID3 tag in both ID3v2.3.0 and later revisions of the format also provides for a “general encapsulated object,” wherein any type of file may be encapsulated.
- FIG. 1 illustrates a music system involving digital audio files.
- FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention.
- FIG. 3 is a high-level block diagram of a system according to an embodiment of the present invention.
- FIG. 4 is a flow diagram illustrating a method according to an embodiment of the present invention.
- FIG. 5 is a flow diagram illustrating a method according to an embodiment of the present invention.
- the processes associated with the presented embodiments may be stored in any storage device, such as, for example, a computer system (non-volatile) memory, an optical disk, magnetic tape, or magnetic disk.
- a computer system non-volatile
- the processes may be programmed when the computer system is manufactured or via a computer-readable medium at a later date.
- Such a medium may include any of the forms listed above with respect to storage devices and may further include, for example, a carrier wave modulated, or otherwise manipulated, to convey instructions that can be read, demodulated/decoded and executed by a computer.
- a method for embedding audio titles involves receiving information, encoded in a format, about an audio program.
- the information is transformed into an audio signal that conveys an audio description of the audio program.
- the audio description and the audio program are then embedded in a predetermined format.
- FIG. 1 illustrates a music system involving digital audio files.
- system 100 includes a digital audio file 110 , such as an MP3 file, recorded on a medium 163 , and digital audio player 120 .
- Digital audio file 110 may have a predetermined format.
- digital audio file 110 comprises a tag 130 , such as an ID3 tag, and audio data 140 , which may include recorded sound, such as music or a narration of a book.
- Tag 130 may comprise metadata, or information about audio data 140 , within digital audio file 110 itself.
- tag 130 may comprise a header 135 and frames 137 .
- Header 135 may specify how tag 130 is structured.
- Frames 137 may comprise information blocks containing data.
- frames 135 may comprise data relating to lyrics 145 , program information 160 , comments 155 , and general encapsulated object 170 .
- Program information 160 may include information relating to the audio data 140 , such as title, artist, and album. Different frames and headers may be included in tag 130 depending on the applicable format of digital audio file 110 .
- Digital audio player 120 may comprise many types of digital audio playback devices.
- digital audio player 120 may comprise a CD player, a car stereo system, a handheld personal digital music player, a digital audio player on a personal computer, and the like.
- Digital audio player 120 plays digital audio selections recorded on medium 163 for the listening enjoyment of a listener. It is to be noted that the present invention may also be applied in the domain of digital video, such as where a digital video file format contains metadata.
- Digital audio player 120 may include a display 150 .
- Display 150 may list information relevant to a selection recorded on medium 163 .
- program information 160 namely title, artist, and album, has been read from frames 137 of tag 130 such that program information 160 from tag 130 is processed and displayed on display 150 of digital audio player 120 .
- Display 150 may also include information relating to the track number of the selection on medium 163 , as well as the time elapsed for a selection that is being played or paused.
- FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention.
- tag 131 Before processing, tag 131 includes general encapsulated object 170 .
- general encapsulated object 170 is a frame containing any kind of encoded data. It is contemplated that other digital audio file formats could be designed to contain such a frame.
- Tag 131 may be processed to produce tag 132 .
- general encapsulated object 170 may include compressed audio data 220 , shown in FIG. 2 by a binary numeric sequence.
- program information 160 including title, artist, and album, are read from tag 131 , recorded to an audio file via a voice synthesizer, such as a text-to-speech (TTS) engine, and embedded within general encapsulated object 210 of tag 132 as compressed audio data 220 .
- TTS text-to-speech
- Compressed audio data 220 may comprise a narrative audio description which includes information besides program information 160 .
- encoded text or other types of data may be embedded within general encapsulated object 210 .
- paragraphs of explanatory text, biographical information relating to an artist, or compressed video data may be embedded.
- the size of a frame, such as general encapsulated object 210 may impose practical limitations on the quantity and type of data suitable for embedding within a digital audio file.
- FIG. 3 is a high-level block diagram of system 300 according to an embodiment of the present invention.
- System 300 comprises transformer 320 , compressor 370 , and embedder 350 .
- Transformer 320 may comprise TTS engine 330 .
- Transformer 320 may receive encoded information 310 , such as tag 131 or an entire digital audio file. Via header information in tag 131 , transformer 320 may extract information from frames in tag 131 .
- Transformer 320 may also receive external information 380 , which is external to tag 131 and may include information from local or remote sources, such as databases accessible via the World Wide Web.
- Transformer 320 may combine information extracted from tag 131 with external information 380 .
- program information relating to a music selection in tag 131 may be combined with words and phrases to produce a narrative sentence, such as “Song 1 is The Star-Spangled Banner by Francis Scott Key.”
- user preferences, system defaults, or a combination thereof may specify which information in tag 131 , and which external information 380 , is included in the narrative.
- TTS engine 330 may process the combined textual data, translating it into speech signals. The speech signals may be recorded as an audio signal 340 and outputted by transformer 320 .
- Compressor 370 receives audio signal 340 from transformer 320 .
- Compressor 370 compresses audio signal 340 into compressed audio data 345 .
- compressor 370 may compress audio signal 340 such that compressed audio data 345 is appropriately compressed for embedding in an MP3 file. It is to be appreciated that compressor 370 may be incorporated into embedder 350 .
- Embedder 350 receives compressed audio data 345 from compressor 370 .
- Embedder 350 embeds compressed audio data 345 within general encapsulated object 210 of a tag.
- Embedder 350 may output information 360 with an audio signal embedded therein.
- information 360 may comprise a digital audio file, such as tag 132 contained in a digital audio file, as shown in FIG. 2 above.
- Information 360 may be stored to a nonvolatile memory, such as a CD-ROM.
- FIG. 4 is a flow diagram illustrating method 400 for embedding audio descriptions according to the present invention.
- block B 410 information about an audio program is received. The information may be encoded in a format when received.
- block B 420 the information is transformed into an audio signal that conveys an audio description relating to an audio program.
- block B 430 the audio description produced by block B 420 is embedded along with an audio program in a predetermined format.
- FIG. 5 is a flow diagram illustrating method 500 for navigating through audio programs according to the present invention.
- an audio signal is played on a playback device.
- the audio signal includes an audio program and an audio description embedded within the audio signal.
- a user may navigate through audio programs based on the audio description, which may be played at predetermined times, such as when the user issues navigation commands to search through various selections on a medium. Such commands may include “play,” “skip,” and “repeat.”
- Method 500 may provide an “eyes-free” method for navigating through audio programs. Indeed, a listener may listen to an audio description to acquire information about selections on a medium. Thus, a playback device need not include an LCD panel and associated circuitry. In some embodiments, a playback device used in conjunction with method 500 may include an LCD panel, and the playback of audio descriptions may be selectively activated or deactivated by a user. In other embodiments, Braille-equipped navigation buttons may be included on the playback device. In other implementations, an audio playback device may incorporate speech processing technologies that process and act upon navigation commands spoken by a user.
- the invention may be implemented in part or in whole as a hard-wired circuit, as a circuit configuration fabricated into an application-specific integrated circuit, or as a firmware program loaded into non-volatile storage or a software program loaded from or into a data storage medium as machine-readable code, such code being instructions executable by an array of logic elements such as a microprocessor or other digital signal processing unit.
- transformer 320 , compressor 370 , and embedder 350 may be implemented individually and interconnected, or implemented as one multifunctional module.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
Abstract
Description
- 1. Field of the Invention
- This invention relates in general to the field of multimedia. Specifically, this invention relates to methods and systems for delivering multimedia information.
- 2. General Background and Related Art
- Digital audio is increasingly being delivered to listeners in compressed form. Sound compression formats, such as MPEG Layer 3 (MP3), Motion Picture Experts Group, ISO/IEC 11172-3 and ISO/IEC 13818-3, substantially compress audio data, yet maintain high sound quality. Tags have been developed to convey information that may relate to audio files. For instance, an ID3 tag may reside at the end of an MP3 file.
- Within a tag conforming to the ID3v2.3.0 informal standard (see http://id3.org), a host of textual information may be included, such as artist, album, song, title, lyrics, and comments. The ID3 tag in both ID3v2.3.0 and later revisions of the format also provides for a “general encapsulated object,” wherein any type of file may be encapsulated.
- When they download compressed music or other such audio for playback on computers, personal digital music players, CD players, or the like, users often like to search for selections among downloaded music selections. Textual information associated with each selection, and derived from a tag, may be displayed on an LCD panel. As such, users must read the information in order to decide which selection they wish to hear. This process is not user-friendly, and it prevents visually-impaired individuals from enjoying the search features of the above devices.
- Therefore, what is needed is a method and system that conveys information about audio programs in audio form.
- FIG. 1 illustrates a music system involving digital audio files.
- FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention.
- FIG. 3 is a high-level block diagram of a system according to an embodiment of the present invention.
- FIG. 4 is a flow diagram illustrating a method according to an embodiment of the present invention.
- FIG. 5 is a flow diagram illustrating a method according to an embodiment of the present invention.
- The following detailed description refers to the accompanying drawings that illustrate exemplary embodiments of the present inventions. Other embodiments are possible and modifications may be made to the embodiments without departing from the spirit and scope of the invention. Therefore, the following detailed description is not meant to limit the invention. Rather, the scope of the invention is defined by the appended claims.
- It will be apparent to one of ordinary skill in the art that the embodiments as described below may be implemented in many different embodiments of software, firmware, and hardware in the entities illustrated in the figures. The actual software code or specialized control hardware used to implement the present invention is not limiting of the present invention. Thus, the operation and behavior of the embodiments will be described without specific reference to the actual software code or specialized hardware components. The absence of such specific references is feasible because it is clearly understood that artisans of ordinary skill would be able to design software and control hardware to implement the embodiments of the present invention based on the description herein with only a reasonable effort and without undue experimentation.
- Moreover, the processes associated with the presented embodiments may be stored in any storage device, such as, for example, a computer system (non-volatile) memory, an optical disk, magnetic tape, or magnetic disk. Furthermore, the processes may be programmed when the computer system is manufactured or via a computer-readable medium at a later date. Such a medium may include any of the forms listed above with respect to storage devices and may further include, for example, a carrier wave modulated, or otherwise manipulated, to convey instructions that can be read, demodulated/decoded and executed by a computer.
- A method for embedding audio titles, as described herein, involves receiving information, encoded in a format, about an audio program. The information is transformed into an audio signal that conveys an audio description of the audio program. The audio description and the audio program are then embedded in a predetermined format.
- FIG. 1 illustrates a music system involving digital audio files. As shown,
system 100 includes adigital audio file 110, such as an MP3 file, recorded on amedium 163, anddigital audio player 120. -
Digital audio file 110 may have a predetermined format. In FIG. 1,digital audio file 110 comprises atag 130, such as an ID3 tag, andaudio data 140, which may include recorded sound, such as music or a narration of a book.Tag 130 may comprise metadata, or information aboutaudio data 140, withindigital audio file 110 itself. - More specifically,
tag 130 may comprise aheader 135 andframes 137.Header 135 may specify howtag 130 is structured.Frames 137 may comprise information blocks containing data. For instance, in FIG. 1,frames 135 may comprise data relating tolyrics 145,program information 160,comments 155, and general encapsulatedobject 170.Program information 160 may include information relating to theaudio data 140, such as title, artist, and album. Different frames and headers may be included intag 130 depending on the applicable format ofdigital audio file 110. -
Digital audio player 120 may comprise many types of digital audio playback devices. For instance,digital audio player 120 may comprise a CD player, a car stereo system, a handheld personal digital music player, a digital audio player on a personal computer, and the like.Digital audio player 120 plays digital audio selections recorded onmedium 163 for the listening enjoyment of a listener. It is to be noted that the present invention may also be applied in the domain of digital video, such as where a digital video file format contains metadata. -
Digital audio player 120 may include adisplay 150.Display 150 may list information relevant to a selection recorded onmedium 163. In FIG. 1,program information 160, namely title, artist, and album, has been read fromframes 137 oftag 130 such thatprogram information 160 fromtag 130 is processed and displayed ondisplay 150 ofdigital audio player 120. As such, a listener listening to, or searching among, various selections recorded onmedium 163 may be informed of relevant information.Display 150 may also include information relating to the track number of the selection onmedium 163, as well as the time elapsed for a selection that is being played or paused. - FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention. Before processing,
tag 131 includes generalencapsulated object 170. In the ID3 format, general encapsulatedobject 170 is a frame containing any kind of encoded data. It is contemplated that other digital audio file formats could be designed to contain such a frame. -
Tag 131 may be processed to producetag 132. Intag 132, generalencapsulated object 170 may include compressed audio data 220, shown in FIG. 2 by a binary numeric sequence. In an exemplary implementation of the present invention,program information 160, including title, artist, and album, are read fromtag 131, recorded to an audio file via a voice synthesizer, such as a text-to-speech (TTS) engine, and embedded within general encapsulatedobject 210 oftag 132 as compressed audio data 220. - Compressed audio data220 may comprise a narrative audio description which includes information besides
program information 160. In other embodiments, encoded text or other types of data may be embedded within general encapsulatedobject 210. For instance, paragraphs of explanatory text, biographical information relating to an artist, or compressed video data may be embedded. The size of a frame, such as general encapsulatedobject 210, may impose practical limitations on the quantity and type of data suitable for embedding within a digital audio file. - FIG. 3 is a high-level block diagram of
system 300 according to an embodiment of the present invention.System 300 comprisestransformer 320,compressor 370, andembedder 350.Transformer 320 may compriseTTS engine 330.Transformer 320 may receive encodedinformation 310, such astag 131 or an entire digital audio file. Via header information intag 131,transformer 320 may extract information from frames intag 131.Transformer 320 may also receiveexternal information 380, which is external to tag 131 and may include information from local or remote sources, such as databases accessible via the World Wide Web. -
Transformer 320 may combine information extracted fromtag 131 withexternal information 380. For instance, program information relating to a music selection intag 131 may be combined with words and phrases to produce a narrative sentence, such as “Song 1 is The Star-Spangled Banner by Francis Scott Key.” In some embodiments, user preferences, system defaults, or a combination thereof may specify which information intag 131, and whichexternal information 380, is included in the narrative.TTS engine 330 may process the combined textual data, translating it into speech signals. The speech signals may be recorded as anaudio signal 340 and outputted bytransformer 320. -
Compressor 370 receivesaudio signal 340 fromtransformer 320.Compressor 370 compressesaudio signal 340 into compressedaudio data 345. For instance, if encodedinformation 310 is received in MP3 format,compressor 370 may compressaudio signal 340 such thatcompressed audio data 345 is appropriately compressed for embedding in an MP3 file. It is to be appreciated thatcompressor 370 may be incorporated intoembedder 350. -
Embedder 350 receives compressedaudio data 345 fromcompressor 370.Embedder 350 embeds compressedaudio data 345 within general encapsulatedobject 210 of a tag.Embedder 350 mayoutput information 360 with an audio signal embedded therein. For instance,information 360 may comprise a digital audio file, such astag 132 contained in a digital audio file, as shown in FIG. 2 above.Information 360 may be stored to a nonvolatile memory, such as a CD-ROM. - FIG. 4 is a flow
diagram illustrating method 400 for embedding audio descriptions according to the present invention. In block B410, information about an audio program is received. The information may be encoded in a format when received. In block B420, the information is transformed into an audio signal that conveys an audio description relating to an audio program. In block B430, the audio description produced by block B420 is embedded along with an audio program in a predetermined format. - FIG. 5 is a flow
diagram illustrating method 500 for navigating through audio programs according to the present invention. In block B510, an audio signal is played on a playback device. The audio signal includes an audio program and an audio description embedded within the audio signal. In block B520, a user may navigate through audio programs based on the audio description, which may be played at predetermined times, such as when the user issues navigation commands to search through various selections on a medium. Such commands may include “play,” “skip,” and “repeat.” -
Method 500 may provide an “eyes-free” method for navigating through audio programs. Indeed, a listener may listen to an audio description to acquire information about selections on a medium. Thus, a playback device need not include an LCD panel and associated circuitry. In some embodiments, a playback device used in conjunction withmethod 500 may include an LCD panel, and the playback of audio descriptions may be selectively activated or deactivated by a user. In other embodiments, Braille-equipped navigation buttons may be included on the playback device. In other implementations, an audio playback device may incorporate speech processing technologies that process and act upon navigation commands spoken by a user. - The foregoing description of the preferred embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments are possible, and the generic principles presented herein may be applied to other embodiments as well. For example, descriptive material may be embedded in digital audio and video files, wherein the material is applicable to all or some selections on a medium. Similarly, timely announcements of advertisements, promotions, concert tours, television specials, and other such external information that may relate to the selections may be embedded within the files. The embedded information may be downloaded from the World Wide Web.
- Further, the invention may be implemented in part or in whole as a hard-wired circuit, as a circuit configuration fabricated into an application-specific integrated circuit, or as a firmware program loaded into non-volatile storage or a software program loaded from or into a data storage medium as machine-readable code, such code being instructions executable by an array of logic elements such as a microprocessor or other digital signal processing unit. Specifically,
transformer 320,compressor 370, andembedder 350 may be implemented individually and interconnected, or implemented as one multifunctional module. - As such, the present invention is not intended to be limited to the embodiments shown above but rather is to be accorded the widest scope consistent with the principles and novel features disclosed in any fashion herein.
Claims (30)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/809,019 US6448485B1 (en) | 2001-03-16 | 2001-03-16 | Method and system for embedding audio titles |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/809,019 US6448485B1 (en) | 2001-03-16 | 2001-03-16 | Method and system for embedding audio titles |
Publications (2)
Publication Number | Publication Date |
---|---|
US6448485B1 US6448485B1 (en) | 2002-09-10 |
US20020129692A1 true US20020129692A1 (en) | 2002-09-19 |
Family
ID=25200353
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/809,019 Expired - Lifetime US6448485B1 (en) | 2001-03-16 | 2001-03-16 | Method and system for embedding audio titles |
Country Status (1)
Country | Link |
---|---|
US (1) | US6448485B1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040125144A1 (en) * | 2002-12-30 | 2004-07-01 | Samsung Electronics Co., Ltd. | Integrated management method and system for multimedia contents |
US20040215461A1 (en) * | 2003-04-24 | 2004-10-28 | Visteon Global Technologies, Inc. | Text-to-speech system for generating information announcements |
US20050137869A1 (en) * | 2003-12-17 | 2005-06-23 | Samsung Electronics Co., Ltd. | Method supporting text-to-speech navigation and multimedia device using the same |
US20070005565A1 (en) * | 2005-07-04 | 2007-01-04 | Samsung Electronics., Ltd. | Database searching method and apparatus |
US20070100904A1 (en) * | 2005-10-31 | 2007-05-03 | Qwest Communications International Inc. | Creation and transmission of rich content media |
US20130300845A1 (en) * | 2008-01-25 | 2013-11-14 | At&T Intellectual Property I, L.P. | System and Method for Digital Video Retrieval Involving Speech Recognition |
Families Citing this family (137)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6757913B2 (en) * | 1996-07-15 | 2004-06-29 | Gregory D. Knox | Wireless music and data transceiver system |
US8645137B2 (en) | 2000-03-16 | 2014-02-04 | Apple Inc. | Fast, language-independent method for user authentication by voice |
US7426513B2 (en) * | 2000-10-12 | 2008-09-16 | Sap Portals Israel Ltd. | Client-based objectifying of text pages |
AU2002214613A1 (en) * | 2000-11-08 | 2002-05-21 | Digimarc Corporation | Content authentication and recovery using digital watermarks |
JP2002334086A (en) * | 2001-05-10 | 2002-11-22 | Sony Corp | Information processor, its method, recording medium, and program |
US6683241B2 (en) | 2001-11-06 | 2004-01-27 | James W. Wieder | Pseudo-live music audio and sound |
US7732697B1 (en) | 2001-11-06 | 2010-06-08 | Wieder James W | Creating music and sound that varies from playback to playback |
US8487176B1 (en) | 2001-11-06 | 2013-07-16 | James W. Wieder | Music and sound that varies from one playback to another playback |
US20030158737A1 (en) * | 2002-02-15 | 2003-08-21 | Csicsatka Tibor George | Method and apparatus for incorporating additional audio information into audio data file identifying information |
US20040186713A1 (en) * | 2003-03-06 | 2004-09-23 | Gomas Steven W. | Content delivery and speech system and apparatus for the blind and print-handicapped |
KR20050107610A (en) * | 2003-03-10 | 2005-11-14 | 코닌클리케 필립스 일렉트로닉스 엔.브이. | Record carrier and apparatus enabling seamless playback |
WO2005003901A2 (en) * | 2003-06-26 | 2005-01-13 | Leave A Little Room Foundation | Compressed media files with intrinsic supplementary content |
US7757173B2 (en) * | 2003-07-18 | 2010-07-13 | Apple Inc. | Voice menu system |
US7213036B2 (en) | 2003-08-12 | 2007-05-01 | Aol Llc | System for incorporating information about a source and usage of a media asset into the asset itself |
US20060160486A1 (en) * | 2005-01-14 | 2006-07-20 | Xm Satellite Radio, Inc. | Method and system for converting streaming digital data to FM modulated data |
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US7825319B2 (en) | 2005-10-06 | 2010-11-02 | Pacing Technologies Llc | System and method for pacing repetitive motion activities |
US20110072955A1 (en) | 2005-10-06 | 2011-03-31 | Turner William D | System and method for pacing repetitive motion activities |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US20080194175A1 (en) * | 2007-02-09 | 2008-08-14 | Intellitoys Llc | Interactive toy providing, dynamic, navigable media content |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US8996376B2 (en) | 2008-04-05 | 2015-03-31 | Apple Inc. | Intelligent text-to-speech conversion |
US10496753B2 (en) | 2010-01-18 | 2019-12-03 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8898568B2 (en) | 2008-09-09 | 2014-11-25 | Apple Inc. | Audio user interface |
JP5428294B2 (en) * | 2008-10-31 | 2014-02-26 | ブラザー工業株式会社 | Exercise content generation system, output terminal, exercise content generation device, content generation method, and content generation program |
US9959870B2 (en) | 2008-12-11 | 2018-05-01 | Apple Inc. | Speech recognition involving a mobile device |
US20120309363A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Triggering notifications associated with tasks items that represent tasks to perform |
US10241752B2 (en) | 2011-09-30 | 2019-03-26 | Apple Inc. | Interface for a virtual digital assistant |
US10241644B2 (en) | 2011-06-03 | 2019-03-26 | Apple Inc. | Actionable reminder entries |
US9858925B2 (en) | 2009-06-05 | 2018-01-02 | Apple Inc. | Using context information to facilitate processing of commands in a virtual assistant |
US9431006B2 (en) | 2009-07-02 | 2016-08-30 | Apple Inc. | Methods and apparatuses for automatic speech recognition |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US10553209B2 (en) | 2010-01-18 | 2020-02-04 | Apple Inc. | Systems and methods for hands-free notification summaries |
US10679605B2 (en) | 2010-01-18 | 2020-06-09 | Apple Inc. | Hands-free list-reading by intelligent automated assistant |
US10705794B2 (en) | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US8977584B2 (en) | 2010-01-25 | 2015-03-10 | Newvaluexchange Global Ai Llp | Apparatuses, methods and systems for a digital conversation management platform |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US10762293B2 (en) | 2010-12-22 | 2020-09-01 | Apple Inc. | Using parts-of-speech tagging and named entity recognition for spelling correction |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US8994660B2 (en) | 2011-08-29 | 2015-03-31 | Apple Inc. | Text correction processing |
JP2013178509A (en) * | 2012-02-07 | 2013-09-09 | Yamaha Corp | Electronic equipment and voice guide program |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US9483461B2 (en) | 2012-03-06 | 2016-11-01 | Apple Inc. | Handling speech synthesis of content for multiple languages |
US9280610B2 (en) | 2012-05-14 | 2016-03-08 | Apple Inc. | Crowd sourcing information to fulfill user requests |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
US9495129B2 (en) | 2012-06-29 | 2016-11-15 | Apple Inc. | Device, method, and user interface for voice-activated navigation and browsing of a document |
US9576574B2 (en) | 2012-09-10 | 2017-02-21 | Apple Inc. | Context-sensitive handling of interruptions by intelligent digital assistant |
US9547647B2 (en) | 2012-09-19 | 2017-01-17 | Apple Inc. | Voice-based media searching |
AU2014214676A1 (en) | 2013-02-07 | 2015-08-27 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US9368114B2 (en) | 2013-03-14 | 2016-06-14 | Apple Inc. | Context-sensitive handling of interruptions |
US9922642B2 (en) | 2013-03-15 | 2018-03-20 | Apple Inc. | Training an at least partial voice command system |
WO2014144579A1 (en) | 2013-03-15 | 2014-09-18 | Apple Inc. | System and method for updating an adaptive speech recognition model |
US9424348B1 (en) | 2013-05-08 | 2016-08-23 | Rock My World, Inc. | Sensor-driven audio playback modification |
WO2014197336A1 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for detecting errors in interactions with a voice-based digital assistant |
WO2014197334A2 (en) | 2013-06-07 | 2014-12-11 | Apple Inc. | System and method for user-specified pronunciation of words for speech synthesis and recognition |
US9582608B2 (en) | 2013-06-07 | 2017-02-28 | Apple Inc. | Unified ranking with entropy-weighted information for phrase-based semantic auto-completion |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
EP3937002A1 (en) | 2013-06-09 | 2022-01-12 | Apple Inc. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101809808B1 (en) | 2013-06-13 | 2017-12-15 | 애플 인크. | System and method for emergency calls initiated by voice command |
US10791216B2 (en) | 2013-08-06 | 2020-09-29 | Apple Inc. | Auto-activating smart responses based on activities from remote devices |
US9620105B2 (en) | 2014-05-15 | 2017-04-11 | Apple Inc. | Analyzing audio input for efficient speech and music recognition |
US10592095B2 (en) | 2014-05-23 | 2020-03-17 | Apple Inc. | Instantaneous speaking of content on touch devices |
US9502031B2 (en) | 2014-05-27 | 2016-11-22 | Apple Inc. | Method for supporting dynamic grammars in WFST-based ASR |
US10289433B2 (en) | 2014-05-30 | 2019-05-14 | Apple Inc. | Domain specific language for encoding assistant dialog |
US9734193B2 (en) | 2014-05-30 | 2017-08-15 | Apple Inc. | Determining domain salience ranking from ambiguous words in natural speech |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
TWI566107B (en) | 2014-05-30 | 2017-01-11 | 蘋果公司 | Method for processing a multi-part voice command, non-transitory computer readable storage medium and electronic device |
US10078631B2 (en) | 2014-05-30 | 2018-09-18 | Apple Inc. | Entropy-guided text prediction using combined word and character n-gram language models |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9785630B2 (en) | 2014-05-30 | 2017-10-10 | Apple Inc. | Text prediction using combined word N-gram and unigram language models |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9760559B2 (en) | 2014-05-30 | 2017-09-12 | Apple Inc. | Predictive text input |
US9842101B2 (en) | 2014-05-30 | 2017-12-12 | Apple Inc. | Predictive conversion of language input |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US10659851B2 (en) | 2014-06-30 | 2020-05-19 | Apple Inc. | Real-time digital assistant knowledge updates |
US10446141B2 (en) | 2014-08-28 | 2019-10-15 | Apple Inc. | Automatic speech recognition based on user feedback |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10789041B2 (en) | 2014-09-12 | 2020-09-29 | Apple Inc. | Dynamic thresholds for always listening speech trigger |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9646609B2 (en) | 2014-09-30 | 2017-05-09 | Apple Inc. | Caching apparatus for serving phonetic pronunciations |
US9886432B2 (en) | 2014-09-30 | 2018-02-06 | Apple Inc. | Parsimonious handling of word inflection via categorical stem + suffix N-gram language models |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10552013B2 (en) | 2014-12-02 | 2020-02-04 | Apple Inc. | Data detection |
US9711141B2 (en) | 2014-12-09 | 2017-07-18 | Apple Inc. | Disambiguating heteronyms in speech synthesis |
US9865280B2 (en) | 2015-03-06 | 2018-01-09 | Apple Inc. | Structured dictation using intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9899019B2 (en) | 2015-03-18 | 2018-02-20 | Apple Inc. | Systems and methods for structured stem and suffix language models |
US9842105B2 (en) | 2015-04-16 | 2017-12-12 | Apple Inc. | Parsimonious continuous-space phrase representations for natural language processing |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10127220B2 (en) | 2015-06-04 | 2018-11-13 | Apple Inc. | Language identification from short strings |
US10101822B2 (en) | 2015-06-05 | 2018-10-16 | Apple Inc. | Language input correction |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US10186254B2 (en) | 2015-06-07 | 2019-01-22 | Apple Inc. | Context-based endpoint detection |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US10255907B2 (en) | 2015-06-07 | 2019-04-09 | Apple Inc. | Automatic accent detection using acoustic models |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US9697820B2 (en) | 2015-09-24 | 2017-07-04 | Apple Inc. | Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks |
US11010550B2 (en) | 2015-09-29 | 2021-05-18 | Apple Inc. | Unified language modeling framework for word prediction, auto-completion and auto-correction |
US10366158B2 (en) | 2015-09-29 | 2019-07-30 | Apple Inc. | Efficient word encoding for recurrent neural network language models |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US10446143B2 (en) | 2016-03-14 | 2019-10-15 | Apple Inc. | Identification of voice inputs providing credentials |
US10114607B1 (en) | 2016-03-31 | 2018-10-30 | Rock My World, Inc. | Physiological state-driven playback tempo modification |
US9934775B2 (en) | 2016-05-26 | 2018-04-03 | Apple Inc. | Unit-selection text-to-speech synthesis based on predicted concatenation parameters |
US9972304B2 (en) | 2016-06-03 | 2018-05-15 | Apple Inc. | Privacy preserving distributed evaluation framework for embedded personalized systems |
US10249300B2 (en) | 2016-06-06 | 2019-04-02 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
DK179588B1 (en) | 2016-06-09 | 2019-02-22 | Apple Inc. | Intelligent automated assistant in a home environment |
US10067938B2 (en) | 2016-06-10 | 2018-09-04 | Apple Inc. | Multilingual word prediction |
US10509862B2 (en) | 2016-06-10 | 2019-12-17 | Apple Inc. | Dynamic phrase expansion of language input |
US10192552B2 (en) | 2016-06-10 | 2019-01-29 | Apple Inc. | Digital assistant providing whispered speech |
US10490187B2 (en) | 2016-06-10 | 2019-11-26 | Apple Inc. | Digital assistant providing automated status report |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
DK179343B1 (en) | 2016-06-11 | 2018-05-14 | Apple Inc | Intelligent task discovery |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179049B1 (en) | 2016-06-11 | 2017-09-18 | Apple Inc | Data driven natural language event detection and classification |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US10593346B2 (en) | 2016-12-22 | 2020-03-17 | Apple Inc. | Rank-reduced token representation for automatic speech recognition |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4731847A (en) * | 1982-04-26 | 1988-03-15 | Texas Instruments Incorporated | Electronic apparatus for simulating singing of song |
DE69434787T2 (en) * | 1993-05-26 | 2007-08-30 | Pioneer Electronic Corp. | Karaoke playback device |
GB2300087B (en) * | 1995-04-18 | 1999-04-28 | Ibm | Voice processing system and method |
EP0915470A3 (en) * | 1995-08-21 | 2004-03-24 | Matsushita Electric Industrial Co., Ltd. | Multimedia optical disk, reproduction apparatus and method for achieving variable scene development based on interactive control |
US6212359B1 (en) * | 1996-07-15 | 2001-04-03 | Gregory D. Knox | Wireless Transceiver System For Digital Music |
US5918303A (en) * | 1996-11-25 | 1999-06-29 | Yamaha Corporation | Performance setting data selecting apparatus |
US5739451A (en) * | 1996-12-27 | 1998-04-14 | Franklin Electronic Publishers, Incorporated | Hand held electronic music encyclopedia with text and note structure search |
US6169543B1 (en) * | 1998-12-28 | 2001-01-02 | Thomson Licensing S.A. | System and method for customizing program guide information to include reminder item or local identifier |
-
2001
- 2001-03-16 US US09/809,019 patent/US6448485B1/en not_active Expired - Lifetime
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040125144A1 (en) * | 2002-12-30 | 2004-07-01 | Samsung Electronics Co., Ltd. | Integrated management method and system for multimedia contents |
US20040215461A1 (en) * | 2003-04-24 | 2004-10-28 | Visteon Global Technologies, Inc. | Text-to-speech system for generating information announcements |
US20050137869A1 (en) * | 2003-12-17 | 2005-06-23 | Samsung Electronics Co., Ltd. | Method supporting text-to-speech navigation and multimedia device using the same |
EP1548704A1 (en) * | 2003-12-17 | 2005-06-29 | Samsung Electronics Co., Ltd. | Method supporting text-to-speech navigation and multimedia device using the same |
US20070005565A1 (en) * | 2005-07-04 | 2007-01-04 | Samsung Electronics., Ltd. | Database searching method and apparatus |
US20070100904A1 (en) * | 2005-10-31 | 2007-05-03 | Qwest Communications International Inc. | Creation and transmission of rich content media |
US8856118B2 (en) * | 2005-10-31 | 2014-10-07 | Qwest Communications International Inc. | Creation and transmission of rich content media |
US20130300845A1 (en) * | 2008-01-25 | 2013-11-14 | At&T Intellectual Property I, L.P. | System and Method for Digital Video Retrieval Involving Speech Recognition |
US9135336B2 (en) * | 2008-01-25 | 2015-09-15 | At&T Intellectual Property I, L.P. | System and method for digital video retrieval involving speech recognition |
US9465870B2 (en) * | 2008-01-25 | 2016-10-11 | At&T Intellectual Property I, L.P. | System and method for digital video retrieval involving speech recognition |
Also Published As
Publication number | Publication date |
---|---|
US6448485B1 (en) | 2002-09-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6448485B1 (en) | Method and system for embedding audio titles | |
US9875735B2 (en) | System and method for synthetically generated speech describing media content | |
EP2491560B1 (en) | Metadata time marking information for indicating a section of an audio object | |
US9153233B2 (en) | Voice-controlled selection of media files utilizing phonetic data | |
JP4965565B2 (en) | Playlist structure for large playlists | |
US20030158737A1 (en) | Method and apparatus for incorporating additional audio information into audio data file identifying information | |
JP2009534696A (en) | Apparatus and method for encoding and decoding a plurality of digital data sets | |
JP2007200495A (en) | Music reproduction apparatus, music reproduction method and music reproduction program | |
US6525251B1 (en) | Method of displaying the title and words of songs on a digital music player | |
CN1332365C (en) | Method and device for sync controlling voice frequency and text information | |
US20040064208A1 (en) | Dual mode store and play digital audio broadcast (DAB) receiver | |
US20060199161A1 (en) | Method of creating multi-lingual lyrics slides video show for sing along | |
KR101143908B1 (en) | Audio reproduction apparatus and method for providing disk jockey service | |
EP2343706A1 (en) | Reproducing device and reproducing method | |
JP4362775B2 (en) | Broadcast recording device | |
KR20010038854A (en) | Method and format of music file for providing text and score | |
US7765198B2 (en) | Data processing apparatus, data processing method, and data processing system | |
CN101031049A (en) | Method and apparatus for playing audio-fequency data and other content simultaneouslly | |
KR100469052B1 (en) | Audio system of vehicle | |
JP4312167B2 (en) | Content playback device | |
JP2007257712A (en) | Content selecting device | |
US20080195925A1 (en) | Compressed Media Files with Intrinsic Supplementary Content | |
JP2005071548A (en) | Audio reproducing device | |
KR20080086793A (en) | Mobile device capable of audio playback | |
JP2004220710A (en) | Audio-reproducing device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTEL CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BARILE, STEVEN E.;REEL/FRAME:011624/0864 Effective date: 20010309 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
REMI | Maintenance fee reminder mailed | ||
FPAY | Fee payment |
Year of fee payment: 8 |
|
SULP | Surcharge for late payment |
Year of fee payment: 7 |
|
FPAY | Fee payment |
Year of fee payment: 12 |