[go: up one dir, main page]

US20020129692A1 - Method and system for embedding audio titles - Google Patents

Method and system for embedding audio titles Download PDF

Info

Publication number
US20020129692A1
US20020129692A1 US09/809,019 US80901901A US2002129692A1 US 20020129692 A1 US20020129692 A1 US 20020129692A1 US 80901901 A US80901901 A US 80901901A US 2002129692 A1 US2002129692 A1 US 2002129692A1
Authority
US
United States
Prior art keywords
audio
information
program
format
description
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US09/809,019
Other versions
US6448485B1 (en
Inventor
Steven Barile
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US09/809,019 priority Critical patent/US6448485B1/en
Assigned to INTEL CORPORATION reassignment INTEL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: BARILE, STEVEN E.
Application granted granted Critical
Publication of US6448485B1 publication Critical patent/US6448485B1/en
Publication of US20020129692A1 publication Critical patent/US20020129692A1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/3027Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/061MP3, i.e. MPEG-1 or MPEG-2 Audio Layer III, lossy audio compression
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/091Info, i.e. juxtaposition of unrelated auxiliary information or commercial messages with or between music files
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2537Optical discs
    • G11B2220/2545CDs

Definitions

  • This invention relates in general to the field of multimedia. Specifically, this invention relates to methods and systems for delivering multimedia information.
  • Digital audio is increasingly being delivered to listeners in compressed form.
  • Sound compression formats such as MPEG Layer 3 (MP3), Motion Picture Experts Group, ISO/IEC 11172-3 and ISO/IEC 13818-3, substantially compress audio data, yet maintain high sound quality.
  • tags have been developed to convey information that may relate to audio files. For instance, an ID3 tag may reside at the end of an MP3 file.
  • ID3v2.3.0 Within a tag conforming to the ID3v2.3.0 informal standard (see http://id3.org), a host of textual information may be included, such as artist, album, song, title, lyrics, and comments.
  • the ID3 tag in both ID3v2.3.0 and later revisions of the format also provides for a “general encapsulated object,” wherein any type of file may be encapsulated.
  • FIG. 1 illustrates a music system involving digital audio files.
  • FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention.
  • FIG. 3 is a high-level block diagram of a system according to an embodiment of the present invention.
  • FIG. 4 is a flow diagram illustrating a method according to an embodiment of the present invention.
  • FIG. 5 is a flow diagram illustrating a method according to an embodiment of the present invention.
  • the processes associated with the presented embodiments may be stored in any storage device, such as, for example, a computer system (non-volatile) memory, an optical disk, magnetic tape, or magnetic disk.
  • a computer system non-volatile
  • the processes may be programmed when the computer system is manufactured or via a computer-readable medium at a later date.
  • Such a medium may include any of the forms listed above with respect to storage devices and may further include, for example, a carrier wave modulated, or otherwise manipulated, to convey instructions that can be read, demodulated/decoded and executed by a computer.
  • a method for embedding audio titles involves receiving information, encoded in a format, about an audio program.
  • the information is transformed into an audio signal that conveys an audio description of the audio program.
  • the audio description and the audio program are then embedded in a predetermined format.
  • FIG. 1 illustrates a music system involving digital audio files.
  • system 100 includes a digital audio file 110 , such as an MP3 file, recorded on a medium 163 , and digital audio player 120 .
  • Digital audio file 110 may have a predetermined format.
  • digital audio file 110 comprises a tag 130 , such as an ID3 tag, and audio data 140 , which may include recorded sound, such as music or a narration of a book.
  • Tag 130 may comprise metadata, or information about audio data 140 , within digital audio file 110 itself.
  • tag 130 may comprise a header 135 and frames 137 .
  • Header 135 may specify how tag 130 is structured.
  • Frames 137 may comprise information blocks containing data.
  • frames 135 may comprise data relating to lyrics 145 , program information 160 , comments 155 , and general encapsulated object 170 .
  • Program information 160 may include information relating to the audio data 140 , such as title, artist, and album. Different frames and headers may be included in tag 130 depending on the applicable format of digital audio file 110 .
  • Digital audio player 120 may comprise many types of digital audio playback devices.
  • digital audio player 120 may comprise a CD player, a car stereo system, a handheld personal digital music player, a digital audio player on a personal computer, and the like.
  • Digital audio player 120 plays digital audio selections recorded on medium 163 for the listening enjoyment of a listener. It is to be noted that the present invention may also be applied in the domain of digital video, such as where a digital video file format contains metadata.
  • Digital audio player 120 may include a display 150 .
  • Display 150 may list information relevant to a selection recorded on medium 163 .
  • program information 160 namely title, artist, and album, has been read from frames 137 of tag 130 such that program information 160 from tag 130 is processed and displayed on display 150 of digital audio player 120 .
  • Display 150 may also include information relating to the track number of the selection on medium 163 , as well as the time elapsed for a selection that is being played or paused.
  • FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention.
  • tag 131 Before processing, tag 131 includes general encapsulated object 170 .
  • general encapsulated object 170 is a frame containing any kind of encoded data. It is contemplated that other digital audio file formats could be designed to contain such a frame.
  • Tag 131 may be processed to produce tag 132 .
  • general encapsulated object 170 may include compressed audio data 220 , shown in FIG. 2 by a binary numeric sequence.
  • program information 160 including title, artist, and album, are read from tag 131 , recorded to an audio file via a voice synthesizer, such as a text-to-speech (TTS) engine, and embedded within general encapsulated object 210 of tag 132 as compressed audio data 220 .
  • TTS text-to-speech
  • Compressed audio data 220 may comprise a narrative audio description which includes information besides program information 160 .
  • encoded text or other types of data may be embedded within general encapsulated object 210 .
  • paragraphs of explanatory text, biographical information relating to an artist, or compressed video data may be embedded.
  • the size of a frame, such as general encapsulated object 210 may impose practical limitations on the quantity and type of data suitable for embedding within a digital audio file.
  • FIG. 3 is a high-level block diagram of system 300 according to an embodiment of the present invention.
  • System 300 comprises transformer 320 , compressor 370 , and embedder 350 .
  • Transformer 320 may comprise TTS engine 330 .
  • Transformer 320 may receive encoded information 310 , such as tag 131 or an entire digital audio file. Via header information in tag 131 , transformer 320 may extract information from frames in tag 131 .
  • Transformer 320 may also receive external information 380 , which is external to tag 131 and may include information from local or remote sources, such as databases accessible via the World Wide Web.
  • Transformer 320 may combine information extracted from tag 131 with external information 380 .
  • program information relating to a music selection in tag 131 may be combined with words and phrases to produce a narrative sentence, such as “Song 1 is The Star-Spangled Banner by Francis Scott Key.”
  • user preferences, system defaults, or a combination thereof may specify which information in tag 131 , and which external information 380 , is included in the narrative.
  • TTS engine 330 may process the combined textual data, translating it into speech signals. The speech signals may be recorded as an audio signal 340 and outputted by transformer 320 .
  • Compressor 370 receives audio signal 340 from transformer 320 .
  • Compressor 370 compresses audio signal 340 into compressed audio data 345 .
  • compressor 370 may compress audio signal 340 such that compressed audio data 345 is appropriately compressed for embedding in an MP3 file. It is to be appreciated that compressor 370 may be incorporated into embedder 350 .
  • Embedder 350 receives compressed audio data 345 from compressor 370 .
  • Embedder 350 embeds compressed audio data 345 within general encapsulated object 210 of a tag.
  • Embedder 350 may output information 360 with an audio signal embedded therein.
  • information 360 may comprise a digital audio file, such as tag 132 contained in a digital audio file, as shown in FIG. 2 above.
  • Information 360 may be stored to a nonvolatile memory, such as a CD-ROM.
  • FIG. 4 is a flow diagram illustrating method 400 for embedding audio descriptions according to the present invention.
  • block B 410 information about an audio program is received. The information may be encoded in a format when received.
  • block B 420 the information is transformed into an audio signal that conveys an audio description relating to an audio program.
  • block B 430 the audio description produced by block B 420 is embedded along with an audio program in a predetermined format.
  • FIG. 5 is a flow diagram illustrating method 500 for navigating through audio programs according to the present invention.
  • an audio signal is played on a playback device.
  • the audio signal includes an audio program and an audio description embedded within the audio signal.
  • a user may navigate through audio programs based on the audio description, which may be played at predetermined times, such as when the user issues navigation commands to search through various selections on a medium. Such commands may include “play,” “skip,” and “repeat.”
  • Method 500 may provide an “eyes-free” method for navigating through audio programs. Indeed, a listener may listen to an audio description to acquire information about selections on a medium. Thus, a playback device need not include an LCD panel and associated circuitry. In some embodiments, a playback device used in conjunction with method 500 may include an LCD panel, and the playback of audio descriptions may be selectively activated or deactivated by a user. In other embodiments, Braille-equipped navigation buttons may be included on the playback device. In other implementations, an audio playback device may incorporate speech processing technologies that process and act upon navigation commands spoken by a user.
  • the invention may be implemented in part or in whole as a hard-wired circuit, as a circuit configuration fabricated into an application-specific integrated circuit, or as a firmware program loaded into non-volatile storage or a software program loaded from or into a data storage medium as machine-readable code, such code being instructions executable by an array of logic elements such as a microprocessor or other digital signal processing unit.
  • transformer 320 , compressor 370 , and embedder 350 may be implemented individually and interconnected, or implemented as one multifunctional module.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing For Digital Recording And Reproducing (AREA)

Abstract

A system and method for embedding audio titles is presented. Information encoded in a format is received about an audio program. The information is transformed into an audio signal conveying an audio description about the audio program. The audio description and the audio program are then embedded in a predetermined format.

Description

    BACKGROUND
  • 1. Field of the Invention [0001]
  • This invention relates in general to the field of multimedia. Specifically, this invention relates to methods and systems for delivering multimedia information. [0002]
  • 2. General Background and Related Art [0003]
  • Digital audio is increasingly being delivered to listeners in compressed form. Sound compression formats, such as MPEG Layer 3 (MP3), Motion Picture Experts Group, ISO/IEC 11172-3 and ISO/IEC 13818-3, substantially compress audio data, yet maintain high sound quality. Tags have been developed to convey information that may relate to audio files. For instance, an ID3 tag may reside at the end of an MP3 file. [0004]
  • Within a tag conforming to the ID3v2.3.0 informal standard (see http://id3.org), a host of textual information may be included, such as artist, album, song, title, lyrics, and comments. The ID3 tag in both ID3v2.3.0 and later revisions of the format also provides for a “general encapsulated object,” wherein any type of file may be encapsulated. [0005]
  • When they download compressed music or other such audio for playback on computers, personal digital music players, CD players, or the like, users often like to search for selections among downloaded music selections. Textual information associated with each selection, and derived from a tag, may be displayed on an LCD panel. As such, users must read the information in order to decide which selection they wish to hear. This process is not user-friendly, and it prevents visually-impaired individuals from enjoying the search features of the above devices. [0006]
  • Therefore, what is needed is a method and system that conveys information about audio programs in audio form.[0007]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a music system involving digital audio files. [0008]
  • FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention. [0009]
  • FIG. 3 is a high-level block diagram of a system according to an embodiment of the present invention. [0010]
  • FIG. 4 is a flow diagram illustrating a method according to an embodiment of the present invention. [0011]
  • FIG. 5 is a flow diagram illustrating a method according to an embodiment of the present invention.[0012]
  • DETAILED DESCRIPTION
  • The following detailed description refers to the accompanying drawings that illustrate exemplary embodiments of the present inventions. Other embodiments are possible and modifications may be made to the embodiments without departing from the spirit and scope of the invention. Therefore, the following detailed description is not meant to limit the invention. Rather, the scope of the invention is defined by the appended claims. [0013]
  • It will be apparent to one of ordinary skill in the art that the embodiments as described below may be implemented in many different embodiments of software, firmware, and hardware in the entities illustrated in the figures. The actual software code or specialized control hardware used to implement the present invention is not limiting of the present invention. Thus, the operation and behavior of the embodiments will be described without specific reference to the actual software code or specialized hardware components. The absence of such specific references is feasible because it is clearly understood that artisans of ordinary skill would be able to design software and control hardware to implement the embodiments of the present invention based on the description herein with only a reasonable effort and without undue experimentation. [0014]
  • Moreover, the processes associated with the presented embodiments may be stored in any storage device, such as, for example, a computer system (non-volatile) memory, an optical disk, magnetic tape, or magnetic disk. Furthermore, the processes may be programmed when the computer system is manufactured or via a computer-readable medium at a later date. Such a medium may include any of the forms listed above with respect to storage devices and may further include, for example, a carrier wave modulated, or otherwise manipulated, to convey instructions that can be read, demodulated/decoded and executed by a computer. [0015]
  • A method for embedding audio titles, as described herein, involves receiving information, encoded in a format, about an audio program. The information is transformed into an audio signal that conveys an audio description of the audio program. The audio description and the audio program are then embedded in a predetermined format. [0016]
  • FIG. 1 illustrates a music system involving digital audio files. As shown, [0017] system 100 includes a digital audio file 110, such as an MP3 file, recorded on a medium 163, and digital audio player 120.
  • [0018] Digital audio file 110 may have a predetermined format. In FIG. 1, digital audio file 110 comprises a tag 130, such as an ID3 tag, and audio data 140, which may include recorded sound, such as music or a narration of a book. Tag 130 may comprise metadata, or information about audio data 140, within digital audio file 110 itself.
  • More specifically, [0019] tag 130 may comprise a header 135 and frames 137. Header 135 may specify how tag 130 is structured. Frames 137 may comprise information blocks containing data. For instance, in FIG. 1, frames 135 may comprise data relating to lyrics 145, program information 160, comments 155, and general encapsulated object 170. Program information 160 may include information relating to the audio data 140, such as title, artist, and album. Different frames and headers may be included in tag 130 depending on the applicable format of digital audio file 110.
  • [0020] Digital audio player 120 may comprise many types of digital audio playback devices. For instance, digital audio player 120 may comprise a CD player, a car stereo system, a handheld personal digital music player, a digital audio player on a personal computer, and the like. Digital audio player 120 plays digital audio selections recorded on medium 163 for the listening enjoyment of a listener. It is to be noted that the present invention may also be applied in the domain of digital video, such as where a digital video file format contains metadata.
  • [0021] Digital audio player 120 may include a display 150. Display 150 may list information relevant to a selection recorded on medium 163. In FIG. 1, program information 160, namely title, artist, and album, has been read from frames 137 of tag 130 such that program information 160 from tag 130 is processed and displayed on display 150 of digital audio player 120. As such, a listener listening to, or searching among, various selections recorded on medium 163 may be informed of relevant information. Display 150 may also include information relating to the track number of the selection on medium 163, as well as the time elapsed for a selection that is being played or paused.
  • FIG. 2 illustrates an ID3 tag before and after processing according to an embodiment of the present invention. Before processing, [0022] tag 131 includes general encapsulated object 170. In the ID3 format, general encapsulated object 170 is a frame containing any kind of encoded data. It is contemplated that other digital audio file formats could be designed to contain such a frame.
  • [0023] Tag 131 may be processed to produce tag 132. In tag 132, general encapsulated object 170 may include compressed audio data 220, shown in FIG. 2 by a binary numeric sequence. In an exemplary implementation of the present invention, program information 160, including title, artist, and album, are read from tag 131, recorded to an audio file via a voice synthesizer, such as a text-to-speech (TTS) engine, and embedded within general encapsulated object 210 of tag 132 as compressed audio data 220.
  • Compressed audio data [0024] 220 may comprise a narrative audio description which includes information besides program information 160. In other embodiments, encoded text or other types of data may be embedded within general encapsulated object 210. For instance, paragraphs of explanatory text, biographical information relating to an artist, or compressed video data may be embedded. The size of a frame, such as general encapsulated object 210, may impose practical limitations on the quantity and type of data suitable for embedding within a digital audio file.
  • FIG. 3 is a high-level block diagram of [0025] system 300 according to an embodiment of the present invention. System 300 comprises transformer 320, compressor 370, and embedder 350. Transformer 320 may comprise TTS engine 330. Transformer 320 may receive encoded information 310, such as tag 131 or an entire digital audio file. Via header information in tag 131, transformer 320 may extract information from frames in tag 131. Transformer 320 may also receive external information 380, which is external to tag 131 and may include information from local or remote sources, such as databases accessible via the World Wide Web.
  • [0026] Transformer 320 may combine information extracted from tag 131 with external information 380. For instance, program information relating to a music selection in tag 131 may be combined with words and phrases to produce a narrative sentence, such as “Song 1 is The Star-Spangled Banner by Francis Scott Key.” In some embodiments, user preferences, system defaults, or a combination thereof may specify which information in tag 131, and which external information 380, is included in the narrative. TTS engine 330 may process the combined textual data, translating it into speech signals. The speech signals may be recorded as an audio signal 340 and outputted by transformer 320.
  • [0027] Compressor 370 receives audio signal 340 from transformer 320. Compressor 370 compresses audio signal 340 into compressed audio data 345. For instance, if encoded information 310 is received in MP3 format, compressor 370 may compress audio signal 340 such that compressed audio data 345 is appropriately compressed for embedding in an MP3 file. It is to be appreciated that compressor 370 may be incorporated into embedder 350.
  • [0028] Embedder 350 receives compressed audio data 345 from compressor 370. Embedder 350 embeds compressed audio data 345 within general encapsulated object 210 of a tag. Embedder 350 may output information 360 with an audio signal embedded therein. For instance, information 360 may comprise a digital audio file, such as tag 132 contained in a digital audio file, as shown in FIG. 2 above. Information 360 may be stored to a nonvolatile memory, such as a CD-ROM.
  • FIG. 4 is a flow [0029] diagram illustrating method 400 for embedding audio descriptions according to the present invention. In block B410, information about an audio program is received. The information may be encoded in a format when received. In block B420, the information is transformed into an audio signal that conveys an audio description relating to an audio program. In block B430, the audio description produced by block B420 is embedded along with an audio program in a predetermined format.
  • FIG. 5 is a flow [0030] diagram illustrating method 500 for navigating through audio programs according to the present invention. In block B510, an audio signal is played on a playback device. The audio signal includes an audio program and an audio description embedded within the audio signal. In block B520, a user may navigate through audio programs based on the audio description, which may be played at predetermined times, such as when the user issues navigation commands to search through various selections on a medium. Such commands may include “play,” “skip,” and “repeat.”
  • [0031] Method 500 may provide an “eyes-free” method for navigating through audio programs. Indeed, a listener may listen to an audio description to acquire information about selections on a medium. Thus, a playback device need not include an LCD panel and associated circuitry. In some embodiments, a playback device used in conjunction with method 500 may include an LCD panel, and the playback of audio descriptions may be selectively activated or deactivated by a user. In other embodiments, Braille-equipped navigation buttons may be included on the playback device. In other implementations, an audio playback device may incorporate speech processing technologies that process and act upon navigation commands spoken by a user.
  • The foregoing description of the preferred embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments are possible, and the generic principles presented herein may be applied to other embodiments as well. For example, descriptive material may be embedded in digital audio and video files, wherein the material is applicable to all or some selections on a medium. Similarly, timely announcements of advertisements, promotions, concert tours, television specials, and other such external information that may relate to the selections may be embedded within the files. The embedded information may be downloaded from the World Wide Web. [0032]
  • Further, the invention may be implemented in part or in whole as a hard-wired circuit, as a circuit configuration fabricated into an application-specific integrated circuit, or as a firmware program loaded into non-volatile storage or a software program loaded from or into a data storage medium as machine-readable code, such code being instructions executable by an array of logic elements such as a microprocessor or other digital signal processing unit. Specifically, [0033] transformer 320, compressor 370, and embedder 350 may be implemented individually and interconnected, or implemented as one multifunctional module.
  • As such, the present invention is not intended to be limited to the embodiments shown above but rather is to be accorded the widest scope consistent with the principles and novel features disclosed in any fashion herein. [0034]

Claims (30)

What is claimed:
1. A method for embedding audio titles, said method comprising:
receiving information, encoded in a format, about an audio program;
transforming said information into an audio signal representing an audio description about said audio program; and
embedding said audio description, with said audio program, in a predetermined format.
2. The method according to claim 1, wherein said information about said audio program includes a title of said audio program.
3. The method according to claim 1, wherein said information about said audio program includes a description about said audio program.
4. The method according to claim 1, wherein said audio program includes a piece of music.
5. The method according to claim 1, wherein said audio description includes information downloaded from an information source, the information source having information different from said information about said audio program.
6. The method according to claim 1, wherein said format of said information includes a text format.
7. The method according to claim 6, wherein said transforming includes transforming said information in said text format into said audio signal using a text-to-speech engine.
8. The method according to claim 1, wherein said format of said information includes an audio format.
9. The method according to claim 1, wherein said predetermined format includes an MP3 ID3 tag.
10. The method of claim 1, further comprising storing said embedded audio description to a nonvolatile memory.
11. The method of claim 1, wherein said audio description includes compressed data.
12. A method for navigating through audio programs, said method comprising:
playing back, on an audio playback device, an audio signal comprising at least one embedded audio program and at least one audio description relating to said at least one audio program; and
navigating through said at least one audio program based on said at least one audio description.
13. The method according to claim 12, wherein said audio playback device includes a CD player.
14. The method according to claim 12, wherein said audio playback device includes a car stereo system.
15. The method according to claim 12, wherein said audio playback device includes a handheld device.
16. The method according to claim 12, wherein said audio playback device includes a personal digital music player.
17. The method according to claim 12, wherein said audio playback device includes an MP3 player on a personal computer.
18. The method according to claim 12, wherein said navigating is achieved by executing navigation commands.
19. The method according to claim 18, wherein said navigation commands include at least one of “play”, “skip”, and “repeat”.
20. The method according to claim 18, wherein said navigation commands are issued through voice commands.
21. A system for embedding audio titles, said system comprising:
a transformer configured to receive information, encoded in a format, about an audio program, said transformer being further configured to transform said information into an audio signal representing an audio description about said audio program; and
an embedder configured to embed said audio description, with said audio program, in a predetermined format.
22. The system according to claim 21, wherein said transformer includes a text-to-speech engine configured to transform said information in said text format into said audio signal.
23. A computer-readable medium encoded with a plurality of processor-executable instruction sequences for:
receiving information, encoded in a format, about an audio program;
transforming said information into an audio signal representing an audio description about said audio program; and
embedding said audio description, with said audio program, in a predetermined format.
24. The computer-readable medium according to claim 23, wherein said information about said audio program includes a description about said audio program.
25. The computer-readable medium according to claim 23, wherein said format of said information includes a text format.
26. The computer-readable medium according to claim 25, wherein said transforming includes transforming said information in said text format into said audio signal using a text-to-speech engine.
27. The computer-readable medium according to claim 23, wherein said predetermined format includes an MP3 ID3 tag.
28. A computer-readable medium encoded with a plurality of processor-executable instruction sequences for:
playing back, on an audio playback device, an audio signal comprising at least one embedded audio program and at least one audio description relating to said at least one audio program; and
navigating through said at least one audio program based on said at least one audio description.
29. The computer-readable medium according to claim 29, wherein said audio playback device includes a personal digital music player.
30. The computer-readable medium according to claim 29, wherein said audio playback device includes a digital music player on a personal computer.
US09/809,019 2001-03-16 2001-03-16 Method and system for embedding audio titles Expired - Lifetime US6448485B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US09/809,019 US6448485B1 (en) 2001-03-16 2001-03-16 Method and system for embedding audio titles

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/809,019 US6448485B1 (en) 2001-03-16 2001-03-16 Method and system for embedding audio titles

Publications (2)

Publication Number Publication Date
US6448485B1 US6448485B1 (en) 2002-09-10
US20020129692A1 true US20020129692A1 (en) 2002-09-19

Family

ID=25200353

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/809,019 Expired - Lifetime US6448485B1 (en) 2001-03-16 2001-03-16 Method and system for embedding audio titles

Country Status (1)

Country Link
US (1) US6448485B1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040125144A1 (en) * 2002-12-30 2004-07-01 Samsung Electronics Co., Ltd. Integrated management method and system for multimedia contents
US20040215461A1 (en) * 2003-04-24 2004-10-28 Visteon Global Technologies, Inc. Text-to-speech system for generating information announcements
US20050137869A1 (en) * 2003-12-17 2005-06-23 Samsung Electronics Co., Ltd. Method supporting text-to-speech navigation and multimedia device using the same
US20070005565A1 (en) * 2005-07-04 2007-01-04 Samsung Electronics., Ltd. Database searching method and apparatus
US20070100904A1 (en) * 2005-10-31 2007-05-03 Qwest Communications International Inc. Creation and transmission of rich content media
US20130300845A1 (en) * 2008-01-25 2013-11-14 At&T Intellectual Property I, L.P. System and Method for Digital Video Retrieval Involving Speech Recognition

Families Citing this family (137)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6757913B2 (en) * 1996-07-15 2004-06-29 Gregory D. Knox Wireless music and data transceiver system
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
US7426513B2 (en) * 2000-10-12 2008-09-16 Sap Portals Israel Ltd. Client-based objectifying of text pages
AU2002214613A1 (en) * 2000-11-08 2002-05-21 Digimarc Corporation Content authentication and recovery using digital watermarks
JP2002334086A (en) * 2001-05-10 2002-11-22 Sony Corp Information processor, its method, recording medium, and program
US6683241B2 (en) 2001-11-06 2004-01-27 James W. Wieder Pseudo-live music audio and sound
US7732697B1 (en) 2001-11-06 2010-06-08 Wieder James W Creating music and sound that varies from playback to playback
US8487176B1 (en) 2001-11-06 2013-07-16 James W. Wieder Music and sound that varies from one playback to another playback
US20030158737A1 (en) * 2002-02-15 2003-08-21 Csicsatka Tibor George Method and apparatus for incorporating additional audio information into audio data file identifying information
US20040186713A1 (en) * 2003-03-06 2004-09-23 Gomas Steven W. Content delivery and speech system and apparatus for the blind and print-handicapped
KR20050107610A (en) * 2003-03-10 2005-11-14 코닌클리케 필립스 일렉트로닉스 엔.브이. Record carrier and apparatus enabling seamless playback
WO2005003901A2 (en) * 2003-06-26 2005-01-13 Leave A Little Room Foundation Compressed media files with intrinsic supplementary content
US7757173B2 (en) * 2003-07-18 2010-07-13 Apple Inc. Voice menu system
US7213036B2 (en) 2003-08-12 2007-05-01 Aol Llc System for incorporating information about a source and usage of a media asset into the asset itself
US20060160486A1 (en) * 2005-01-14 2006-07-20 Xm Satellite Radio, Inc. Method and system for converting streaming digital data to FM modulated data
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US7825319B2 (en) 2005-10-06 2010-11-02 Pacing Technologies Llc System and method for pacing repetitive motion activities
US20110072955A1 (en) 2005-10-06 2011-03-31 Turner William D System and method for pacing repetitive motion activities
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US20080194175A1 (en) * 2007-02-09 2008-08-14 Intellitoys Llc Interactive toy providing, dynamic, navigable media content
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US10496753B2 (en) 2010-01-18 2019-12-03 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8898568B2 (en) 2008-09-09 2014-11-25 Apple Inc. Audio user interface
JP5428294B2 (en) * 2008-10-31 2014-02-26 ブラザー工業株式会社 Exercise content generation system, output terminal, exercise content generation device, content generation method, and content generation program
US9959870B2 (en) 2008-12-11 2018-05-01 Apple Inc. Speech recognition involving a mobile device
US20120309363A1 (en) 2011-06-03 2012-12-06 Apple Inc. Triggering notifications associated with tasks items that represent tasks to perform
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US10553209B2 (en) 2010-01-18 2020-02-04 Apple Inc. Systems and methods for hands-free notification summaries
US10679605B2 (en) 2010-01-18 2020-06-09 Apple Inc. Hands-free list-reading by intelligent automated assistant
US10705794B2 (en) 2010-01-18 2020-07-07 Apple Inc. Automatically adapting user interfaces for hands-free interaction
US8977584B2 (en) 2010-01-25 2015-03-10 Newvaluexchange Global Ai Llp Apparatuses, methods and systems for a digital conversation management platform
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US10762293B2 (en) 2010-12-22 2020-09-01 Apple Inc. Using parts-of-speech tagging and named entity recognition for spelling correction
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
JP2013178509A (en) * 2012-02-07 2013-09-09 Yamaha Corp Electronic equipment and voice guide program
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9483461B2 (en) 2012-03-06 2016-11-01 Apple Inc. Handling speech synthesis of content for multiple languages
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9495129B2 (en) 2012-06-29 2016-11-15 Apple Inc. Device, method, and user interface for voice-activated navigation and browsing of a document
US9576574B2 (en) 2012-09-10 2017-02-21 Apple Inc. Context-sensitive handling of interruptions by intelligent digital assistant
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
AU2014214676A1 (en) 2013-02-07 2015-08-27 Apple Inc. Voice trigger for a digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US9368114B2 (en) 2013-03-14 2016-06-14 Apple Inc. Context-sensitive handling of interruptions
US9922642B2 (en) 2013-03-15 2018-03-20 Apple Inc. Training an at least partial voice command system
WO2014144579A1 (en) 2013-03-15 2014-09-18 Apple Inc. System and method for updating an adaptive speech recognition model
US9424348B1 (en) 2013-05-08 2016-08-23 Rock My World, Inc. Sensor-driven audio playback modification
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
EP3937002A1 (en) 2013-06-09 2022-01-12 Apple Inc. Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
KR101809808B1 (en) 2013-06-13 2017-12-15 애플 인크. System and method for emergency calls initiated by voice command
US10791216B2 (en) 2013-08-06 2020-09-29 Apple Inc. Auto-activating smart responses based on activities from remote devices
US9620105B2 (en) 2014-05-15 2017-04-11 Apple Inc. Analyzing audio input for efficient speech and music recognition
US10592095B2 (en) 2014-05-23 2020-03-17 Apple Inc. Instantaneous speaking of content on touch devices
US9502031B2 (en) 2014-05-27 2016-11-22 Apple Inc. Method for supporting dynamic grammars in WFST-based ASR
US10289433B2 (en) 2014-05-30 2019-05-14 Apple Inc. Domain specific language for encoding assistant dialog
US9734193B2 (en) 2014-05-30 2017-08-15 Apple Inc. Determining domain salience ranking from ambiguous words in natural speech
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
TWI566107B (en) 2014-05-30 2017-01-11 蘋果公司 Method for processing a multi-part voice command, non-transitory computer readable storage medium and electronic device
US10078631B2 (en) 2014-05-30 2018-09-18 Apple Inc. Entropy-guided text prediction using combined word and character n-gram language models
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9785630B2 (en) 2014-05-30 2017-10-10 Apple Inc. Text prediction using combined word N-gram and unigram language models
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9760559B2 (en) 2014-05-30 2017-09-12 Apple Inc. Predictive text input
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US10446141B2 (en) 2014-08-28 2019-10-15 Apple Inc. Automatic speech recognition based on user feedback
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10552013B2 (en) 2014-12-02 2020-02-04 Apple Inc. Data detection
US9711141B2 (en) 2014-12-09 2017-07-18 Apple Inc. Disambiguating heteronyms in speech synthesis
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US10114607B1 (en) 2016-03-31 2018-10-30 Rock My World, Inc. Physiological state-driven playback tempo modification
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179588B1 (en) 2016-06-09 2019-02-22 Apple Inc. Intelligent automated assistant in a home environment
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK179560B1 (en) 2017-05-16 2019-02-18 Apple Inc. Far-field extension for digital assistant services

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4731847A (en) * 1982-04-26 1988-03-15 Texas Instruments Incorporated Electronic apparatus for simulating singing of song
DE69434787T2 (en) * 1993-05-26 2007-08-30 Pioneer Electronic Corp. Karaoke playback device
GB2300087B (en) * 1995-04-18 1999-04-28 Ibm Voice processing system and method
EP0915470A3 (en) * 1995-08-21 2004-03-24 Matsushita Electric Industrial Co., Ltd. Multimedia optical disk, reproduction apparatus and method for achieving variable scene development based on interactive control
US6212359B1 (en) * 1996-07-15 2001-04-03 Gregory D. Knox Wireless Transceiver System For Digital Music
US5918303A (en) * 1996-11-25 1999-06-29 Yamaha Corporation Performance setting data selecting apparatus
US5739451A (en) * 1996-12-27 1998-04-14 Franklin Electronic Publishers, Incorporated Hand held electronic music encyclopedia with text and note structure search
US6169543B1 (en) * 1998-12-28 2001-01-02 Thomson Licensing S.A. System and method for customizing program guide information to include reminder item or local identifier

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040125144A1 (en) * 2002-12-30 2004-07-01 Samsung Electronics Co., Ltd. Integrated management method and system for multimedia contents
US20040215461A1 (en) * 2003-04-24 2004-10-28 Visteon Global Technologies, Inc. Text-to-speech system for generating information announcements
US20050137869A1 (en) * 2003-12-17 2005-06-23 Samsung Electronics Co., Ltd. Method supporting text-to-speech navigation and multimedia device using the same
EP1548704A1 (en) * 2003-12-17 2005-06-29 Samsung Electronics Co., Ltd. Method supporting text-to-speech navigation and multimedia device using the same
US20070005565A1 (en) * 2005-07-04 2007-01-04 Samsung Electronics., Ltd. Database searching method and apparatus
US20070100904A1 (en) * 2005-10-31 2007-05-03 Qwest Communications International Inc. Creation and transmission of rich content media
US8856118B2 (en) * 2005-10-31 2014-10-07 Qwest Communications International Inc. Creation and transmission of rich content media
US20130300845A1 (en) * 2008-01-25 2013-11-14 At&T Intellectual Property I, L.P. System and Method for Digital Video Retrieval Involving Speech Recognition
US9135336B2 (en) * 2008-01-25 2015-09-15 At&T Intellectual Property I, L.P. System and method for digital video retrieval involving speech recognition
US9465870B2 (en) * 2008-01-25 2016-10-11 At&T Intellectual Property I, L.P. System and method for digital video retrieval involving speech recognition

Also Published As

Publication number Publication date
US6448485B1 (en) 2002-09-10

Similar Documents

Publication Publication Date Title
US6448485B1 (en) Method and system for embedding audio titles
US9875735B2 (en) System and method for synthetically generated speech describing media content
EP2491560B1 (en) Metadata time marking information for indicating a section of an audio object
US9153233B2 (en) Voice-controlled selection of media files utilizing phonetic data
JP4965565B2 (en) Playlist structure for large playlists
US20030158737A1 (en) Method and apparatus for incorporating additional audio information into audio data file identifying information
JP2009534696A (en) Apparatus and method for encoding and decoding a plurality of digital data sets
JP2007200495A (en) Music reproduction apparatus, music reproduction method and music reproduction program
US6525251B1 (en) Method of displaying the title and words of songs on a digital music player
CN1332365C (en) Method and device for sync controlling voice frequency and text information
US20040064208A1 (en) Dual mode store and play digital audio broadcast (DAB) receiver
US20060199161A1 (en) Method of creating multi-lingual lyrics slides video show for sing along
KR101143908B1 (en) Audio reproduction apparatus and method for providing disk jockey service
EP2343706A1 (en) Reproducing device and reproducing method
JP4362775B2 (en) Broadcast recording device
KR20010038854A (en) Method and format of music file for providing text and score
US7765198B2 (en) Data processing apparatus, data processing method, and data processing system
CN101031049A (en) Method and apparatus for playing audio-fequency data and other content simultaneouslly
KR100469052B1 (en) Audio system of vehicle
JP4312167B2 (en) Content playback device
JP2007257712A (en) Content selecting device
US20080195925A1 (en) Compressed Media Files with Intrinsic Supplementary Content
JP2005071548A (en) Audio reproducing device
KR20080086793A (en) Mobile device capable of audio playback
JP2004220710A (en) Audio-reproducing device

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTEL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BARILE, STEVEN E.;REEL/FRAME:011624/0864

Effective date: 20010309

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 8

SULP Surcharge for late payment

Year of fee payment: 7

FPAY Fee payment

Year of fee payment: 12