US20150032238A1 - Method and Device for Audio Input Routing - Google Patents
Method and Device for Audio Input Routing Download PDFInfo
- Publication number
- US20150032238A1 US20150032238A1 US14/142,190 US201314142190A US2015032238A1 US 20150032238 A1 US20150032238 A1 US 20150032238A1 US 201314142190 A US201314142190 A US 201314142190A US 2015032238 A1 US2015032238 A1 US 2015032238A1
- Authority
- US
- United States
- Prior art keywords
- audio input
- audio
- mobile device
- input
- microphone
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G06F17/3074—
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/20—Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/60—Substation equipment, e.g. for use by subscribers including speech amplifiers
- H04M1/6033—Substation equipment, e.g. for use by subscribers including speech amplifiers for providing handsfree use or a loudspeaker mode in telephone sets
- H04M1/6041—Portable telephones adapted for handsfree use
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/26—Devices for calling a subscriber
- H04M1/27—Devices whereby a plurality of signals may be stored simultaneously
- H04M1/271—Devices whereby a plurality of signals may be stored simultaneously controlled by voice recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/74—Details of telephonic subscriber devices with voice recognition means
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D30/00—Reducing energy consumption in communication networks
- Y02D30/70—Reducing energy consumption in communication networks in wireless communication networks
Definitions
- the present disclosure relates to processing audio signals and, more particularly, to methods and devices for routing audio signals including voice or speech.
- speech recognition has been around for decades, the quality of speech recognition software and hardware has only recently reached a high enough level to appeal to a large number of consumers.
- One area in which speech recognition has become very popular in recent years is the smartphone and tablet computer industry.
- a speech recognition-enabled device a consumer can perform such tasks as making phone calls, writing emails, and navigating with GPS, strictly by voice.
- Traditional voice recognition systems may receive an audio input from several input sources, such as a built-in microphone, a Bluetooth headset, or wired headset.
- input sources such as a built-in microphone, a Bluetooth headset, or wired headset.
- the behavior of the traditional systems is typically the same regardless of the input source except that a voice response may be played back from a different audio output, such as a phone speaker or the corresponding headset.
- FIG. 1 is a block diagram illustrating a mobile device, according to an embodiment
- FIG. 2 is a block diagram of example components of a mobile device, according to an embodiment
- FIG. 3 is a block diagram of a mobile device for receiving audio input from multiple input sources, according to an embodiment.
- FIG. 4 illustrates a process flow of a method for audio input routing that may be performed by the mobile device of FIG. 3 , according to an embodiment
- the speech is converted by a microphone to an audio input (e.g., an analog or digital signal).
- the audio input may be further processed, such as converted from analog-to-digital or encoded using an audio codec, before the mobile device recognizes the speech with an audio processor.
- Various input characteristics for the audio input affect its quality or recognition capability, such as a sampling rate or frequency range. These input characteristics may be dependent on the quality or features of the microphone or other components in an audio input path between the microphone and the audio processor.
- the various embodiments described herein allow a mobile device to determine the input characteristics for the audio input and to recognize the speech based on those characteristics.
- the mobile device configures or “tunes” the audio processor, for example, to improve accuracy, increase speed, or reduce power consumption for the voice recognition.
- the mobile device in one example performs a lookup of predetermined input characteristics for an input source, such as a microphone with a fixed sampling rate.
- the mobile device dynamically determines the input characteristics, for example, based on information associated with the audio input.
- the mobile device receives a trigger for the audio input that indicates which input source will provide the audio input, such as a microphone or headset.
- the trigger may further indicate an audio input path for the audio input (e.g., a wired path or wireless path).
- the mobile device determines at least one parameter for the audio processor based on input characteristics for the audio input or audio input path.
- the parameters may include the input characteristics themselves, such as a sampling rate of a microphone or latency of the audio input path.
- the parameter is an indicator for which voice recognition engine the audio processor should use, such as a high quality or low quality voice recognition engine.
- the mobile device routes the audio input to the audio processor, which then performs voice recognition based on the parameters.
- a mobile device receives a trigger for an audio input.
- the mobile device determines at least one parameter for an audio processor based on at least one input characteristic for the audio input.
- the mobile device routes the audio input to the audio processor with the at least one parameter.
- a mobile device receives a trigger for an audio input.
- the mobile device selects a microphone from a set of microphones based on the trigger.
- the set of microphones includes a local microphone of the mobile device and a remote microphone.
- the mobile device determines at least one parameter for an audio processor based on the selected microphone.
- the mobile device receives the audio input from the selected microphone and provides the audio input to the audio processor with the at least one parameter.
- the mobile device 100 may be any type of device capable of storing and executing multiple applications. Examples of the mobile device 100 include, but are not limited to, mobile devices, smart phones, smart watches, wireless devices, tablet computing devices, personal digital assistants, personal navigation devices, touch screen input device, touch or pen-based input devices, portable video and/or audio players, and the like. It is to be understood that the mobile device 100 may take the form of a variety of form factors, such as, but not limited to, bar, tablet, flip/clam, slider, rotator, and wearable form factors.
- the mobile device 100 has a housing 101 comprising a front surface 103 which includes a visible display 105 and a user interface.
- the user interface may be a touch screen including a touch-sensitive surface that overlays the display 105 .
- the user interface or touch screen of the mobile device 100 may include a touch-sensitive surface supported by the housing 101 that does not overlay any type of display.
- the user interface of the mobile device 100 may include one or more input keys 107 . Examples of the input key or keys 107 include, but are not limited to, keys of an alpha or numeric keypad or keyboard, a physical keys, touch-sensitive surfaces, mechanical surfaces, multipoint directional keys and side buttons or keys 107 .
- the mobile device 100 may also comprise a speaker 109 and microphone 111 for audio output and input at the surface. It is to be understood that the mobile device 100 may include a variety of different combination of displays and interfaces.
- the mobile device 100 includes one or more sensors 113 positioned at or within an exterior boundary of the housing 101 .
- the sensor or sensors 113 may be positioned at the front surface 103 and/or another surface (such as one or more side surfaces 115 ) of the exterior boundary of the housing 101 .
- the sensor or sensors 113 may include an exterior sensor supported at the exterior boundary to detect an environmental condition associated with an environment external to the housing.
- the sensor or sensors 113 may also, or in the alternative, include an interior sensors supported within the exterior boundary (i.e., internal to the housing) to detect a condition of the device itself. Examples of the sensors 113 are described below in reference to FIG. 2 .
- the components 200 include one or more wireless transceivers 201 , one or more processors 203 , one or more memories 205 , one or more output components 207 , and one or more input components 209 .
- the mobile device 100 includes a user interface, including the touch screen display 105 that comprises one or more of the output components 207 and one or more of the input components 209 .
- the mobile device 100 includes a plurality of the sensors 113 , several of which are described in more detail below. In the present embodiment, the sensors 113 are in communication with (so as to provide sensor signals to or receive control signals from) a sensor hub 224 .
- the components 200 include a device interface 215 to provide a direct connection to auxiliary components or accessories for additional or enhanced functionality.
- the internal components 200 include a power source or supply 217 , such as a portable battery, for providing power to the other internal components and allow portability of the mobile device 100 .
- all of the components 200 and particularly the wireless transceivers 201 , processors 203 , memories 205 , output components 207 , input components 209 , sensor hub 224 , device interface 215 , and power supply 217 , are coupled directly or indirectly with one another by way of one or more internal communication link(s) 218 (e.g., an internal communications bus).
- internal communication link(s) 218 e.g., an internal communications bus
- the wireless transceivers 201 particularly include a cellular transceiver 211 and a Wi-Fi transceiver 213 .
- the wireless transceivers 201 particularly include two of the wireless transceivers 211 and 213 , the present disclosure is intended to encompass numerous embodiments in which any arbitrary number of (e.g., more than two) wireless transceivers employing any arbitrary number of (e.g., two or more) communication technologies are present.
- the cellular transceiver 211 is configured to conduct cellular communications, such as 3G, 4G, 4G-LTE, vis-à-vis cell towers (not shown), albeit in other embodiments, the cellular transceiver 211 can be configured to utilize any of a variety of other cellular-based communication technologies such as analog communications (using AMPS), digital communications (using CDMA, TDMA, GSM, iDEN, GPRS, EDGE, etc.), or next generation communications (using UMTS, WCDMA, LTE, IEEE 802.16, etc.) or variants thereof.
- analog communications using AMPS
- digital communications using CDMA, TDMA, GSM, iDEN, GPRS, EDGE, etc.
- next generation communications using UMTS, WCDMA, LTE, IEEE 802.16, etc.
- the Wi-Fi transceiver 213 is a wireless local area network (WLAN) transceiver configured to conduct Wi-Fi communications in accordance with the IEEE 802.11 (a, b, g, or n) standard with access points.
- the Wi-Fi transceiver 213 can instead (or in addition) conduct other types of communications commonly understood as being encompassed within Wi-Fi communications such as some types of peer-to-peer (e.g., Wi-Fi Peer-to-Peer) communications.
- the Wi-Fi transceiver 213 can be replaced or supplemented with one or more other wireless transceivers configured for non-cellular wireless communications including, for example, wireless transceivers employing ad hoc communication technologies such as HomeRF (radio frequency), Home Node B (3G femtocell), Bluetooth, or other wireless communication technologies such as infrared technology.
- ad hoc communication technologies such as HomeRF (radio frequency), Home Node B (3G femtocell), Bluetooth, or other wireless communication technologies such as infrared technology.
- each of the wireless transceivers 201 serves as or includes both a respective transmitter and a respective receiver, it should be appreciated that the wireless transceivers are also intended to encompass one or more receiver(s) that are distinct from any transmitter(s), as well as one or more transmitter(s) that are distinct from any receiver(s).
- the wireless transceiver 201 includes at least one receiver that is a baseband receiver.
- Exemplary operation of the wireless transceivers 201 in conjunction with others of the components 200 of the mobile device 100 can take a variety of forms and can include, for example, operation in which, upon reception of wireless signals (as provided, for example, by remote device(s)), the internal components detect communication signals and the transceivers 201 demodulate the communication signals to recover incoming information, such as voice or data, transmitted by the wireless signals. After receiving the incoming information from the transceivers 201 , the processors 203 format the incoming information for the one or more output components 207 .
- the processors 203 format outgoing information, which can but need not be activated by the input components 209 , and convey the outgoing information to one or more of the wireless transceivers 201 for modulation so as to provide modulated communication signals to be transmitted.
- the wireless transceiver(s) 201 convey the modulated communication signals by way of wireless (as well as possibly wired) communication links to other devices (e.g., remote devices).
- the wireless transceivers 201 in one example allow the mobile device 100 to exchange messages with remote devices, for example, a remote network entity (not shown) of a cellular network or WLAN network.
- Examples of the remote network entity include an application server, web server, database server, or other network entity accessible through the wireless transceivers 201 either directly or indirectly via one or more intermediate devices or networks (e.g., via a WLAN access point, the Internet, LTE network, or other network).
- the output and input components 207 , 209 of the components 200 can include a variety of visual, audio, or mechanical outputs.
- the output device(s) 207 can include one or more visual output devices such as a cathode ray tube, liquid crystal display, plasma display, video screen, incandescent light, fluorescent light, front or rear projection display, and light emitting diode indicator, one or more audio output devices such as a speaker, alarm, or buzzer, or one or more mechanical output devices such as a vibrating mechanism or motion-based mechanism.
- the input device(s) 209 can include one or more visual input devices such as an optical sensor (for example, a camera lens and photosensor), one or more audio input devices such as a microphone, and one or more mechanical input devices such as a flip sensor, keyboard, keypad, selection button, navigation cluster, touch pad, capacitive sensor, motion sensor, and switch.
- visual input devices such as an optical sensor (for example, a camera lens and photosensor)
- audio input devices such as a microphone
- mechanical input devices such as a flip sensor, keyboard, keypad, selection button, navigation cluster, touch pad, capacitive sensor, motion sensor, and switch.
- the various sensors 113 in the present embodiment can be controlled by the sensor hub 224 , which can operate in response to or independent of the processor(s) 203 .
- the various sensors 113 may include, but are not limited to, power sensors, temperature sensors, pressure sensors, moisture sensors, ambient noise sensors, motion sensors (e.g., accelerometers or Gyro sensors), light sensors, proximity sensors (e.g., a light detecting sensor, an ultrasound transceiver or an infrared transceiver), other touch sensors, altitude sensors, one or more location circuits/components that can include, for example, a Global Positioning System (GPS) receiver, a triangulation receiver, an accelerometer, a tilt sensor, a gyroscope, or any other information collecting device that can identify a current location or user-device interface (carry mode) of the mobile device 100 .
- GPS Global Positioning System
- the processor(s) can include any one or more processing or control devices such as, for example, a microprocessor, digital signal processor, microcomputer, application-specific integrated circuit, etc.
- the processors 203 can generate commands, for example, based on information received from the one or more input components 209 .
- the processor(s) 203 can process the received information alone or in combination with other data, such as information stored in the memories 205 .
- the memories 205 of the components 200 can be used by the processors 203 to store and retrieve data.
- the memories (or memory portions) 205 of the components 200 can encompass one or more memory devices of any of a variety of forms (e.g., read-only memory, random access memory, static random access memory, dynamic random access memory, etc.), and can be used by the processors 203 to store and retrieve data.
- one or more of the memories 205 can be integrated with one or more of the processors 203 in a single device (e.g., a processing device including memory or processor-in-memory (PIM)), albeit such a single device will still typically have distinct portions/sections that perform the different processing and memory functions and that can be considered separate devices.
- the data that is stored by the memories 205 can include, but need not be limited to, operating systems, applications, and informational data.
- Each operating system includes executable code that controls basic functions of the mobile device 100 , such as interaction among the various components included among the components 200 , communication with external devices or networks via the wireless transceivers 201 or the device interface 215 , and storage and retrieval of applications and data, to and from the memories 205 .
- Each application includes executable code that utilizes an operating system to provide more specific functionality, such as file system service and handling of protected and unprotected data stored in the memories 205 .
- Such operating system or application information can include software update information (which can be understood to potentially encompass updates to either application(s) or operating system(s) or both).
- informational data this is non-executable code or information that can be referenced or manipulated by an operating system or application for performing functions of the mobile device 100 .
- FIG. 2 is provided for illustrative purposes only and for illustrating components of an mobile device in accordance with various embodiments, and is not intended to be a complete schematic diagram of the various components required for an mobile device. Therefore, an mobile device can include various other components not shown in FIG. 2 , or can include a combination of two or more components or a division of a particular component into two or more separate components, and still be within the scope of the disclosed embodiments.
- a block diagram 300 illustrates another embodiment of the mobile device 100 along with a remote microphone 319 .
- the mobile device 100 comprises the wireless transceiver 201 , the memory 205 , and the device interface 215 .
- the mobile device further comprises an audio processor 303 , a buffer 305 , and a local microphone 311 .
- the audio processor 303 in one example is an instance of the processor 203 , such as a digital signal processor or an application processor.
- the buffer 305 in one example is an instance of the memory 205 that is available as an intermediate buffer for audio inputs from the local microphone 311 .
- the local microphone 311 is an instance of the microphone 111 . In one example, multiple instances of the local microphone 311 cooperate to provide the audio input.
- the remote microphone 321 is remotely located from the mobile device 100 and not integrated with the mobile device 100 .
- the remote microphone 321 comprises a headset accessory for the mobile device 100 .
- the remote microphone 321 is a wireless-enabled headset that provides the audio input to the mobile device 100 via the wireless transceiver 201 (e.g., a Bluetooth transceiver via a synchronous connection-oriented link).
- the remote microphone 321 is a wired headset that provides the audio input to the mobile device 100 via the device interface 215 .
- the remote microphone 321 is located in an electronic device, such as a voice-activated household appliance (e.g., a television, thermostat, entertainment console, or lighting system), an automobile, desktop computer, or other devices as will be apparent to those skilled in the art.
- the remote microphone 321 provides the audio input to the mobile device 100 via the wireless transceiver 201 , such as the cellular transceiver 211 , WLAN transceiver 213 , or a Bluetooth transceiver.
- a plurality of remote microphones 321 cooperates to provide the voice input. For example, multiple microphones 321 may be spread throughout a user's home to provide a voice activation capability.
- a plurality of audio input paths are available for providing audio input to the mobile device 100 .
- a first path starts with the local microphone 311 and proceeds “directly” to the audio processor 303 (e.g., without substantial processing by other components).
- the audio input may be provided to the audio processor 303 substantially in real-time.
- the mobile device 100 comprises an Integrated Interchip Sound (“I2S”) bus for providing the audio input from the local microphone 311 to the audio processor 303 .
- I2S Integrated Interchip Sound
- a second path starts with the local microphone 311 and proceeds through the buffer 305 to reach the audio processor 303 .
- the mobile device 100 stores the audio input in the buffer 305 before providing the audio input to the audio processor 303 .
- Storage in an intermediate buffer e.g., the buffer 305 or the memory 205 ) allows the mobile device 100 additional time for initializing the audio processor 303 or to reduce the effects of high latency in receiving the audio input.
- a third path is based on both the first path and the second path. In this case, a first portion of the audio input is provided via the second path and buffered, for example, while the audio processor 303 is initialized. Once initialized, the mobile device 100 uses the first path to provide a second portion of the audio input directly from the local microphone 311 .
- the mobile device 100 uses the third path to provide a “one-shot” voice recognition feature of an always-on voice system.
- the mobile device 100 may listen for a trigger phrase (e.g., “OK Google Now”) and then buffer a command phrase that occurs after the trigger phrase while the audio processor 303 is initialized.
- a trigger phrase e.g., “OK Google Now”
- a fourth path for audio input starts with the remote microphone 321 and proceeds through the device interface 215 to the audio processor 303 .
- the remote microphone 321 or the device interface 215 may include a buffer (not shown) for buffering a portion of the audio input.
- a fifth path starts with the remote microphone 321 and proceeds through the wireless transceiver 201 to the audio processor 303 .
- a sixth path starts with the memory 205 and proceeds to the audio processor 303 .
- a software program or application records or stores the audio input in the memory 205 .
- the trigger e.g., a software trigger or inter-process communication trigger
- the audio input or a portion thereof may be stored in the memory 205 for access by the audio processor 303 while still being considered “direct” processing.
- the mobile device 100 in one example stores the audio input in the memory 205 to reduce the effects of high latency (e.g., from a remote microphone 321 over a Wi-Fi connection).
- a process flow 400 illustrates a method for audio input routing that may be performed by the mobile device 100 .
- the mobile device 100 receives ( 402 ) a trigger for an audio input.
- the trigger indicates which input source will provide the audio input, such as the local microphone 311 , the remote microphone 321 , or an application via the memory 205 .
- the trigger may further indicate an audio input path for the audio input (e.g., a wired path or wireless path), as described above, or other input characteristics for the audio input.
- the audio input path may also include multiple sub-paths for using multiple microphones in cooperation.
- the mobile device selects ( 403 ) the input source (or multiple sources) for the audio input.
- the trigger may be a processor interrupt, Bluetooth multi-function button trigger, software trigger, inter-process communication trigger, push notification, button press, audio keyword detection indicator (e.g., “OK Google Now”), or other user input.
- the mobile device 100 determines ( 404 ) whether the audio input path has an available buffer, such as the buffer 305 , a buffer integrated with the remote microphone 321 , or a buffer integrated with the wireless transceiver 201 .
- the mobile device 100 determines ( 406 ) at least one input characteristic for the audio input. Examples of the input characteristic include a sampling rate (e.g., 8 kHz, 44.1 kHz), quality indicator (e.g., “High Definition)”, frequency range (e.g., 300 Hz to 6 kHz), codec type, background noise level, compression feature (e.g., compression type or ratio), noise separation feature, or noise canceling feature for the audio input.
- the input characteristics may also include a transmission latency for the corresponding audio input path.
- the mobile device 100 in one example performs a lookup of predetermined input characteristics with an identifier of the input source, such as a microphone name or model number.
- the mobile device 100 dynamically determines the input characteristics, for example, based on information associated with the audio input.
- the mobile device 100 may determine a sampling rate or codec based on the audio input, such as from a header portion of a file that contains the audio input.
- the mobile device 100 determines ( 408 ) at least one audio parameter for the audio processor 303 .
- the audio parameters are used by the audio processor 303 for performing the voice recognition.
- the audio parameters may include one or more of the input characteristics, such as the sampling rate, frequency range, or availability of an intermediate buffer (e.g., the buffer 305 ).
- the local microphone 311 and the remote microphone 321 support different sampling rates.
- the mobile device 100 selects a sampling rate based on the microphone used for the audio input. For example, a Bluetooth headset may support an 8K sampling rate while the local microphone 311 supports a 16K sampling rate.
- the audio parameters may further include an indication of the paths used for the audio input.
- the audio parameter is an indicator for which voice recognition engine the audio processor should use, such as a high quality or low quality voice recognition engine. If the audio input path includes a wireless path (e.g., Bluetooth, Wi-Fi, or cellular), the mobile device 100 in one example selects the parameters based on a latency of the wireless path.
- a wireless path e.g., Bluetooth, Wi-Fi, or cellular
- the mobile device 100 initializes ( 410 ) the audio processor 303 with the audio parameters. Using the audio parameters, the mobile device 100 configures or “tunes” the audio processor 303 , for example, to improve accuracy, increase speed, or reduce power consumption for the voice recognition. In one example, the mobile device 100 sends a wakeup signal to the audio processor 303 for the initialization. In another example, the mobile device 100 passes the audio parameters in a function call to a voice recognition application running on the audio processor 303 . While initialization ( 410 ) is shown after the determination ( 408 ) of the audio parameters, in alternative implementations the mobile device 100 begins the initialization after receiving ( 402 ) the audio trigger.
- the mobile device 100 may perform the initialization in two or more steps, such as waking the audio processor 303 followed by configuration with the audio parameters.
- the audio processor 303 may also be started or “running” prior to the audio trigger, for example, as a background process or service, or in response to another voice recognition request.
- the mobile device 100 routes ( 412 ) the audio input from the input source to the audio processor 303 .
- the mobile device 100 streams the audio input to the audio processor 303 substantially in real-time.
- the mobile device 100 receives the audio input as an analog audio input from the local microphone 311 or the remote microphone 321 (via the transceiver 201 ).
- the mobile device 100 receives the audio input as a digital audio input from the local microphone 311 or the remote microphone 321 (via the transceiver 201 ).
- the audio input has two or more input sources such as an intermediate buffer and a direct input (e.g., the local microphone 311 )
- the mobile device 100 instructs the audio processor 303 to change from processing the audio input from the intermediate buffer to the direct input.
- the mobile device 100 may also provide a prompt or indication to the user when the mobile device 100 is ready to receive the audio input. For example, if an intermediate buffer is not available, the mobile device 100 may provide an audio indication, instruction, or “beep” once the audio processor 303 has been initialized. In another implementation, the mobile device 100 provides instructions to the user upon receipt of the trigger.
- the apparatus described herein may include a processor, a memory for storing program data to be executed by the processor, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, touch panel, keys, buttons, etc.
- these software modules may be stored as program instructions or computer readable code executable by the processor on a non-transitory computer-readable media such as magnetic storage media (e.g., magnetic tapes, hard disks, floppy disks), optical recording media (e.g., CD-ROMs, Digital Versatile Discs (DVDs), etc.), and solid state memory (e.g., random-access memory (RAM), read-only memory (ROM), static random-access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), flash memory, thumb drives, etc.).
- the computer readable recording media may also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. This computer readable recording media may be read by the computer, stored in the memory, and executed by the processor.
- the disclosed embodiments may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions.
- the disclosed embodiments may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices.
- the elements of the disclosed embodiments are implemented using software programming or software elements
- the disclosed embodiments may be implemented with any programming or scripting language such as C, C++, JAVA®, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements.
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Signal Processing (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Telephone Function (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
Description
- The present application claims priority to U.S. Provisional Patent Application 61/857,696, filed Jul. 23, 2013, and U.S. Provisional Patent Application 61/889,938, filed Oct. 11, 2013, the contents of which are hereby incorporated by reference herein.
- The present disclosure relates to processing audio signals and, more particularly, to methods and devices for routing audio signals including voice or speech.
- Although speech recognition has been around for decades, the quality of speech recognition software and hardware has only recently reached a high enough level to appeal to a large number of consumers. One area in which speech recognition has become very popular in recent years is the smartphone and tablet computer industry. Using a speech recognition-enabled device, a consumer can perform such tasks as making phone calls, writing emails, and navigating with GPS, strictly by voice.
- Traditional voice recognition systems may receive an audio input from several input sources, such as a built-in microphone, a Bluetooth headset, or wired headset. However, the behavior of the traditional systems is typically the same regardless of the input source except that a voice response may be played back from a different audio output, such as a phone speaker or the corresponding headset.
- While the appended claims set forth the features of the present techniques with particularity, these techniques, together with their objects and advantages, may be best understood from the following detailed description taken in conjunction with the accompanying drawings of which:
-
FIG. 1 is a block diagram illustrating a mobile device, according to an embodiment; -
FIG. 2 is a block diagram of example components of a mobile device, according to an embodiment; -
FIG. 3 is a block diagram of a mobile device for receiving audio input from multiple input sources, according to an embodiment. -
FIG. 4 illustrates a process flow of a method for audio input routing that may be performed by the mobile device ofFIG. 3 , according to an embodiment; - Turning to the drawings, wherein like reference numerals refer to like elements, techniques of the present disclosure are illustrated as being implemented in a suitable environment. The following description is based on embodiments of the claims and should not be taken as limiting the claims with regard to alternative embodiments that are not explicitly described herein.
- When a user provides speech for voice recognition by a mobile device, the speech is converted by a microphone to an audio input (e.g., an analog or digital signal). The audio input may be further processed, such as converted from analog-to-digital or encoded using an audio codec, before the mobile device recognizes the speech with an audio processor. Various input characteristics for the audio input affect its quality or recognition capability, such as a sampling rate or frequency range. These input characteristics may be dependent on the quality or features of the microphone or other components in an audio input path between the microphone and the audio processor.
- The various embodiments described herein allow a mobile device to determine the input characteristics for the audio input and to recognize the speech based on those characteristics. The mobile device configures or “tunes” the audio processor, for example, to improve accuracy, increase speed, or reduce power consumption for the voice recognition. The mobile device in one example performs a lookup of predetermined input characteristics for an input source, such as a microphone with a fixed sampling rate. In another example, the mobile device dynamically determines the input characteristics, for example, based on information associated with the audio input.
- The mobile device receives a trigger for the audio input that indicates which input source will provide the audio input, such as a microphone or headset. The trigger may further indicate an audio input path for the audio input (e.g., a wired path or wireless path). The mobile device determines at least one parameter for the audio processor based on input characteristics for the audio input or audio input path. The parameters may include the input characteristics themselves, such as a sampling rate of a microphone or latency of the audio input path. In another example, the parameter is an indicator for which voice recognition engine the audio processor should use, such as a high quality or low quality voice recognition engine. The mobile device routes the audio input to the audio processor, which then performs voice recognition based on the parameters.
- In one embodiment, a mobile device receives a trigger for an audio input. The mobile device determines at least one parameter for an audio processor based on at least one input characteristic for the audio input. The mobile device routes the audio input to the audio processor with the at least one parameter.
- In another embodiment, a mobile device receives a trigger for an audio input. The mobile device selects a microphone from a set of microphones based on the trigger. The set of microphones includes a local microphone of the mobile device and a remote microphone. The mobile device determines at least one parameter for an audio processor based on the selected microphone. The mobile device receives the audio input from the selected microphone and provides the audio input to the audio processor with the at least one parameter.
- Referring to
FIG. 1 , there is illustrated a perspective view of an examplemobile device 100. Themobile device 100 may be any type of device capable of storing and executing multiple applications. Examples of themobile device 100 include, but are not limited to, mobile devices, smart phones, smart watches, wireless devices, tablet computing devices, personal digital assistants, personal navigation devices, touch screen input device, touch or pen-based input devices, portable video and/or audio players, and the like. It is to be understood that themobile device 100 may take the form of a variety of form factors, such as, but not limited to, bar, tablet, flip/clam, slider, rotator, and wearable form factors. - For one embodiment, the
mobile device 100 has ahousing 101 comprising afront surface 103 which includes avisible display 105 and a user interface. For example, the user interface may be a touch screen including a touch-sensitive surface that overlays thedisplay 105. For another embodiment, the user interface or touch screen of themobile device 100 may include a touch-sensitive surface supported by thehousing 101 that does not overlay any type of display. For yet another embodiment, the user interface of themobile device 100 may include one ormore input keys 107. Examples of the input key orkeys 107 include, but are not limited to, keys of an alpha or numeric keypad or keyboard, a physical keys, touch-sensitive surfaces, mechanical surfaces, multipoint directional keys and side buttons orkeys 107. Themobile device 100 may also comprise aspeaker 109 and microphone 111 for audio output and input at the surface. It is to be understood that themobile device 100 may include a variety of different combination of displays and interfaces. - The
mobile device 100 includes one ormore sensors 113 positioned at or within an exterior boundary of thehousing 101. For example, as illustrated byFIG. 1 , the sensor orsensors 113 may be positioned at thefront surface 103 and/or another surface (such as one or more side surfaces 115) of the exterior boundary of thehousing 101. The sensor orsensors 113 may include an exterior sensor supported at the exterior boundary to detect an environmental condition associated with an environment external to the housing. The sensor orsensors 113 may also, or in the alternative, include an interior sensors supported within the exterior boundary (i.e., internal to the housing) to detect a condition of the device itself. Examples of thesensors 113 are described below in reference toFIG. 2 . - Referring to
FIG. 2 , there is shown a block diagram representing example components (e.g., internal components) 200 of themobile device 100 ofFIG. 1 . In the present embodiment, thecomponents 200 include one or morewireless transceivers 201, one ormore processors 203, one ormore memories 205, one ormore output components 207, and one ormore input components 209. As already noted above, themobile device 100 includes a user interface, including thetouch screen display 105 that comprises one or more of theoutput components 207 and one or more of theinput components 209. Also as already discussed above, themobile device 100 includes a plurality of thesensors 113, several of which are described in more detail below. In the present embodiment, thesensors 113 are in communication with (so as to provide sensor signals to or receive control signals from) asensor hub 224. - Further, the
components 200 include adevice interface 215 to provide a direct connection to auxiliary components or accessories for additional or enhanced functionality. In addition, theinternal components 200 include a power source orsupply 217, such as a portable battery, for providing power to the other internal components and allow portability of themobile device 100. As shown, all of thecomponents 200, and particularly thewireless transceivers 201,processors 203,memories 205,output components 207,input components 209,sensor hub 224,device interface 215, andpower supply 217, are coupled directly or indirectly with one another by way of one or more internal communication link(s) 218 (e.g., an internal communications bus). - Further, in the present embodiment of
FIG. 2 , thewireless transceivers 201 particularly include acellular transceiver 211 and a Wi-Fi transceiver 213. Although in the present embodiment thewireless transceivers 201 particularly include two of thewireless transceivers cellular transceiver 211 is configured to conduct cellular communications, such as 3G, 4G, 4G-LTE, vis-à-vis cell towers (not shown), albeit in other embodiments, thecellular transceiver 211 can be configured to utilize any of a variety of other cellular-based communication technologies such as analog communications (using AMPS), digital communications (using CDMA, TDMA, GSM, iDEN, GPRS, EDGE, etc.), or next generation communications (using UMTS, WCDMA, LTE, IEEE 802.16, etc.) or variants thereof. - By contrast, the Wi-
Fi transceiver 213 is a wireless local area network (WLAN) transceiver configured to conduct Wi-Fi communications in accordance with the IEEE 802.11 (a, b, g, or n) standard with access points. In other embodiments, the Wi-Fi transceiver 213 can instead (or in addition) conduct other types of communications commonly understood as being encompassed within Wi-Fi communications such as some types of peer-to-peer (e.g., Wi-Fi Peer-to-Peer) communications. Further, in other embodiments, the Wi-Fi transceiver 213 can be replaced or supplemented with one or more other wireless transceivers configured for non-cellular wireless communications including, for example, wireless transceivers employing ad hoc communication technologies such as HomeRF (radio frequency), Home Node B (3G femtocell), Bluetooth, or other wireless communication technologies such as infrared technology. Although in the present embodiment each of thewireless transceivers 201 serves as or includes both a respective transmitter and a respective receiver, it should be appreciated that the wireless transceivers are also intended to encompass one or more receiver(s) that are distinct from any transmitter(s), as well as one or more transmitter(s) that are distinct from any receiver(s). In one example embodiment encompassed herein, thewireless transceiver 201 includes at least one receiver that is a baseband receiver. - Exemplary operation of the
wireless transceivers 201 in conjunction with others of thecomponents 200 of themobile device 100 can take a variety of forms and can include, for example, operation in which, upon reception of wireless signals (as provided, for example, by remote device(s)), the internal components detect communication signals and thetransceivers 201 demodulate the communication signals to recover incoming information, such as voice or data, transmitted by the wireless signals. After receiving the incoming information from thetransceivers 201, theprocessors 203 format the incoming information for the one ormore output components 207. Likewise, for transmission of wireless signals, theprocessors 203 format outgoing information, which can but need not be activated by theinput components 209, and convey the outgoing information to one or more of thewireless transceivers 201 for modulation so as to provide modulated communication signals to be transmitted. The wireless transceiver(s) 201 convey the modulated communication signals by way of wireless (as well as possibly wired) communication links to other devices (e.g., remote devices). Thewireless transceivers 201 in one example allow themobile device 100 to exchange messages with remote devices, for example, a remote network entity (not shown) of a cellular network or WLAN network. Examples of the remote network entity include an application server, web server, database server, or other network entity accessible through thewireless transceivers 201 either directly or indirectly via one or more intermediate devices or networks (e.g., via a WLAN access point, the Internet, LTE network, or other network). - Depending upon the embodiment, the output and
input components components 200 can include a variety of visual, audio, or mechanical outputs. For example, the output device(s) 207 can include one or more visual output devices such as a cathode ray tube, liquid crystal display, plasma display, video screen, incandescent light, fluorescent light, front or rear projection display, and light emitting diode indicator, one or more audio output devices such as a speaker, alarm, or buzzer, or one or more mechanical output devices such as a vibrating mechanism or motion-based mechanism. Likewise, by example, the input device(s) 209 can include one or more visual input devices such as an optical sensor (for example, a camera lens and photosensor), one or more audio input devices such as a microphone, and one or more mechanical input devices such as a flip sensor, keyboard, keypad, selection button, navigation cluster, touch pad, capacitive sensor, motion sensor, and switch. - As already noted, the
various sensors 113 in the present embodiment can be controlled by thesensor hub 224, which can operate in response to or independent of the processor(s) 203. Examples of thevarious sensors 113 may include, but are not limited to, power sensors, temperature sensors, pressure sensors, moisture sensors, ambient noise sensors, motion sensors (e.g., accelerometers or Gyro sensors), light sensors, proximity sensors (e.g., a light detecting sensor, an ultrasound transceiver or an infrared transceiver), other touch sensors, altitude sensors, one or more location circuits/components that can include, for example, a Global Positioning System (GPS) receiver, a triangulation receiver, an accelerometer, a tilt sensor, a gyroscope, or any other information collecting device that can identify a current location or user-device interface (carry mode) of themobile device 100. - With respect to the processor(s) 203, the processor(s) can include any one or more processing or control devices such as, for example, a microprocessor, digital signal processor, microcomputer, application-specific integrated circuit, etc. The
processors 203 can generate commands, for example, based on information received from the one ormore input components 209. The processor(s) 203 can process the received information alone or in combination with other data, such as information stored in thememories 205. Thus, thememories 205 of thecomponents 200 can be used by theprocessors 203 to store and retrieve data. - Further, the memories (or memory portions) 205 of the
components 200 can encompass one or more memory devices of any of a variety of forms (e.g., read-only memory, random access memory, static random access memory, dynamic random access memory, etc.), and can be used by theprocessors 203 to store and retrieve data. In some embodiments, one or more of thememories 205 can be integrated with one or more of theprocessors 203 in a single device (e.g., a processing device including memory or processor-in-memory (PIM)), albeit such a single device will still typically have distinct portions/sections that perform the different processing and memory functions and that can be considered separate devices. The data that is stored by thememories 205 can include, but need not be limited to, operating systems, applications, and informational data. - Each operating system includes executable code that controls basic functions of the
mobile device 100, such as interaction among the various components included among thecomponents 200, communication with external devices or networks via thewireless transceivers 201 or thedevice interface 215, and storage and retrieval of applications and data, to and from thememories 205. Each application includes executable code that utilizes an operating system to provide more specific functionality, such as file system service and handling of protected and unprotected data stored in thememories 205. Such operating system or application information can include software update information (which can be understood to potentially encompass updates to either application(s) or operating system(s) or both). As for informational data, this is non-executable code or information that can be referenced or manipulated by an operating system or application for performing functions of themobile device 100. - It is to be understood that
FIG. 2 is provided for illustrative purposes only and for illustrating components of an mobile device in accordance with various embodiments, and is not intended to be a complete schematic diagram of the various components required for an mobile device. Therefore, an mobile device can include various other components not shown inFIG. 2 , or can include a combination of two or more components or a division of a particular component into two or more separate components, and still be within the scope of the disclosed embodiments. - Turning to
FIG. 3 , a block diagram 300 illustrates another embodiment of themobile device 100 along with a remote microphone 319. As shown inFIG. 3 , themobile device 100 comprises thewireless transceiver 201, thememory 205, and thedevice interface 215. The mobile device further comprises anaudio processor 303, abuffer 305, and alocal microphone 311. Theaudio processor 303 in one example is an instance of theprocessor 203, such as a digital signal processor or an application processor. Thebuffer 305 in one example is an instance of thememory 205 that is available as an intermediate buffer for audio inputs from thelocal microphone 311. Thelocal microphone 311 is an instance of themicrophone 111. In one example, multiple instances of thelocal microphone 311 cooperate to provide the audio input. - The
remote microphone 321 is remotely located from themobile device 100 and not integrated with themobile device 100. For example, theremote microphone 321 comprises a headset accessory for themobile device 100. In the implementation shown inFIG. 3 , theremote microphone 321 is a wireless-enabled headset that provides the audio input to themobile device 100 via the wireless transceiver 201 (e.g., a Bluetooth transceiver via a synchronous connection-oriented link). In another example, theremote microphone 321 is a wired headset that provides the audio input to themobile device 100 via thedevice interface 215. In yet another example, theremote microphone 321 is located in an electronic device, such as a voice-activated household appliance (e.g., a television, thermostat, entertainment console, or lighting system), an automobile, desktop computer, or other devices as will be apparent to those skilled in the art. In this case, theremote microphone 321 provides the audio input to themobile device 100 via thewireless transceiver 201, such as thecellular transceiver 211,WLAN transceiver 213, or a Bluetooth transceiver. In one example, a plurality ofremote microphones 321 cooperates to provide the voice input. For example,multiple microphones 321 may be spread throughout a user's home to provide a voice activation capability. - Referring to
FIG. 3 , a plurality of audio input paths are available for providing audio input to themobile device 100. A first path starts with thelocal microphone 311 and proceeds “directly” to the audio processor 303 (e.g., without substantial processing by other components). In this case, the audio input may be provided to theaudio processor 303 substantially in real-time. In one example, themobile device 100 comprises an Integrated Interchip Sound (“I2S”) bus for providing the audio input from thelocal microphone 311 to theaudio processor 303. - A second path starts with the
local microphone 311 and proceeds through thebuffer 305 to reach theaudio processor 303. In this case, themobile device 100 stores the audio input in thebuffer 305 before providing the audio input to theaudio processor 303. Storage in an intermediate buffer (e.g., thebuffer 305 or the memory 205) allows themobile device 100 additional time for initializing theaudio processor 303 or to reduce the effects of high latency in receiving the audio input. A third path is based on both the first path and the second path. In this case, a first portion of the audio input is provided via the second path and buffered, for example, while theaudio processor 303 is initialized. Once initialized, themobile device 100 uses the first path to provide a second portion of the audio input directly from thelocal microphone 311. Themobile device 100 in one example uses the third path to provide a “one-shot” voice recognition feature of an always-on voice system. For example, themobile device 100 may listen for a trigger phrase (e.g., “OK Google Now”) and then buffer a command phrase that occurs after the trigger phrase while theaudio processor 303 is initialized. - A fourth path for audio input starts with the
remote microphone 321 and proceeds through thedevice interface 215 to theaudio processor 303. In this case, theremote microphone 321 or thedevice interface 215 may include a buffer (not shown) for buffering a portion of the audio input. A fifth path starts with theremote microphone 321 and proceeds through thewireless transceiver 201 to theaudio processor 303. A sixth path starts with thememory 205 and proceeds to theaudio processor 303. In this case, a software program or application records or stores the audio input in thememory 205. Upon receipt of the trigger (e.g., a software trigger or inter-process communication trigger), themobile device 100 obtains the pre-recorded audio input from thememory 205. For the paths described herein, the audio input or a portion thereof may be stored in thememory 205 for access by theaudio processor 303 while still being considered “direct” processing. Themobile device 100 in one example stores the audio input in thememory 205 to reduce the effects of high latency (e.g., from aremote microphone 321 over a Wi-Fi connection). - Turning to
FIG. 4 , aprocess flow 400 illustrates a method for audio input routing that may be performed by themobile device 100. Themobile device 100 receives (402) a trigger for an audio input. The trigger indicates which input source will provide the audio input, such as thelocal microphone 311, theremote microphone 321, or an application via thememory 205. The trigger may further indicate an audio input path for the audio input (e.g., a wired path or wireless path), as described above, or other input characteristics for the audio input. The audio input path may also include multiple sub-paths for using multiple microphones in cooperation. Based on the audio trigger, the mobile device selects (403) the input source (or multiple sources) for the audio input. The trigger may be a processor interrupt, Bluetooth multi-function button trigger, software trigger, inter-process communication trigger, push notification, button press, audio keyword detection indicator (e.g., “OK Google Now”), or other user input. - The
mobile device 100 determines (404) whether the audio input path has an available buffer, such as thebuffer 305, a buffer integrated with theremote microphone 321, or a buffer integrated with thewireless transceiver 201. Themobile device 100 determines (406) at least one input characteristic for the audio input. Examples of the input characteristic include a sampling rate (e.g., 8 kHz, 44.1 kHz), quality indicator (e.g., “High Definition)”, frequency range (e.g., 300 Hz to 6 kHz), codec type, background noise level, compression feature (e.g., compression type or ratio), noise separation feature, or noise canceling feature for the audio input. The input characteristics may also include a transmission latency for the corresponding audio input path. Themobile device 100 in one example performs a lookup of predetermined input characteristics with an identifier of the input source, such as a microphone name or model number. In another example, themobile device 100 dynamically determines the input characteristics, for example, based on information associated with the audio input. In this case, themobile device 100 may determine a sampling rate or codec based on the audio input, such as from a header portion of a file that contains the audio input. - After determination (406) of the input characteristics, the
mobile device 100 determines (408) at least one audio parameter for theaudio processor 303. The audio parameters are used by theaudio processor 303 for performing the voice recognition. The audio parameters may include one or more of the input characteristics, such as the sampling rate, frequency range, or availability of an intermediate buffer (e.g., the buffer 305). In one example, thelocal microphone 311 and theremote microphone 321 support different sampling rates. In this case, themobile device 100 selects a sampling rate based on the microphone used for the audio input. For example, a Bluetooth headset may support an 8K sampling rate while thelocal microphone 311 supports a 16K sampling rate. If the intermediate buffer is available or theremote microphone 321 is selected, the audio parameters may further include an indication of the paths used for the audio input. In another example, the audio parameter is an indicator for which voice recognition engine the audio processor should use, such as a high quality or low quality voice recognition engine. If the audio input path includes a wireless path (e.g., Bluetooth, Wi-Fi, or cellular), themobile device 100 in one example selects the parameters based on a latency of the wireless path. - The
mobile device 100 initializes (410) theaudio processor 303 with the audio parameters. Using the audio parameters, themobile device 100 configures or “tunes” theaudio processor 303, for example, to improve accuracy, increase speed, or reduce power consumption for the voice recognition. In one example, themobile device 100 sends a wakeup signal to theaudio processor 303 for the initialization. In another example, themobile device 100 passes the audio parameters in a function call to a voice recognition application running on theaudio processor 303. While initialization (410) is shown after the determination (408) of the audio parameters, in alternative implementations themobile device 100 begins the initialization after receiving (402) the audio trigger. In this case, themobile device 100 may perform the initialization in two or more steps, such as waking theaudio processor 303 followed by configuration with the audio parameters. Theaudio processor 303 may also be started or “running” prior to the audio trigger, for example, as a background process or service, or in response to another voice recognition request. - When the
audio processor 303 is ready to receive the audio input, themobile device 100 routes (412) the audio input from the input source to theaudio processor 303. In one example, themobile device 100 streams the audio input to theaudio processor 303 substantially in real-time. In another example, themobile device 100 receives the audio input as an analog audio input from thelocal microphone 311 or the remote microphone 321 (via the transceiver 201). In yet another example, themobile device 100 receives the audio input as a digital audio input from thelocal microphone 311 or the remote microphone 321 (via the transceiver 201). Where the audio input has two or more input sources such as an intermediate buffer and a direct input (e.g., the local microphone 311), themobile device 100 instructs theaudio processor 303 to change from processing the audio input from the intermediate buffer to the direct input. - The
mobile device 100 may also provide a prompt or indication to the user when themobile device 100 is ready to receive the audio input. For example, if an intermediate buffer is not available, themobile device 100 may provide an audio indication, instruction, or “beep” once theaudio processor 303 has been initialized. In another implementation, themobile device 100 provides instructions to the user upon receipt of the trigger. - It can be seen from the foregoing that a method and system for audio input routing have been described. In view of the many possible embodiments to which the principles of the present discussion may be applied, it should be recognized that the embodiments described herein with respect to the drawing figures are meant to be illustrative only and should not be taken as limiting the scope of the claims. Therefore, the techniques as described herein contemplate all such embodiments as may come within the scope of the following claims and equivalents thereof.
- The apparatus described herein may include a processor, a memory for storing program data to be executed by the processor, a permanent storage such as a disk drive, a communications port for handling communications with external devices, and user interface devices, including a display, touch panel, keys, buttons, etc. When software modules are involved, these software modules may be stored as program instructions or computer readable code executable by the processor on a non-transitory computer-readable media such as magnetic storage media (e.g., magnetic tapes, hard disks, floppy disks), optical recording media (e.g., CD-ROMs, Digital Versatile Discs (DVDs), etc.), and solid state memory (e.g., random-access memory (RAM), read-only memory (ROM), static random-access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), flash memory, thumb drives, etc.). The computer readable recording media may also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. This computer readable recording media may be read by the computer, stored in the memory, and executed by the processor.
- The disclosed embodiments may be described in terms of functional block components and various processing steps. Such functional blocks may be realized by any number of hardware and/or software components configured to perform the specified functions. For example, the disclosed embodiments may employ various integrated circuit components, e.g., memory elements, processing elements, logic elements, look-up tables, and the like, which may carry out a variety of functions under the control of one or more microprocessors or other control devices. Similarly, where the elements of the disclosed embodiments are implemented using software programming or software elements, the disclosed embodiments may be implemented with any programming or scripting language such as C, C++, JAVA®, assembler, or the like, with the various algorithms being implemented with any combination of data structures, objects, processes, routines or other programming elements. Functional aspects may be implemented in algorithms that execute on one or more processors. Furthermore, the disclosed embodiments may employ any number of conventional techniques for electronics configuration, signal processing and/or control, data processing and the like. Finally, the steps of all methods described herein may be performed in any suitable order unless otherwise indicated herein or otherwise clearly contradicted by context.
- For the sake of brevity, conventional electronics, control systems, software development and other functional aspects of the systems (and components of the individual operating components of the systems) may not be described in detail. Furthermore, the connecting lines, or connectors shown in the various figures presented are intended to represent exemplary functional relationships and/or physical or logical couplings between the various elements. It should be noted that many alternative or additional functional relationships, physical connections or logical connections may be present in a practical device. The words “mechanism”, “element”, “unit”, “structure”, “means”, “device”, “controller”, and “construction” are used broadly and are not limited to mechanical or physical embodiments, but may include software routines in conjunction with processors, etc.
- No item or component is essential to the practice of the disclosed embodiments unless the element is specifically described as “essential” or “critical”. It will also be recognized that the terms “comprises,” “comprising,” “includes,” “including,” “has,” and “having,” as used herein, are specifically intended to be read as open-ended terms of art. The use of the terms “a” and “an” and “the” and similar referents in the context of describing the disclosed embodiments (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless the context clearly indicates otherwise. In addition, it should be understood that although the terms “first,” “second,” etc. may be used herein to describe various elements, these elements should not be limited by these terms, which are only used to distinguish one element from another. Furthermore, recitation of ranges of values herein are merely intended to serve as a shorthand method of referring individually to each separate value falling within the range, unless otherwise indicated herein, and each separate value is incorporated into the specification as if it were individually recited herein.
- The use of any and all examples, or exemplary language (e.g., “such as”) provided herein, is intended merely to better illuminate the disclosed embodiments and does not pose a limitation on the scope of the disclosed embodiments unless otherwise claimed. Numerous modifications and adaptations will be readily apparent to those of ordinary skill in this art.
Claims (20)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/142,190 US20150032238A1 (en) | 2013-07-23 | 2013-12-27 | Method and Device for Audio Input Routing |
PCT/US2014/047448 WO2015013201A2 (en) | 2013-07-23 | 2014-07-21 | Method and device for audio imput routing |
US16/703,003 US11363128B2 (en) | 2013-07-23 | 2019-12-04 | Method and device for audio input routing |
US17/833,446 US11876922B2 (en) | 2013-07-23 | 2022-06-06 | Method and device for audio input routing |
US18/537,386 US20240137435A1 (en) | 2013-07-23 | 2023-12-12 | Method and device for audio input routing |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361857696P | 2013-07-23 | 2013-07-23 | |
US201361889938P | 2013-10-11 | 2013-10-11 | |
US14/142,190 US20150032238A1 (en) | 2013-07-23 | 2013-12-27 | Method and Device for Audio Input Routing |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/703,003 Continuation US11363128B2 (en) | 2013-07-23 | 2019-12-04 | Method and device for audio input routing |
Publications (1)
Publication Number | Publication Date |
---|---|
US20150032238A1 true US20150032238A1 (en) | 2015-01-29 |
Family
ID=52390940
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/142,234 Abandoned US20150031416A1 (en) | 2013-07-23 | 2013-12-27 | Method and Device For Command Phrase Validation |
US14/142,190 Abandoned US20150032238A1 (en) | 2013-07-23 | 2013-12-27 | Method and Device for Audio Input Routing |
US16/703,003 Active 2034-04-01 US11363128B2 (en) | 2013-07-23 | 2019-12-04 | Method and device for audio input routing |
US17/833,446 Active US11876922B2 (en) | 2013-07-23 | 2022-06-06 | Method and device for audio input routing |
US18/537,386 Pending US20240137435A1 (en) | 2013-07-23 | 2023-12-12 | Method and device for audio input routing |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/142,234 Abandoned US20150031416A1 (en) | 2013-07-23 | 2013-12-27 | Method and Device For Command Phrase Validation |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/703,003 Active 2034-04-01 US11363128B2 (en) | 2013-07-23 | 2019-12-04 | Method and device for audio input routing |
US17/833,446 Active US11876922B2 (en) | 2013-07-23 | 2022-06-06 | Method and device for audio input routing |
US18/537,386 Pending US20240137435A1 (en) | 2013-07-23 | 2023-12-12 | Method and device for audio input routing |
Country Status (2)
Country | Link |
---|---|
US (5) | US20150031416A1 (en) |
WO (1) | WO2015013201A2 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9799337B2 (en) * | 2016-03-02 | 2017-10-24 | Merry Electronics(Suzhou) Co., Ltd. | Microphone apparatus for enhancing power conservation |
US20180069815A1 (en) * | 2016-09-02 | 2018-03-08 | Bose Corporation | Application-based messaging system using headphones |
KR20190090532A (en) * | 2018-01-25 | 2019-08-02 | 삼성전자주식회사 | Application processor for low power operation, electronic device including the same and method of operating the same |
CN112652317A (en) * | 2019-10-09 | 2021-04-13 | 深圳云程科技有限公司 | High-sampling-rate audio playing method and system based on Android system |
US20210110824A1 (en) * | 2019-10-10 | 2021-04-15 | Samsung Electronics Co., Ltd. | Electronic apparatus and controlling method thereof |
US20210118448A1 (en) * | 2014-10-09 | 2021-04-22 | Google Llc | Hotword Detection on Multiple Devices |
CN112817387A (en) * | 2019-11-18 | 2021-05-18 | 英业达科技有限公司 | Radio component |
Families Citing this family (138)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8677377B2 (en) | 2005-09-08 | 2014-03-18 | Apple Inc. | Method and apparatus for building an intelligent automated assistant |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US8977255B2 (en) | 2007-04-03 | 2015-03-10 | Apple Inc. | Method and system for operating a multi-function portable electronic device using voice-activation |
US10002189B2 (en) | 2007-12-20 | 2018-06-19 | Apple Inc. | Method and apparatus for searching using an active ontology |
US9330720B2 (en) | 2008-01-03 | 2016-05-03 | Apple Inc. | Methods and apparatus for altering audio output signals |
US20100030549A1 (en) | 2008-07-31 | 2010-02-04 | Lee Michael M | Mobile device having human language translation capability with positional feedback |
US8676904B2 (en) | 2008-10-02 | 2014-03-18 | Apple Inc. | Electronic devices with voice command and contextual data processing capabilities |
US20120311585A1 (en) | 2011-06-03 | 2012-12-06 | Apple Inc. | Organizing task items that represent tasks to perform |
US10276170B2 (en) | 2010-01-18 | 2019-04-30 | Apple Inc. | Intelligent automated assistant |
US8682667B2 (en) | 2010-02-25 | 2014-03-25 | Apple Inc. | User profiling for selecting user specific voice input processing information |
US9262612B2 (en) | 2011-03-21 | 2016-02-16 | Apple Inc. | Device access using voice authentication |
US10057736B2 (en) | 2011-06-03 | 2018-08-21 | Apple Inc. | Active transport based notifications |
US10134385B2 (en) | 2012-03-02 | 2018-11-20 | Apple Inc. | Systems and methods for name pronunciation |
US10417037B2 (en) | 2012-05-15 | 2019-09-17 | Apple Inc. | Systems and methods for integrating third party services with a digital assistant |
US9721563B2 (en) | 2012-06-08 | 2017-08-01 | Apple Inc. | Name recognition system |
AU2014214676A1 (en) | 2013-02-07 | 2015-08-27 | Apple Inc. | Voice trigger for a digital assistant |
US10652394B2 (en) | 2013-03-14 | 2020-05-12 | Apple Inc. | System and method for processing voicemail |
US10748529B1 (en) | 2013-03-15 | 2020-08-18 | Apple Inc. | Voice activated device for use with a voice-based digital assistant |
WO2014197335A1 (en) | 2013-06-08 | 2014-12-11 | Apple Inc. | Interpreting and acting upon commands that involve sharing information with remote devices |
US10176167B2 (en) | 2013-06-09 | 2019-01-08 | Apple Inc. | System and method for inferring user intent from speech inputs |
KR101922663B1 (en) | 2013-06-09 | 2018-11-28 | 애플 인크. | Device, method, and graphical user interface for enabling conversation persistence across two or more instances of a digital assistant |
US20150031416A1 (en) | 2013-07-23 | 2015-01-29 | Motorola Mobility Llc | Method and Device For Command Phrase Validation |
DE112014003653B4 (en) | 2013-08-06 | 2024-04-18 | Apple Inc. | Automatically activate intelligent responses based on activities from remote devices |
US9508345B1 (en) | 2013-09-24 | 2016-11-29 | Knowles Electronics, Llc | Continuous voice sensing |
US10296160B2 (en) | 2013-12-06 | 2019-05-21 | Apple Inc. | Method for extracting salient dialog usage from live data |
US9953634B1 (en) | 2013-12-17 | 2018-04-24 | Knowles Electronics, Llc | Passive training for automatic speech recognition |
US9437188B1 (en) | 2014-03-28 | 2016-09-06 | Knowles Electronics, Llc | Buffered reprocessing for multi-microphone automatic speech recognition assist |
US10170123B2 (en) | 2014-05-30 | 2019-01-01 | Apple Inc. | Intelligent assistant for home automation |
US9715875B2 (en) | 2014-05-30 | 2017-07-25 | Apple Inc. | Reducing the need for manual start/end-pointing and trigger phrases |
US9633004B2 (en) | 2014-05-30 | 2017-04-25 | Apple Inc. | Better resolution when referencing to concepts |
US9966065B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | Multi-command single utterance input method |
US9430463B2 (en) | 2014-05-30 | 2016-08-30 | Apple Inc. | Exemplar-based natural language processing |
US9338493B2 (en) | 2014-06-30 | 2016-05-10 | Apple Inc. | Intelligent automated assistant for TV user interactions |
US9818400B2 (en) | 2014-09-11 | 2017-11-14 | Apple Inc. | Method and apparatus for discovering trending terms in speech requests |
US10074360B2 (en) | 2014-09-30 | 2018-09-11 | Apple Inc. | Providing an indication of the suitability of speech recognition |
US9668121B2 (en) | 2014-09-30 | 2017-05-30 | Apple Inc. | Social reminders |
US10127911B2 (en) | 2014-09-30 | 2018-11-13 | Apple Inc. | Speaker identification and unsupervised speaker adaptation techniques |
US10152299B2 (en) | 2015-03-06 | 2018-12-11 | Apple Inc. | Reducing response latency of intelligent automated assistants |
US9886953B2 (en) | 2015-03-08 | 2018-02-06 | Apple Inc. | Virtual assistant activation |
US10567477B2 (en) | 2015-03-08 | 2020-02-18 | Apple Inc. | Virtual assistant continuity |
US9721566B2 (en) | 2015-03-08 | 2017-08-01 | Apple Inc. | Competing devices responding to voice triggers |
US10460227B2 (en) | 2015-05-15 | 2019-10-29 | Apple Inc. | Virtual assistant in a communication session |
US10083688B2 (en) | 2015-05-27 | 2018-09-25 | Apple Inc. | Device voice control for selecting a displayed affordance |
US10200824B2 (en) | 2015-05-27 | 2019-02-05 | Apple Inc. | Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device |
US9578173B2 (en) | 2015-06-05 | 2017-02-21 | Apple Inc. | Virtual assistant aided communication with 3rd party service in a communication session |
US11025565B2 (en) | 2015-06-07 | 2021-06-01 | Apple Inc. | Personalized prediction of responses for instant messaging |
US20160378747A1 (en) | 2015-06-29 | 2016-12-29 | Apple Inc. | Virtual assistant for media playback |
US10747498B2 (en) | 2015-09-08 | 2020-08-18 | Apple Inc. | Zero latency digital assistant |
US10740384B2 (en) | 2015-09-08 | 2020-08-11 | Apple Inc. | Intelligent automated assistant for media search and playback |
US10331312B2 (en) | 2015-09-08 | 2019-06-25 | Apple Inc. | Intelligent automated assistant in a media environment |
US10671428B2 (en) | 2015-09-08 | 2020-06-02 | Apple Inc. | Distributed personal assistant |
US11587559B2 (en) | 2015-09-30 | 2023-02-21 | Apple Inc. | Intelligent device identification |
US9679497B2 (en) | 2015-10-09 | 2017-06-13 | Microsoft Technology Licensing, Llc | Proxies for speech generating devices |
US10262555B2 (en) | 2015-10-09 | 2019-04-16 | Microsoft Technology Licensing, Llc | Facilitating awareness and conversation throughput in an augmentative and alternative communication system |
US10148808B2 (en) * | 2015-10-09 | 2018-12-04 | Microsoft Technology Licensing, Llc | Directed personal communication for speech generating devices |
US10691473B2 (en) | 2015-11-06 | 2020-06-23 | Apple Inc. | Intelligent automated assistant in a messaging environment |
US10956666B2 (en) | 2015-11-09 | 2021-03-23 | Apple Inc. | Unconventional virtual assistant interactions |
US10049668B2 (en) | 2015-12-02 | 2018-08-14 | Apple Inc. | Applying neural network language models to weighted finite state transducers for automatic speech recognition |
US11087775B2 (en) * | 2015-12-11 | 2021-08-10 | Sony Corporation | Device and method of noise suppression based on noise source positions |
US10762245B2 (en) * | 2015-12-17 | 2020-09-01 | Ncr Corporation | Input peripheral device security |
US10223066B2 (en) | 2015-12-23 | 2019-03-05 | Apple Inc. | Proactive assistance based on dialog communication between devices |
US11227589B2 (en) | 2016-06-06 | 2022-01-18 | Apple Inc. | Intelligent list reading |
US10049663B2 (en) | 2016-06-08 | 2018-08-14 | Apple, Inc. | Intelligent automated assistant for media exploration |
US12223282B2 (en) | 2016-06-09 | 2025-02-11 | Apple Inc. | Intelligent automated assistant in a home environment |
US10586535B2 (en) | 2016-06-10 | 2020-03-10 | Apple Inc. | Intelligent digital assistant in a multi-tasking environment |
DK201670540A1 (en) | 2016-06-11 | 2018-01-08 | Apple Inc | Application integration with a digital assistant |
DK179415B1 (en) | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
US12197817B2 (en) | 2016-06-11 | 2025-01-14 | Apple Inc. | Intelligent device arbitration and control |
US10474753B2 (en) | 2016-09-07 | 2019-11-12 | Apple Inc. | Language identification using recurrent neural networks |
US10043516B2 (en) | 2016-09-23 | 2018-08-07 | Apple Inc. | Intelligent automated assistant |
US11281993B2 (en) | 2016-12-05 | 2022-03-22 | Apple Inc. | Model and ensemble compression for metric learning |
US11204787B2 (en) | 2017-01-09 | 2021-12-21 | Apple Inc. | Application integration with a digital assistant |
DK201770383A1 (en) | 2017-05-09 | 2018-12-14 | Apple Inc. | User interface for correcting recognition errors |
US10417266B2 (en) | 2017-05-09 | 2019-09-17 | Apple Inc. | Context-aware ranking of intelligent response suggestions |
US10726832B2 (en) | 2017-05-11 | 2020-07-28 | Apple Inc. | Maintaining privacy of personal information |
DK180048B1 (en) | 2017-05-11 | 2020-02-04 | Apple Inc. | MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION |
US10395654B2 (en) | 2017-05-11 | 2019-08-27 | Apple Inc. | Text normalization based on a data-driven learning network |
DK201770439A1 (en) | 2017-05-11 | 2018-12-13 | Apple Inc. | Offline personal assistant |
DK179496B1 (en) | 2017-05-12 | 2019-01-15 | Apple Inc. | USER-SPECIFIC Acoustic Models |
DK201770429A1 (en) | 2017-05-12 | 2018-12-14 | Apple Inc. | Low-latency intelligent automated assistant |
DK179745B1 (en) | 2017-05-12 | 2019-05-01 | Apple Inc. | SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT |
US11301477B2 (en) | 2017-05-12 | 2022-04-12 | Apple Inc. | Feedback analysis of a digital assistant |
DK201770432A1 (en) | 2017-05-15 | 2018-12-21 | Apple Inc. | Hierarchical belief states for digital assistants |
DK201770411A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | MULTI-MODAL INTERFACES |
DK201770431A1 (en) | 2017-05-15 | 2018-12-20 | Apple Inc. | Optimizing dialogue policy decisions for digital assistants using implicit feedback |
US10311144B2 (en) | 2017-05-16 | 2019-06-04 | Apple Inc. | Emoji word sense disambiguation |
US20180336275A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Intelligent automated assistant for media exploration |
US20180336892A1 (en) | 2017-05-16 | 2018-11-22 | Apple Inc. | Detecting a trigger of a digital assistant |
DK179560B1 (en) | 2017-05-16 | 2019-02-18 | Apple Inc. | Far-field extension for digital assistant services |
US10403278B2 (en) | 2017-05-16 | 2019-09-03 | Apple Inc. | Methods and systems for phonetic matching in digital assistant services |
US10657328B2 (en) | 2017-06-02 | 2020-05-19 | Apple Inc. | Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling |
US10445429B2 (en) | 2017-09-21 | 2019-10-15 | Apple Inc. | Natural language understanding using vocabularies with compressed serialized tries |
US10755051B2 (en) | 2017-09-29 | 2020-08-25 | Apple Inc. | Rule-based natural language processing |
US10636424B2 (en) | 2017-11-30 | 2020-04-28 | Apple Inc. | Multi-turn canned dialog |
US10733982B2 (en) | 2018-01-08 | 2020-08-04 | Apple Inc. | Multi-directional dialog |
US10733375B2 (en) | 2018-01-31 | 2020-08-04 | Apple Inc. | Knowledge-based framework for improving natural language understanding |
US10789959B2 (en) | 2018-03-02 | 2020-09-29 | Apple Inc. | Training speaker recognition models for digital assistants |
US10592604B2 (en) | 2018-03-12 | 2020-03-17 | Apple Inc. | Inverse text normalization for automatic speech recognition |
US10818288B2 (en) | 2018-03-26 | 2020-10-27 | Apple Inc. | Natural assistant interaction |
US10909331B2 (en) | 2018-03-30 | 2021-02-02 | Apple Inc. | Implicit identification of translation payload with neural machine translation |
US11145294B2 (en) | 2018-05-07 | 2021-10-12 | Apple Inc. | Intelligent automated assistant for delivering content from user experiences |
US10928918B2 (en) | 2018-05-07 | 2021-02-23 | Apple Inc. | Raise to speak |
US10984780B2 (en) | 2018-05-21 | 2021-04-20 | Apple Inc. | Global semantic word embeddings using bi-directional recurrent neural networks |
DK180639B1 (en) | 2018-06-01 | 2021-11-04 | Apple Inc | DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT |
US11386266B2 (en) | 2018-06-01 | 2022-07-12 | Apple Inc. | Text correction |
US10892996B2 (en) | 2018-06-01 | 2021-01-12 | Apple Inc. | Variable latency device coordination |
DK201870355A1 (en) | 2018-06-01 | 2019-12-16 | Apple Inc. | Virtual assistant operation in multi-device environments |
DK179822B1 (en) | 2018-06-01 | 2019-07-12 | Apple Inc. | Voice interaction at a primary device to access call functionality of a companion device |
US10496705B1 (en) | 2018-06-03 | 2019-12-03 | Apple Inc. | Accelerated task performance |
US11080121B2 (en) * | 2018-06-27 | 2021-08-03 | International Business Machines Corporation | Generating runbooks for problem events |
US11010561B2 (en) | 2018-09-27 | 2021-05-18 | Apple Inc. | Sentiment prediction from textual data |
US11170166B2 (en) | 2018-09-28 | 2021-11-09 | Apple Inc. | Neural typographical error modeling via generative adversarial networks |
US10839159B2 (en) | 2018-09-28 | 2020-11-17 | Apple Inc. | Named entity normalization in a spoken dialog system |
US11462215B2 (en) | 2018-09-28 | 2022-10-04 | Apple Inc. | Multi-modal inputs for voice commands |
US11475898B2 (en) | 2018-10-26 | 2022-10-18 | Apple Inc. | Low-latency multi-speaker speech recognition |
US11638059B2 (en) | 2019-01-04 | 2023-04-25 | Apple Inc. | Content playback on multiple devices |
US11348573B2 (en) | 2019-03-18 | 2022-05-31 | Apple Inc. | Multimodality in digital assistant systems |
US11307752B2 (en) | 2019-05-06 | 2022-04-19 | Apple Inc. | User configurable task triggers |
US11475884B2 (en) | 2019-05-06 | 2022-10-18 | Apple Inc. | Reducing digital assistant latency when a language is incorrectly determined |
DK201970509A1 (en) | 2019-05-06 | 2021-01-15 | Apple Inc | Spoken notifications |
US11423908B2 (en) | 2019-05-06 | 2022-08-23 | Apple Inc. | Interpreting spoken requests |
US11140099B2 (en) | 2019-05-21 | 2021-10-05 | Apple Inc. | Providing message response suggestions |
DK180129B1 (en) | 2019-05-31 | 2020-06-02 | Apple Inc. | USER ACTIVITY SHORTCUT SUGGESTIONS |
US11496600B2 (en) | 2019-05-31 | 2022-11-08 | Apple Inc. | Remote execution of machine-learned models |
US11289073B2 (en) | 2019-05-31 | 2022-03-29 | Apple Inc. | Device text to speech |
DK201970511A1 (en) | 2019-05-31 | 2021-02-15 | Apple Inc | Voice identification in digital assistant systems |
US11227599B2 (en) | 2019-06-01 | 2022-01-18 | Apple Inc. | Methods and user interfaces for voice-based control of electronic devices |
US11360641B2 (en) | 2019-06-01 | 2022-06-14 | Apple Inc. | Increasing the relevance of new available information |
US11488406B2 (en) | 2019-09-25 | 2022-11-01 | Apple Inc. | Text detection using global geometry estimators |
US11183193B1 (en) | 2020-05-11 | 2021-11-23 | Apple Inc. | Digital assistant hardware abstraction |
US11810578B2 (en) | 2020-05-11 | 2023-11-07 | Apple Inc. | Device arbitration for digital assistant-based intercom systems |
US11061543B1 (en) | 2020-05-11 | 2021-07-13 | Apple Inc. | Providing relevant data items based on context |
US11755276B2 (en) | 2020-05-12 | 2023-09-12 | Apple Inc. | Reducing description length based on confidence |
US11490204B2 (en) | 2020-07-20 | 2022-11-01 | Apple Inc. | Multi-device audio adjustment coordination |
US11438683B2 (en) | 2020-07-21 | 2022-09-06 | Apple Inc. | User identification using headphones |
US11128955B1 (en) | 2020-09-15 | 2021-09-21 | Motorola Solutions, Inc. | Method and apparatus for managing audio processing in a converged portable communication device |
US11620999B2 (en) | 2020-09-18 | 2023-04-04 | Apple Inc. | Reducing device processing of unintended audio |
US20220293109A1 (en) * | 2021-03-11 | 2022-09-15 | Google Llc | Device arbitration for local execution of automatic speech recognition |
Citations (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5553220A (en) * | 1993-09-07 | 1996-09-03 | Cirrus Logic, Inc. | Managing audio data using a graphics display controller |
US6496743B1 (en) * | 1998-04-17 | 2002-12-17 | International Business Machines Corporation | Method and apparatus for providing audio sample rate conversion within a data-processing system |
US20030187639A1 (en) * | 2002-03-26 | 2003-10-02 | Sbc Technology Resources, Inc. | Method and system for evaluating automatic speech recognition telephone services |
US20050180464A1 (en) * | 2002-10-01 | 2005-08-18 | Adondo Corporation | Audio communication with a computer |
US20060053011A1 (en) * | 2004-09-07 | 2006-03-09 | Lg Electronics Inc. | Baseband modem for speech recognition and mobile communication terminal using the same |
US7099825B1 (en) * | 2002-03-15 | 2006-08-29 | Sprint Communications Company L.P. | User mobility in a voice recognition environment |
US20060215856A1 (en) * | 2003-05-07 | 2006-09-28 | Rolf Meyer | Detection device |
US20070043563A1 (en) * | 2005-08-22 | 2007-02-22 | International Business Machines Corporation | Methods and apparatus for buffering data for use in accordance with a speech recognition system |
US20070054686A1 (en) * | 2005-07-15 | 2007-03-08 | Andrew Allen | Methods and apparatus for providing PTT data buffering support indications from mobile devices and PTT data buffering control by wireless networks |
US20070133826A1 (en) * | 2005-12-13 | 2007-06-14 | Theodore Burk | Digital microphone interface, audio codec and methods for use therewith |
US20070255435A1 (en) * | 2005-03-28 | 2007-11-01 | Sound Id | Personal Sound System Including Multi-Mode Ear Level Module with Priority Logic |
US20080019537A1 (en) * | 2004-10-26 | 2008-01-24 | Rajeev Nongpiur | Multi-channel periodic signal enhancement system |
US20080167868A1 (en) * | 2007-01-04 | 2008-07-10 | Dimitri Kanevsky | Systems and methods for intelligent control of microphones for speech recognition applications |
US20080300025A1 (en) * | 2007-05-31 | 2008-12-04 | Motorola, Inc. | Method and system to configure audio processing paths for voice recognition |
US20090007159A1 (en) * | 2007-06-30 | 2009-01-01 | Microsoft Corporation | Interfaces for digital media processing |
US20090081953A1 (en) * | 2005-04-30 | 2009-03-26 | Dan Tian | Method for Supporting Simultaneously Multiple-Path Bluetooth Audio Applications |
US20090287489A1 (en) * | 2008-05-15 | 2009-11-19 | Palm, Inc. | Speech processing for plurality of users |
US20100161340A1 (en) * | 2008-12-24 | 2010-06-24 | Plantronics, Inc. | Dynamic Codec Switching |
US20100169092A1 (en) * | 2008-11-26 | 2010-07-01 | Backes Steven J | Voice interface ocx |
US20110237217A1 (en) * | 2010-03-29 | 2011-09-29 | Motorola, Inc. | Method and apparatus for enhanced safety in a public safety communication system |
US20120072221A1 (en) * | 1999-04-12 | 2012-03-22 | Ben Franklin Patent Holding, Llc | Distributed voice user interface |
US20120148067A1 (en) * | 2008-12-05 | 2012-06-14 | Audioasics A/S | Wind noise detection method and system |
US20120166186A1 (en) * | 2010-12-23 | 2012-06-28 | Microsoft Corporation | Dual-Band Speech Encoding |
US20130132072A1 (en) * | 2011-11-21 | 2013-05-23 | Rajesh Pradhan | Engine for human language comprehension of intent and command execution |
US20140122378A1 (en) * | 2012-10-29 | 2014-05-01 | Qualcomm Incorporated | Rules engine as a platform for mobile applications |
US20140156269A1 (en) * | 2012-12-03 | 2014-06-05 | Lg Electronics Inc. | Portable device and method for providing voice recognition service |
US20140257813A1 (en) * | 2013-03-08 | 2014-09-11 | Analog Devices A/S | Microphone circuit assembly and system with speech recognition |
US20140270260A1 (en) * | 2013-03-13 | 2014-09-18 | Aliphcom | Speech detection using low power microelectrical mechanical systems sensor |
US20140331092A1 (en) * | 2013-05-02 | 2014-11-06 | Microsoft Corporation | Activity based sampling of diagnostics data |
US20140337036A1 (en) * | 2013-05-09 | 2014-11-13 | Dsp Group Ltd. | Low power activation of a voice activated device |
US20140358552A1 (en) * | 2013-05-31 | 2014-12-04 | Cirrus Logic, Inc. | Low-power voice gate for device wake-up |
US20140372122A1 (en) * | 2013-06-14 | 2014-12-18 | Mitsubishi Electric Research Laboratories, Inc. | Determining Word Sequence Constraints for Low Cognitive Speech Recognition |
US20150074524A1 (en) * | 2013-09-10 | 2015-03-12 | Lenovo (Singapore) Pte. Ltd. | Management of virtual assistant action items |
US9100735B1 (en) * | 2011-02-10 | 2015-08-04 | Dolby Laboratories Licensing Corporation | Vector noise cancellation |
US20150221307A1 (en) * | 2013-12-20 | 2015-08-06 | Saurin Shah | Transition from low power always listening mode to high power speech recognition mode |
US9112989B2 (en) * | 2010-04-08 | 2015-08-18 | Qualcomm Incorporated | System and method of smart audio logging for mobile devices |
US20150340042A1 (en) * | 2013-03-12 | 2015-11-26 | Nuance Communications, Inc. | Methods and apparatus for detecting a voice command |
Family Cites Families (57)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5632002A (en) | 1992-12-28 | 1997-05-20 | Kabushiki Kaisha Toshiba | Speech recognition interface system suitable for window systems and speech mail systems |
US6070140A (en) * | 1995-06-05 | 2000-05-30 | Tran; Bao Q. | Speech recognizer |
JPH10105191A (en) * | 1996-09-30 | 1998-04-24 | Toshiba Corp | Speech recognition device and microphone frequency characteristic converting method |
US5950157A (en) | 1997-02-28 | 1999-09-07 | Sri International | Method for establishing handset-dependent normalizing models for speaker recognition |
US6587867B1 (en) * | 1997-05-22 | 2003-07-01 | Mci Communications Corporation | Internet-based subscriber profile management of a communications system |
US5970446A (en) * | 1997-11-25 | 1999-10-19 | At&T Corp | Selective noise/channel/coding models and recognizers for automatic speech recognition |
US6499013B1 (en) | 1998-09-09 | 2002-12-24 | One Voice Technologies, Inc. | Interactive user interface using speech recognition and natural language processing |
US6449496B1 (en) * | 1999-02-08 | 2002-09-10 | Qualcomm Incorporated | Voice recognition user interface for telephone handsets |
US6574599B1 (en) | 1999-03-31 | 2003-06-03 | Microsoft Corporation | Voice-recognition-based methods for establishing outbound communication through a unified messaging system including intelligent calendar interface |
US6553345B1 (en) | 1999-08-26 | 2003-04-22 | Matsushita Electric Industrial Co., Ltd. | Universal remote control allowing natural language modality for television and multimedia searches and requests |
US6760704B1 (en) | 2000-09-29 | 2004-07-06 | Intel Corporation | System for generating speech and non-speech audio messages |
US7451085B2 (en) * | 2000-10-13 | 2008-11-11 | At&T Intellectual Property Ii, L.P. | System and method for providing a compensated speech recognition model for speech recognition |
US7457750B2 (en) * | 2000-10-13 | 2008-11-25 | At&T Corp. | Systems and methods for dynamic re-configurable speech recognition |
US20020082834A1 (en) | 2000-11-16 | 2002-06-27 | Eaves George Paul | Simplified and robust speech recognizer |
US20020087320A1 (en) * | 2000-12-29 | 2002-07-04 | Lee Victor Wai Leung | Computer-implemented fuzzy logic based data verification method and system |
US6704707B2 (en) | 2001-03-14 | 2004-03-09 | Intel Corporation | Method for automatically and dynamically switching between speech technologies |
US6985862B2 (en) | 2001-03-22 | 2006-01-10 | Tellme Networks, Inc. | Histogram grammar weighting and error corrective training of grammar weights |
GB0120672D0 (en) * | 2001-08-24 | 2001-10-17 | Mitel Knowledge Corp | Intermediate voice and DTMF detector device for improved speech recognition utilisation and penetration |
US7526431B2 (en) | 2001-09-05 | 2009-04-28 | Voice Signal Technologies, Inc. | Speech recognition using ambiguous or phone key spelling and/or filtering |
US7203644B2 (en) | 2001-12-31 | 2007-04-10 | Intel Corporation | Automating tuning of speech recognition systems |
US20030190037A1 (en) * | 2002-04-04 | 2003-10-09 | John Hruska | Method for automated dialing of telephone number prefixes |
US7570943B2 (en) | 2002-08-29 | 2009-08-04 | Nokia Corporation | System and method for providing context sensitive recommendations to digital services |
JP2005025381A (en) | 2003-06-30 | 2005-01-27 | Toshiba Corp | Electronic apparatus, and power control method |
US6983244B2 (en) * | 2003-08-29 | 2006-01-03 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for improved speech recognition with supplementary information |
US7664639B2 (en) * | 2004-01-14 | 2010-02-16 | Art Advanced Recognition Technologies, Inc. | Apparatus and methods for speech recognition |
US7136459B2 (en) * | 2004-02-05 | 2006-11-14 | Avaya Technology Corp. | Methods and apparatus for data caching to improve name recognition in large namespaces |
US7899671B2 (en) * | 2004-02-05 | 2011-03-01 | Avaya, Inc. | Recognition results postprocessor for use in voice recognition systems |
US20050204009A1 (en) | 2004-03-09 | 2005-09-15 | Devapratim Hazarika | System, method and computer program product for prioritizing messages |
US20050246325A1 (en) | 2004-04-30 | 2005-11-03 | Microsoft Corporation | Method and system for recording and accessing usage of an item in a computer system |
JP3827317B2 (en) * | 2004-06-03 | 2006-09-27 | 任天堂株式会社 | Command processing unit |
US20060009974A1 (en) | 2004-07-09 | 2006-01-12 | Matsushita Electric Industrial Co., Ltd. | Hands-free voice dialing for portable and remote devices |
US7580363B2 (en) | 2004-08-16 | 2009-08-25 | Nokia Corporation | Apparatus and method for facilitating contact selection in communication devices |
US8009678B2 (en) | 2005-03-17 | 2011-08-30 | Microsoft Corporation | System and method for generating a dynamic prioritized contact list |
US20070060114A1 (en) | 2005-09-14 | 2007-03-15 | Jorey Ramer | Predictive text completion for a mobile communication facility |
US8433041B2 (en) * | 2005-09-23 | 2013-04-30 | Bce Inc. | Method and system to enable touch-free incoming call handling and touch-free outgoing call origination |
US7689420B2 (en) * | 2006-04-06 | 2010-03-30 | Microsoft Corporation | Personalizing a context-free grammar using a dictation language model |
ATE405088T1 (en) | 2006-08-30 | 2008-08-15 | Research In Motion Ltd | METHOD, COMPUTER PROGRAM AND APPARATUS FOR CLEARLY IDENTIFYING A CONTACT IN A CONTACT DATABASE THROUGH A SINGLE VOICE UTTERANCE |
US8060373B2 (en) | 2007-03-21 | 2011-11-15 | At&T Intellectual Property I, L.P. | System and method of identifying contact information |
US7523012B2 (en) | 2007-07-03 | 2009-04-21 | Motorola, Inc. | Method and system for controlling a handheld electronic device |
US20090037413A1 (en) | 2007-07-30 | 2009-02-05 | Research In Motion Limited | Method and system for generating address lists |
US8374316B2 (en) | 2007-08-28 | 2013-02-12 | International Business Machines Corporation | System for recording spoken phone numbers during a voice call |
US20090228513A1 (en) | 2008-03-06 | 2009-09-10 | Nokia Corporation | Methods, apparatuses, and computer program products for modeling contact networks |
US8909361B2 (en) | 2008-06-19 | 2014-12-09 | Broadcom Corporation | Method and system for processing high quality audio in a hardware audio codec for audio transmission |
US8411603B2 (en) | 2008-06-19 | 2013-04-02 | Broadcom Corporation | Method and system for dual digital microphone processing in an audio CODEC |
US8600741B2 (en) * | 2008-08-20 | 2013-12-03 | General Motors Llc | Method of using microphone characteristics to optimize speech recognition performance |
US8626511B2 (en) | 2010-01-22 | 2014-01-07 | Google Inc. | Multi-dimensional disambiguation of voice commands |
US8694313B2 (en) * | 2010-05-19 | 2014-04-08 | Google Inc. | Disambiguation of contact information using historical data |
US9992745B2 (en) * | 2011-11-01 | 2018-06-05 | Qualcomm Incorporated | Extraction and analysis of buffered audio data using multiple codec rates each greater than a low-power processor rate |
US20130144618A1 (en) * | 2011-12-02 | 2013-06-06 | Liang-Che Sun | Methods and electronic devices for speech recognition |
KR20130133629A (en) * | 2012-05-29 | 2013-12-09 | 삼성전자주식회사 | Method and apparatus for executing voice command in electronic device |
US9142215B2 (en) * | 2012-06-15 | 2015-09-22 | Cypress Semiconductor Corporation | Power-efficient voice activation |
US9704486B2 (en) * | 2012-12-11 | 2017-07-11 | Amazon Technologies, Inc. | Speech recognition power management |
DE102013001219B4 (en) * | 2013-01-25 | 2019-08-29 | Inodyn Newmedia Gmbh | Method and system for voice activation of a software agent from a standby mode |
US10395651B2 (en) * | 2013-02-28 | 2019-08-27 | Sony Corporation | Device and method for activating with voice input |
US9349386B2 (en) * | 2013-03-07 | 2016-05-24 | Analog Device Global | System and method for processor wake-up based on sensor data |
US9703350B2 (en) * | 2013-03-15 | 2017-07-11 | Maxim Integrated Products, Inc. | Always-on low-power keyword spotting |
US20150031416A1 (en) | 2013-07-23 | 2015-01-29 | Motorola Mobility Llc | Method and Device For Command Phrase Validation |
-
2013
- 2013-12-27 US US14/142,234 patent/US20150031416A1/en not_active Abandoned
- 2013-12-27 US US14/142,190 patent/US20150032238A1/en not_active Abandoned
-
2014
- 2014-07-21 WO PCT/US2014/047448 patent/WO2015013201A2/en active Application Filing
-
2019
- 2019-12-04 US US16/703,003 patent/US11363128B2/en active Active
-
2022
- 2022-06-06 US US17/833,446 patent/US11876922B2/en active Active
-
2023
- 2023-12-12 US US18/537,386 patent/US20240137435A1/en active Pending
Patent Citations (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5553220A (en) * | 1993-09-07 | 1996-09-03 | Cirrus Logic, Inc. | Managing audio data using a graphics display controller |
US6496743B1 (en) * | 1998-04-17 | 2002-12-17 | International Business Machines Corporation | Method and apparatus for providing audio sample rate conversion within a data-processing system |
US20120072221A1 (en) * | 1999-04-12 | 2012-03-22 | Ben Franklin Patent Holding, Llc | Distributed voice user interface |
US7099825B1 (en) * | 2002-03-15 | 2006-08-29 | Sprint Communications Company L.P. | User mobility in a voice recognition environment |
US20030187639A1 (en) * | 2002-03-26 | 2003-10-02 | Sbc Technology Resources, Inc. | Method and system for evaluating automatic speech recognition telephone services |
US20050180464A1 (en) * | 2002-10-01 | 2005-08-18 | Adondo Corporation | Audio communication with a computer |
US20060215856A1 (en) * | 2003-05-07 | 2006-09-28 | Rolf Meyer | Detection device |
US20060053011A1 (en) * | 2004-09-07 | 2006-03-09 | Lg Electronics Inc. | Baseband modem for speech recognition and mobile communication terminal using the same |
US20080019537A1 (en) * | 2004-10-26 | 2008-01-24 | Rajeev Nongpiur | Multi-channel periodic signal enhancement system |
US20110250932A1 (en) * | 2005-03-28 | 2011-10-13 | Sound Id | Personal Sound System Including Multi-Mode Ear Level Module with Priority Logic |
US20070255435A1 (en) * | 2005-03-28 | 2007-11-01 | Sound Id | Personal Sound System Including Multi-Mode Ear Level Module with Priority Logic |
US20090081953A1 (en) * | 2005-04-30 | 2009-03-26 | Dan Tian | Method for Supporting Simultaneously Multiple-Path Bluetooth Audio Applications |
US20070054686A1 (en) * | 2005-07-15 | 2007-03-08 | Andrew Allen | Methods and apparatus for providing PTT data buffering support indications from mobile devices and PTT data buffering control by wireless networks |
US20070043563A1 (en) * | 2005-08-22 | 2007-02-22 | International Business Machines Corporation | Methods and apparatus for buffering data for use in accordance with a speech recognition system |
US20070133826A1 (en) * | 2005-12-13 | 2007-06-14 | Theodore Burk | Digital microphone interface, audio codec and methods for use therewith |
US20080167868A1 (en) * | 2007-01-04 | 2008-07-10 | Dimitri Kanevsky | Systems and methods for intelligent control of microphones for speech recognition applications |
US20080300025A1 (en) * | 2007-05-31 | 2008-12-04 | Motorola, Inc. | Method and system to configure audio processing paths for voice recognition |
US20090007159A1 (en) * | 2007-06-30 | 2009-01-01 | Microsoft Corporation | Interfaces for digital media processing |
US20090287489A1 (en) * | 2008-05-15 | 2009-11-19 | Palm, Inc. | Speech processing for plurality of users |
US20100169092A1 (en) * | 2008-11-26 | 2010-07-01 | Backes Steven J | Voice interface ocx |
US20120148067A1 (en) * | 2008-12-05 | 2012-06-14 | Audioasics A/S | Wind noise detection method and system |
US20100161340A1 (en) * | 2008-12-24 | 2010-06-24 | Plantronics, Inc. | Dynamic Codec Switching |
US20110237217A1 (en) * | 2010-03-29 | 2011-09-29 | Motorola, Inc. | Method and apparatus for enhanced safety in a public safety communication system |
US9112989B2 (en) * | 2010-04-08 | 2015-08-18 | Qualcomm Incorporated | System and method of smart audio logging for mobile devices |
US20120166186A1 (en) * | 2010-12-23 | 2012-06-28 | Microsoft Corporation | Dual-Band Speech Encoding |
US9100735B1 (en) * | 2011-02-10 | 2015-08-04 | Dolby Laboratories Licensing Corporation | Vector noise cancellation |
US20130132072A1 (en) * | 2011-11-21 | 2013-05-23 | Rajesh Pradhan | Engine for human language comprehension of intent and command execution |
US20140122378A1 (en) * | 2012-10-29 | 2014-05-01 | Qualcomm Incorporated | Rules engine as a platform for mobile applications |
US20140156269A1 (en) * | 2012-12-03 | 2014-06-05 | Lg Electronics Inc. | Portable device and method for providing voice recognition service |
US20140257813A1 (en) * | 2013-03-08 | 2014-09-11 | Analog Devices A/S | Microphone circuit assembly and system with speech recognition |
US20150340042A1 (en) * | 2013-03-12 | 2015-11-26 | Nuance Communications, Inc. | Methods and apparatus for detecting a voice command |
US20140270260A1 (en) * | 2013-03-13 | 2014-09-18 | Aliphcom | Speech detection using low power microelectrical mechanical systems sensor |
US20140331092A1 (en) * | 2013-05-02 | 2014-11-06 | Microsoft Corporation | Activity based sampling of diagnostics data |
US20140337036A1 (en) * | 2013-05-09 | 2014-11-13 | Dsp Group Ltd. | Low power activation of a voice activated device |
US20140358552A1 (en) * | 2013-05-31 | 2014-12-04 | Cirrus Logic, Inc. | Low-power voice gate for device wake-up |
US20140372122A1 (en) * | 2013-06-14 | 2014-12-18 | Mitsubishi Electric Research Laboratories, Inc. | Determining Word Sequence Constraints for Low Cognitive Speech Recognition |
US20150074524A1 (en) * | 2013-09-10 | 2015-03-12 | Lenovo (Singapore) Pte. Ltd. | Management of virtual assistant action items |
US20150221307A1 (en) * | 2013-12-20 | 2015-08-06 | Saurin Shah | Transition from low power always listening mode to high power speech recognition mode |
Non-Patent Citations (2)
Title |
---|
AES 42 specification; c2006 * |
AES3 wikipedia page from archive.org dated 8/2012 * |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11557299B2 (en) * | 2014-10-09 | 2023-01-17 | Google Llc | Hotword detection on multiple devices |
US12254884B2 (en) * | 2014-10-09 | 2025-03-18 | Google Llc | Hotword detection on multiple devices |
US20240169992A1 (en) * | 2014-10-09 | 2024-05-23 | Google Llc | Hotword detection on multiple devices |
US11915706B2 (en) * | 2014-10-09 | 2024-02-27 | Google Llc | Hotword detection on multiple devices |
US20210118448A1 (en) * | 2014-10-09 | 2021-04-22 | Google Llc | Hotword Detection on Multiple Devices |
US9799337B2 (en) * | 2016-03-02 | 2017-10-24 | Merry Electronics(Suzhou) Co., Ltd. | Microphone apparatus for enhancing power conservation |
US20180069815A1 (en) * | 2016-09-02 | 2018-03-08 | Bose Corporation | Application-based messaging system using headphones |
US10754804B2 (en) | 2018-01-25 | 2020-08-25 | Samsung Electronics Co., Ltd. | Application processor for low power operation, electronic device including the same and method of operating the same |
US11216399B2 (en) | 2018-01-25 | 2022-01-04 | Samsung Electronics Co., Ltd. | Application processor for low power operation, electronic device including the same and method of operating the same |
KR102453656B1 (en) | 2018-01-25 | 2022-10-12 | 삼성전자주식회사 | Application processor for low power operation, electronic device including the same and method of operating the same |
KR20190090532A (en) * | 2018-01-25 | 2019-08-02 | 삼성전자주식회사 | Application processor for low power operation, electronic device including the same and method of operating the same |
CN112652317A (en) * | 2019-10-09 | 2021-04-13 | 深圳云程科技有限公司 | High-sampling-rate audio playing method and system based on Android system |
US20210110824A1 (en) * | 2019-10-10 | 2021-04-15 | Samsung Electronics Co., Ltd. | Electronic apparatus and controlling method thereof |
US12008988B2 (en) * | 2019-10-10 | 2024-06-11 | Samsung Electronics Co., Ltd. | Electronic apparatus and controlling method thereof |
CN112817387A (en) * | 2019-11-18 | 2021-05-18 | 英业达科技有限公司 | Radio component |
Also Published As
Publication number | Publication date |
---|---|
WO2015013201A3 (en) | 2015-03-19 |
US11363128B2 (en) | 2022-06-14 |
US20150031416A1 (en) | 2015-01-29 |
US11876922B2 (en) | 2024-01-16 |
WO2015013201A2 (en) | 2015-01-29 |
US20220303377A1 (en) | 2022-09-22 |
US20240137435A1 (en) | 2024-04-25 |
US20200106872A1 (en) | 2020-04-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11363128B2 (en) | Method and device for audio input routing | |
US10510337B2 (en) | Method and device for voice recognition training | |
US20110053506A1 (en) | Methods and Devices for Controlling Particular User Interface Functions of a Mobile Communication Device in a Vehicle | |
US9225815B2 (en) | Method and device for loudness level determination | |
WO2015074567A1 (en) | Clicking control method and terminal | |
CN104902116B (en) | A kind of time unifying method and device of voice data and reference signal | |
KR20170132240A (en) | Communication message transmission method and device | |
US20140278392A1 (en) | Method and Apparatus for Pre-Processing Audio Signals | |
WO2021036711A1 (en) | Network control method and related product | |
WO2014105398A1 (en) | Recommendations for applications based on device context | |
CN108683799A (en) | wearable device searching method and related device | |
US20170336848A1 (en) | Information processing apparatus, information processing system, and information processing method | |
CN107888765B (en) | Method for switching scene mode and mobile terminal | |
WO2017215635A1 (en) | Sound effect processing method and mobile terminal | |
CN112997471B (en) | Audio channel switching method and device, readable storage medium and electronic equipment | |
CN108881778B (en) | A wearable device-based video output method and wearable device | |
JP2022546542A (en) | Communication method, communication device, communication system, server and computer program | |
CN107145275A (en) | Method for quickly operating application program and smart watch | |
CN106793010B (en) | Network access method and equipment | |
CN113254092B (en) | Processing method, apparatus and storage medium | |
CN113253826A (en) | Control method, control device, terminal and storage medium | |
JP7404389B2 (en) | Movement trajectory recording method and related devices | |
WO2015067169A1 (en) | Method, apparatus and terminal device for obtaining call log | |
CN106713668B (en) | Method and device for controlling vibration of terminal | |
US20230300549A1 (en) | Audio playing method and device, storage medium, and mobile terminal |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MOTOROLA MOBILITY LLC, ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ONDO, KAZUHIRO;LABOWICZ, MICHAEL P;YOSHINO, HIDEKI;SIGNING DATES FROM 20131114 TO 20131211;REEL/FRAME:031854/0198 |
|
AS | Assignment |
Owner name: GOOGLE TECHNOLOGY HOLDINGS LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA MOBILITY LLC;REEL/FRAME:034500/0001 Effective date: 20141028 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |