[go: up one dir, main page]

CN110111786A - A kind of audio-frequency processing method, equipment and computer readable storage medium - Google Patents

A kind of audio-frequency processing method, equipment and computer readable storage medium Download PDF

Info

Publication number
CN110111786A
CN110111786A CN201910360041.4A CN201910360041A CN110111786A CN 110111786 A CN110111786 A CN 110111786A CN 201910360041 A CN201910360041 A CN 201910360041A CN 110111786 A CN110111786 A CN 110111786A
Authority
CN
China
Prior art keywords
wearable device
audio processing
audio
grip
receiving
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910360041.4A
Other languages
Chinese (zh)
Other versions
CN110111786B (en
Inventor
夏小聪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xi'an Zhidian Decibel Information Technology Co.,Ltd.
Original Assignee
Nubia Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nubia Technology Co Ltd filed Critical Nubia Technology Co Ltd
Priority to CN201910360041.4A priority Critical patent/CN110111786B/en
Publication of CN110111786A publication Critical patent/CN110111786A/en
Application granted granted Critical
Publication of CN110111786B publication Critical patent/CN110111786B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

This application discloses a kind of audio-frequency processing method, equipment and computer readable storage mediums, wherein this method comprises: determining the wearing state of the wearable device;Then, under the wearing state, grip of wearable device during carrying out voice transmitting-receiving is identified;Subsequently, according to the grip and ear position region, the audio processing region of the voice transmitting-receiving is determined in the wearable device;Finally, audio processing components are arranged in the audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel.A kind of audio processing scheme of hommization is realized, so that user's receiving and transmitting voice when using wearable device is more convenient, the clarity of receiving and transmitting voice is improved, enhances user experience.

Description

A kind of audio-frequency processing method, equipment and computer readable storage medium
Technical field
This application involves mobile communication field more particularly to a kind of audio-frequency processing methods, equipment and computer-readable storage Medium.
Background technique
In the prior art, with the fast development of intelligent terminal, occur being different from the wearing of normal procedure intelligent mobile phone Formula equipment, for example, the wearable devices such as smartwatch or Intelligent bracelet.Since wearable device is compared to traditional intelligent hand Machine, the particularity such as soft and hardware environment, mode of operation and operating environment, if the manipulation scheme of traditional smart phone is turned With to wearable device, then may make troubles place, user experience to the operation of user are bad.
Summary of the invention
In order to solve above-mentioned technological deficiency in the prior art, the invention proposes a kind of audio-frequency processing method, this method Include:
Determine the wearing state of the wearable device;
Under the wearing state, grip of wearable device during carrying out voice transmitting-receiving is identified;
According to the grip and ear position region, the voice transmitting-receiving is determined in the wearable device Audio processing region;
Audio processing components are set in the audio processing region, wherein the audio processing components include vibrating diaphragm, line Circle and magnet steel.
Optionally, the wearing state of the determination wearable device, comprising:
Obtain the connection status of the wearable device buckling component;
If the buckling component is in connection status, it is determined that the wearable device is in wearing state, if the button Seaming element is in an off state, it is determined that the wearable device is in disengaged position.
Optionally, the wearing state of the determination wearable device, further includes:
Under the wearing state, according to the closed position of buckling component and the pressure sensitivity value of fastening, current pendant is determined Wear position;
Under the disengaged position, according to the screen trigger area of the wearable device and/or edge pressure sensitivity region, really Gripping area before settled.
Optionally, described under the wearing state, identify the wearable device carry out voice transmitting-receiving during Grip, comprising:
Obtain the operational order of the voice transmitting-receiving;
Grip of the wearable device under current state is identified according to the operational order, wherein described to hold Holding gesture includes the first grip under wearing state, and the second grip in a disengaged condition.
Optionally, described under the wearing state, identify the wearable device carry out voice transmitting-receiving during Grip, further includes:
During the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained;
First audio processing ginseng of the gesture voice transmitting-receiving under first grip is adjusted according to the kinematic parameter Number is joined alternatively, adjusting the second audio processing of the gesture voice transmitting-receiving under second grip according to the kinematic parameter Number.
Optionally, described according to the grip and ear position region, institute is determined in the wearable device State the audio processing region of voice transmitting-receiving, comprising:
By the ultrasonic module and/or image recognition component of the wearable device, the ear position of voice transmitting-receiving person is identified Set region;
In the first grip state, determine that the wearable device is associated with the ear position region Region, as the audio processing region, alternatively, in the second grip state, determine the wearable device with The associated region in ear position region, as the audio processing region.
Optionally, described that audio processing components are set in the audio processing region, wherein the audio processing components Including vibrating diaphragm, coil and magnet steel, comprising:
One or more groups of audio processing components are set in one or more audio processing regions;
The audio processing components include at least two groups of coils and multiple groups magnet steel.
Optionally, described that audio processing components are set in the audio processing region, wherein the audio processing components Including vibrating diaphragm, coil and magnet steel, further includes:
The vibrating diaphragm, first coil, the second coil and magnet steel are successively arranged in the audio processing region;
When the audio processing components receive audible signal, second coil is triggered by the audible signal and is existed The first active force is generated in the magnetic field that the magnet steel generates, vibrating diaphragm is driven by first active force, generates acoustic signals;
When the vibrating diaphragm of the audio processing components receives the acoustic signals, the first coil is driven to generate vibration, Electromotive force is obtained by the vibration in the magnetic field, and converts the electric signal that quiets down for the electromotive force.
The invention also provides a kind of audio processing equipment, which includes:
Memory, processor and it is stored in the computer program that can be run on the memory and on the processor;
The step of computer program realizes method as described in any one of the above embodiments when being executed by the processor.
The invention also provides a kind of computer readable storage medium, audio is stored on the computer readable storage medium Processing routine, the audio processing program realize the step of audio-frequency processing method as described in any one of the above embodiments when being executed by processor Suddenly.
The beneficial effects of the present invention are pass through the wearing state of the determination wearable device;Then, in the wearing Under state, grip of wearable device during carrying out voice transmitting-receiving is identified;Subsequently, according to the gripping hand Gesture and ear position region determine the audio processing region of the voice transmitting-receiving in the wearable device;Finally, institute State setting audio processing components in audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel. A kind of audio processing scheme of hommization is realized, so that user's receiving and transmitting voice when using wearable device is more convenient, is mentioned The high clarity of receiving and transmitting voice, enhances user experience.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows and meets implementation of the invention Example, and be used to explain the principle of the present invention together with specification.
In order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, to embodiment or will show below There is attached drawing needed in technical description to be briefly described, it should be apparent that, for those of ordinary skill in the art Speech, without any creative labor, is also possible to obtain other drawings based on these drawings.
Fig. 1 is a kind of hardware structural diagram of embodiment of wearable device provided in an embodiment of the present invention;
Fig. 2 is a kind of hardware schematic of embodiment of wearable device provided by the embodiments of the present application;
Fig. 3 is a kind of hardware schematic of embodiment of wearable device provided by the embodiments of the present application;
Fig. 4 is a kind of hardware schematic of embodiment of wearable device provided by the embodiments of the present application;
Fig. 5 is a kind of hardware schematic of embodiment of wearable device provided by the embodiments of the present application;
Fig. 6 is the flow chart of audio-frequency processing method first embodiment of the present invention;
Fig. 7 is the flow chart of audio-frequency processing method second embodiment of the present invention;
Fig. 8 is the flow chart of audio-frequency processing method 3rd embodiment of the present invention;
Fig. 9 is the flow chart of audio-frequency processing method fourth embodiment of the present invention;
Figure 10 is the flow chart of the 5th embodiment of audio-frequency processing method of the present invention;
Figure 11 is the flow chart of audio-frequency processing method sixth embodiment of the present invention;
Figure 12 is the flow chart of the 7th embodiment of audio-frequency processing method of the present invention;
Figure 13 is the flow chart of the 8th embodiment of audio-frequency processing method of the present invention;
Figure 14 is the corresponding audio processing components structure chart of audio-frequency processing method of the present invention.
Specific embodiment
It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, it is not intended to limit the present invention.
In subsequent description, it is only using the suffix for indicating such as " module ", " component " or " unit " of element Be conducive to explanation of the invention, itself there is no a specific meaning.Therefore, " module ", " component " or " unit " can mix Ground uses.
The wearable device provided in the embodiment of the present invention includes that Intelligent bracelet, smartwatch and smart phone etc. move Dynamic terminal.With the continuous development of Screen Technology, the appearance of the screens form such as flexible screen, Folding screen, smart phone etc. is mobile eventually End can also be used as wearable device.The wearable device provided in the embodiment of the present invention may include: RF (Radio Frequency, radio frequency) unit, WiFi module, audio output unit, A/V (audio/video) input unit, sensor, display The components such as unit, user input unit, interface unit, memory, processor and power supply.
It will be illustrated by taking wearable device as an example in subsequent descriptions, referring to Fig. 1, its each implementation to realize the present invention A kind of hardware structural diagram of wearable device of example, which may include: RF (Radio Frequency, radio frequency) unit 101, WiFi module 102, audio output unit 103, A/V (audio/video) input unit 104, Sensor 105, display unit 106, user input unit 107, interface unit 108, memory 109, processor 110, Yi Ji electricity The components such as source 111.It will be understood by those skilled in the art that wearable device structure shown in Fig. 1 is not constituted to wearable The restriction of equipment, wearable device may include perhaps combining certain components or difference than illustrating more or fewer components Component layout.
It is specifically introduced below with reference to all parts of the Fig. 1 to wearable device:
Radio frequency unit 101 can be used for receiving and sending messages or communication process in, signal sends and receivees, specifically, radio frequency list Uplink information can be sent to base station by member 101, and after the downlink information that in addition can also be sent base station receives, being sent to can be worn The processor 110 for wearing equipment is handled, and base station can be to the downlink information that radio frequency unit 101 is sent and be sent out according to radio frequency unit 101 What the uplink information sent generated, it is also possible to actively push away to radio frequency unit 101 after the information update for detecting wearable device It send, for example, base station can penetrating to wearable device after detecting that geographical location locating for wearable device changes Frequency unit 101 sends the message informing of geographical location variation, and radio frequency unit 101, can should after receiving the message informing The processor 110 that message informing is sent to wearable device is handled, and it is logical that the processor 110 of wearable device can control the message Know on the display panel 1061 for being shown in wearable device;In general, radio frequency unit 101 include but is not limited to antenna, at least one Amplifier, transceiver, coupler, low-noise amplifier, duplexer etc..In addition, radio frequency unit 101 can also pass through channel radio Letter communicated with network and other equipment, specifically may include: by wireless communication with the server communication in network system, example Such as, wearable device can download file resource from server by wireless communication, for example can download and answer from server With program, after wearable device completes the downloading of a certain application program, if the corresponding file of the application program in server Resource updates, then the server can be by wireless communication to the message informing of wearable device push resource updates, to remind User is updated the application program.Any communication standard or agreement can be used in above-mentioned wireless communication, including but not limited to GSM (Global System of Mobile communication, global system for mobile communications), GPRS (General Packet Radio Service, general packet radio service), CDMA2000 (Code Division Multiple Access 2000, CDMA 2000), (Wideband Code Division Multiple Access, wideband code division are more by WCDMA Location), TD-SCDMA (Time Division-Synchronous Code Division Multiple Access, time division synchronous CDMA), (Frequency Division Duplexing-Long Term Evolution, frequency division duplex are long by FDD-LTE Phase evolution) and TDD-LTE (Time Division Duplexing-Long Term Evolution, time division duplex are drilled for a long time Into) etc..
In one embodiment, wearable device 100 can access existing communication network by insertion SIM card.
In another embodiment, wearable device 100 can be come real by the way that esim card (Embedded-SIM) is arranged Existing communication network is now accessed, by the way of esim card, the inner space of wearable device can be saved, reduce thickness.
It is understood that although Fig. 1 shows radio frequency unit 101, but it is understood that, radio frequency unit 101 its And it is not belonging to must be configured into for wearable device, it can according to need within the scope of not changing the essence of the invention and save completely Slightly., wearable device 100 can realize the communication connection with other equipment or communication network separately through wifi module 102, The embodiment of the present invention is not limited thereto.
WiFi belongs to short range wireless transmission technology, and wearable device can help user to receive and dispatch by WiFi module 102 Email, browsing webpage and access streaming video etc., it provides wireless broadband internet access for user.Although Fig. 1 WiFi module 102 is shown, but it is understood that, and it is not belonging to must be configured into for wearable device, it completely can root It is omitted within the scope of not changing the essence of the invention according to needs.
Audio output unit 103 can be in call signal reception pattern, call mode, record in wearable device 100 When under the isotypes such as mode, speech recognition mode, broadcast reception mode, by radio frequency unit 101 or WiFi module 102 it is received or The audio data that person stores in memory 109 is converted into audio signal and exports to be sound.Moreover, audio output unit 103 can also provide audio output relevant to the specific function that wearable device 100 executes (for example, call signal reception sound Sound, message sink sound etc.).Audio output unit 103 may include loudspeaker, buzzer etc..
A/V input unit 104 is for receiving audio or video signal.A/V input unit 104 may include graphics processor (Graphics Processing Unit, GPU) 1041 and microphone 1042, graphics processor 1041 is in video acquisition mode Or the image data of the static images or video obtained in image capture mode by image capture apparatus (such as camera) carries out Reason.Treated, and picture frame may be displayed on display unit 106.Through graphics processor 1041, treated that picture frame can be deposited Storage is sent in memory 109 (or other storage mediums) or via radio frequency unit 101 or WiFi module 102.Mike Wind 1042 can connect in telephone calling model, logging mode, speech recognition mode etc. operational mode via microphone 1042 Quiet down sound (audio data), and can be audio data by such acoustic processing.Audio that treated (voice) data can To be converted to the format output that can be sent to mobile communication base station via radio frequency unit 101 in the case where telephone calling model. Microphone 1042 can be implemented various types of noises elimination (or inhibition) algorithms and send and receive sound to eliminate (or inhibition) The noise generated during frequency signal or interference.
In one embodiment, wearable device 100 includes one or more cameras, by opening camera, It can be realized the capture to image, realize the functions such as take pictures, record a video, the position of camera, which can according to need, to be configured.
Wearable device 100 further includes at least one sensor 105, for example, optical sensor, motion sensor and other Sensor.Specifically, optical sensor includes ambient light sensor and proximity sensor, wherein ambient light sensor can be according to ring The light and shade of border light adjusts the brightness of display panel 1061, proximity sensor can when wearable device 100 is moved in one's ear, Close display panel 1061 and/or backlight.As a kind of motion sensor, accelerometer sensor can detect in all directions The size of (generally three axis) acceleration, can detect that size and the direction of gravity, can be used to identify mobile phone posture when static It (for example pedometer, is struck using (such as horizontal/vertical screen switching, dependent game, magnetometer pose calibrating), Vibration identification correlation function Hit) etc..
In one embodiment, wearable device 100 further includes proximity sensor, can by using proximity sensor Wearable device can be realized non-contact manipulation, provide more modes of operation.
In one embodiment, wearable device 100 further includes heart rate sensor, when wearing, by close to using Person can be realized the detecting of heart rate.
In one embodiment, wearable device 100 can also include that fingerprint sensor can by reading fingerprint Realize the functions such as safety verification.
Display unit 106 is for showing information input by user or being supplied to the information of user.Display unit 106 can wrap Display panel 1061 is included, liquid crystal display (Liquid Crystal Display, LCD), Organic Light Emitting Diode can be used Forms such as (Organic Light-Emitting Diode, OLED) configure display panel 1061.
In one embodiment, display panel 1061 uses flexible display screen, and wearable using flexible display screen sets For when wearing, screen is able to carry out bending, to more be bonded.Optionally, the flexible display screen can use OLED screen Body and graphene screen body, in other embodiments, the flexible display screen is also possible to other display materials, the present embodiment It is not limited thereto.
In one embodiment, the display panel 1061 of wearable device can take rectangle, ring when convenient for wearing Around.In other embodiments, other modes can also be taken.
User input unit 107 can be used for receiving the number or character information of input, and generate and wearable device User setting and the related key signals input of function control.Specifically, user input unit 107 may include touch panel 1071 And other input equipments 1072.Touch panel 1071, also referred to as touch screen collect the touch behaviour of user on it or nearby Make (for example user uses any suitable objects or attachment such as finger, stylus on touch panel 1071 or in touch panel Operation near 1071), and corresponding attachment device is driven according to preset formula.Touch panel 1071 may include touching Two parts of detection device and touch controller.Wherein, the touch orientation of touch detecting apparatus detection user, and detect touch behaviour Make bring signal, transmits a signal to touch controller;Touch controller receives touch information from touch detecting apparatus, and It is converted into contact coordinate, then gives processor 110, and order that processor 110 is sent can be received and executed.This Outside, touch panel 1071 can be realized using multiple types such as resistance-type, condenser type, infrared ray and surface acoustic waves.In addition to touching Panel 1071 is controlled, user input unit 107 can also include other input equipments 1072.Specifically, other input equipments 1072 It can include but is not limited to physical keyboard, function key (such as volume control button, switch key etc.), trace ball, mouse, operation One of bar etc. is a variety of, specifically herein without limitation.
In one embodiment, one or more buttons have can be set in the side of wearable device 100.Button can be with The various ways such as short-press, long-pressing, rotation are realized, to realize a variety of operating effects.The quantity of button can be different to be multiple It can be applied in combination between button, realize a variety of operating functions.
Further, touch panel 1071 can cover display panel 1061, when touch panel 1071 detect on it or After neighbouring touch operation, processor 110 is sent to determine the type of touch event, is followed by subsequent processing device 110 according to touch thing The type of part provides corresponding visual output on display panel 1061.Although in Fig. 1, touch panel 1071 and display panel 1061 be the function that outputs and inputs of realizing wearable device as two independent components, but in certain embodiments, Touch panel 1071 and display panel 1061 can be integrated and be realized the function that outputs and inputs of wearable device, specifically herein Without limitation.For example, processor 110 can be controlled when receiving the message informing of a certain application program by radio frequency unit 101 The message informing show in a certain predeterminable area of display panel 1061 by system, the predeterminable area and touch panel 1071 certain One region is corresponding, can be to corresponding to area on display panel 1061 by carrying out touch control operation to a certain region of touch panel 1071 The message informing shown in domain is controlled.
Interface unit 108 be used as at least one external device (ED) connect with wearable device 100 can by interface.Example Such as, external device (ED) may include wired or wireless headphone port, external power supply (or battery charger) port, You Xianhuo Wireless data communications port, memory card port, the port for connecting the device with identification module, audio input/output (I/O) end Mouth, video i/o port, ear port etc..Interface unit 108 can be used for receiving the input from external device (ED) (for example, number It is believed that breath, electric power etc.) and the input received is transferred to one or more elements in wearable device 100 or can For transmitting data between wearable device 100 and external device (ED).
In one embodiment, wearable device 100 interface unit 108 using contact structure, by contact with Corresponding other equipment connection, realizes the functions such as charging, connection.Use contact can be with waterproof.
Memory 109 can be used for storing software program and various data.Memory 109 can mainly include storing program area The storage data area and, wherein storing program area can (such as the sound of application program needed for storage program area, at least one function Sound playing function, image player function etc.) etc.;Storage data area can store according to mobile phone use created data (such as Audio data, phone directory etc.) etc..In addition, memory 109 may include high-speed random access memory, it can also include non-easy The property lost memory, a for example, at least disk memory, flush memory device or other volatile solid-state parts.
Processor 110 is the control centre of wearable device, utilizes various interfaces and the entire wearable device of connection Various pieces, by running or execute the software program and/or module that are stored in memory 109, and call and be stored in Data in memory 109 execute the various functions and processing data of wearable device, to carry out to wearable device whole Monitoring.Processor 110 may include one or more processing units;Preferably, processor 110 can integrate application processor and modulation Demodulation processor, wherein the main processing operation system of application processor, user interface and application program etc., modulation /demodulation processing Device mainly handles wireless communication.It is understood that above-mentioned modem processor can not also be integrated into processor 110.
Wearable device 100 can also include the power supply 111 (such as battery) powered to all parts, it is preferred that power supply 111 can be logically contiguous by power-supply management system and processor 110, thus charged by power-supply management system realization management, The functions such as electric discharge and power managed.
Although Fig. 1 is not shown, wearable device 100 can also be including bluetooth module etc., and details are not described herein.It is wearable to set Standby 100, by bluetooth, can connect with other terminal devices, realize communication and the interaction of information.
Fig. 2-Fig. 4 is please referred to, is the structure under a kind of a kind of embodiment of wearable device provided in an embodiment of the present invention Schematic diagram.Wearable device in the embodiment of the present invention, including flexible screen.In wearable device expansion, flexible screen is in Strip;When wearable device is in wearing state, flexible screen bending is annular in shape.Fig. 2 and Fig. 3 show wearable device Structural schematic diagram when screen is unfolded, Fig. 4 show structural schematic diagram when wearable device screen-bending.
Based on above-mentioned each embodiment, it can be seen that if the equipment is wrist-watch, bracelet or wearable device When, the screen of the equipment can not overlay device watchband region, can also be with the watchband region of overlay device.Here, this Shen It please propose a kind of optional embodiment, in the present embodiment, the equipment for wrist-watch, bracelet or wearable can be set Standby, the equipment includes screen and interconnecting piece.The screen can be flexible screen, and the interconnecting piece can be watchband.It can Choosing, the screen of the equipment or the viewing area of screen can be partly or completely covered on the watchband of equipment.Such as Fig. 5 Shown, Fig. 5 is a kind of a kind of hardware schematic of embodiment of wearable device provided by the embodiments of the present application, the equipment Screen extends to two sides, and part is covered on the watchband of equipment.In other embodiments, the screen of the equipment can also be with It is all covered on the watchband of the equipment, the embodiment of the present application is not limited thereto.
Embodiment one
Fig. 6 is the flow chart of audio-frequency processing method first embodiment of the present invention.A kind of audio-frequency processing method, this method packet It includes:
S1, the wearing state for determining the wearable device;
S2, under the wearing state, identify grip of wearable device during carrying out voice transmitting-receiving;
S3, according to the grip and ear position region, determine that the voice is received in the wearable device The audio processing region of hair;
S4, audio processing components are set in the audio processing region, wherein the audio processing components include vibration Film, coil and magnet steel.
In the present embodiment, firstly, determining the wearing state of the wearable device;Then, under the wearing state, Identify grip of wearable device during carrying out voice transmitting-receiving;Subsequently, according to the grip and Ear position region determines the audio processing region of the voice transmitting-receiving in the wearable device;Finally, in the audio Audio processing components are set in processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel.
Optionally, in the present embodiment, wearable device has a variety of manners, for example, table chain fastens, absorption type Fasten etc., then led to by the way that sensory package is arranged in the one or both ends of fastening to identify the buckling state of wearable device Cross buckling state identification wearing state;
Optionally, in the present embodiment, according to the different use demand of wearable device user and different uses Scene determines the optimal audio transceiver location, for example, when user answers under the grip in conjunction with current grip Used grip is different when used grip is from hearing voice messages when making a phone call, in another example, user's It is different that left hand and right hand is respectively used to used grip when hearing voice messages;
Optionally, in the present embodiment, audio processing region is set in one or more positions of wearable device, it can be with Understand, due to the different use demands and different usage scenarios of user, receives and dispatches position with different optimal audios It sets, therefore, audio processing region is arranged in the one or more positions of corresponding setting;
Optionally, in the present embodiment, one group of audio processing group is at least set in one or more audio processing regions Part, wherein the audio processing components include vibrating diaphragm, coil and magnet steel.
The beneficial effect of the present embodiment is, passes through the wearing state of the determination wearable device;Then, it is worn described It wears under state, identifies grip of wearable device during carrying out voice transmitting-receiving;Subsequently, according to the gripping Gesture and ear position region determine the audio processing region of the voice transmitting-receiving in the wearable device;Finally, Audio processing components are set in the audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnetic Steel.A kind of audio processing scheme of hommization is realized, so that user's receiving and transmitting voice when using wearable device is more convenient, The clarity for improving receiving and transmitting voice, enhances user experience.
Embodiment two
Fig. 7 is the flow chart of audio-frequency processing method second embodiment of the present invention, based on the above embodiment, described in the determination The wearing state of wearable device, comprising:
S11, the connection status for obtaining the wearable device buckling component;
If S12, the buckling component are in connection status, it is determined that the wearable device is in wearing state, if institute It states buckling component to be in an off state, it is determined that the wearable device is in disengaged position.
In the present embodiment, firstly, obtaining the connection status of the wearable device buckling component;Then, if the button Seaming element is in connection status, it is determined that the wearable device is in wearing state, if the buckling component, which is in, disconnects shape State, it is determined that the wearable device is in disengaged position.
Optionally, by the way that pressure sensor module or sensing contact component is arranged in the one or both ends of fastening, to know The buckling state of other wearable device, then identifies wearing state by buckling state;
Optionally, by the way that warming sensory package is arranged in the one or both ends of fastening, to be worn according to shell temperature identification Wear the wearing state of formula equipment.
The beneficial effect of the present embodiment is, by the connection status for obtaining the wearable device buckling component;Then, If the buckling component is in connection status, it is determined that the wearable device is in wearing state, if at the buckling component In off-state, it is determined that the wearable device is in disengaged position.Realize a kind of humanized audio processing side Case improves the clarity of receiving and transmitting voice, enhances use so that user's receiving and transmitting voice when using wearable device is more convenient Family experience.
Embodiment three
Fig. 8 is the flow chart of audio-frequency processing method 3rd embodiment of the present invention, based on the above embodiment, described in the determination The wearing state of wearable device, further includes:
S13, under the wearing state, according to the closed position of buckling component and the pressure sensitivity value of fastening, determine current Wearing position;
S14, under the disengaged position, according to the screen trigger area of the wearable device and/or edge pressure sensitivity area Domain determines current gripping area.
In the present embodiment, firstly, under the wearing state, according to the closed position of buckling component and the pressure of fastening Inductance value determines current wearing position;Then, under the disengaged position, according to the screen trigger region of the wearable device Domain and/or edge pressure sensitivity region, determine current gripping area.
Optionally, under the disengaged position, according to preset one or more screen triggerings of the wearable device Region determines current gripping area;
Optionally, it under the disengaged position, according to the both sides of the edge pressure sensitivity region of the wearable device, determines current Gripping area.
The beneficial effect of the present embodiment is, by under the wearing state, according to the closed position of buckling component with And the pressure sensitivity value fastened, determine current wearing position;Then, under the disengaged position, according to the wearable device Screen trigger area and/or edge pressure sensitivity region, determine current gripping area.It realizes at the humanized audio of one kind Reason scheme improves the clarity of receiving and transmitting voice, enhances so that user's receiving and transmitting voice when using wearable device is more convenient User experience.
Example IV
Fig. 9 is the flow chart of audio-frequency processing method fourth embodiment of the present invention, based on the above embodiment, described to wear described It wears under state, identifies grip of wearable device during carrying out voice transmitting-receiving, comprising:
S21, the operational order for obtaining the voice transmitting-receiving;
S22, grip of the wearable device under current state is identified according to the operational order, wherein institute Stating grip includes the first grip under wearing state, and the second grip in a disengaged condition.
In the present embodiment, firstly, obtaining the operational order of the voice transmitting-receiving;Then, known according to the operational order Grip of the not described wearable device under current state, wherein the grip includes under wearing state One grip, and the second grip in a disengaged condition.
Optionally, grip of the wearable device under current state is identified according to the operational order, wherein The grip includes the first grip under wearing state, and the second grip in a disengaged condition, more Specifically, in buckling state, it is also possible to it is in disengaged position, due in this state, under, wearable device is still with the side of gripping Formula carries out voice transmitting-receiving and the gesture under such state is therefore attributed to second grip of the present embodiment.
The beneficial effect of the present embodiment is, by the operational order for obtaining the voice transmitting-receiving;Then, according to the behaviour Make grip of the wearable device described in instruction identification under current state, wherein the grip is included in wearing shape The first grip under state, and the second grip in a disengaged condition.Realize a kind of humanized audio Processing scheme improves the clarity of receiving and transmitting voice, increases so that user's receiving and transmitting voice when using wearable device is more convenient Strong user experience.
Embodiment five
Figure 10 is the flow chart of the 5th embodiment of audio-frequency processing method of the present invention, based on the above embodiment, described described Under wearing state, grip of wearable device during carrying out voice transmitting-receiving is identified, further includes:
S23, during the voice transmitting-receiving, obtain the kinematic parameter of the wearable device;
S24, gesture voice transmitting-receiving is adjusted at the first audio under first grip according to the kinematic parameter Parameter is managed, alternatively, adjusting gesture voice transmitting-receiving at the second audio under second grip according to the kinematic parameter Manage parameter.
In the present embodiment, firstly, during the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained; Then, first audio processing parameters of the gesture voice transmitting-receiving under first grip are adjusted according to the kinematic parameter, Alternatively, adjusting second audio processing parameters of the gesture voice transmitting-receiving under second grip according to the kinematic parameter.
Optionally, during the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained, wherein the movement Parameter includes the parameter sensing of three-axis gyroscope, acceleration sensing parameter etc.;
Optionally, first audio of the gesture voice transmitting-receiving under first grip is adjusted according to the kinematic parameter Processing parameter, alternatively, adjusting second audio of the gesture voice transmitting-receiving under second grip according to the kinematic parameter Processing parameter, wherein voice when the first audio processing parameters and the second audio processing parameters include voice transmitting-receiving plays ginseng Several and voice collecting parameter.
The beneficial effect of the present embodiment is, by during the voice transmitting-receiving, obtaining the wearable device Kinematic parameter;Then, first audio of the gesture voice transmitting-receiving under first grip is adjusted according to the kinematic parameter Processing parameter, alternatively, adjusting second audio of the gesture voice transmitting-receiving under second grip according to the kinematic parameter Processing parameter.A kind of humanized audio processing scheme is realized, so that user receives and dispatches language when using wearable device Sound is more convenient, improves the clarity of receiving and transmitting voice, enhances user experience.
Embodiment six
Figure 11 is the flow chart of audio-frequency processing method sixth embodiment of the present invention, based on the above embodiment, described according to institute Grip and ear position region are stated, the audio processing region of the voice transmitting-receiving is determined in the wearable device, Include:
S31, ultrasonic module and/or image recognition component by the wearable device identify the ear of voice transmitting-receiving person The portion band of position;
S32, in the first grip state, determine that the wearable device is related to the ear position region The region of connection alternatively, in the second grip state, determines described wearable set as the audio processing region Standby region associated with the ear position region, as the audio processing region.
In the present embodiment, firstly, by the ultrasonic module and/or image recognition component of the wearable device, identification The ear position region of voice transmitting-receiving person;Then, in the first grip state, the wearable device and institute are determined The associated region in ear position region is stated, as the audio processing region, alternatively, in the second grip state In, wearable device region associated with the ear position region is determined, as the audio processing region.
Optionally, by the ultrasonic module of the wearable device and/or image recognition component, voice transmitting-receiving person is identified Ear position region, wherein the ear position region of identification includes auricle region and/or ear canal region.
The beneficial effect of the present embodiment is, passes through the ultrasonic module and/or image recognition by the wearable device Component identifies the ear position region of voice transmitting-receiving person;Then, in the first grip state, the wearing is determined Formula equipment region associated with the ear position region, as the audio processing region, alternatively, being held described second In gesture state, wearable device region associated with the ear position region is determined, as the audio processing Region.A kind of humanized audio processing scheme is realized, so that receiving and transmitting voice is more when using wearable device by user Be it is convenient, improve the clarity of receiving and transmitting voice, enhance user experience.
Embodiment seven
Figure 12 is the flow chart of the 7th embodiment of audio-frequency processing method of the present invention, based on the above embodiment, described described Audio processing components are set in audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel, packet It includes:
S41, one or more groups of audio processing components are set in one or more audio processing regions;
S42, the audio processing components include at least two groups of coils and multiple groups magnet steel.
In the present embodiment, firstly, one or more groups of sounds are arranged in one or more audio processing regions Frequency processing component;Then, the audio processing components include at least two groups of coils and multiple groups magnet steel.
Optionally, one or more groups of audio processing components are set in one or more audio processing regions, Wherein, one or more groups of audio processing components of different zones share the one or more of above-mentioned coil and above-mentioned magnet steel Audio component.
The beneficial effect of the present embodiment is, by being arranged one group or more in one or more audio processing regions The group audio processing components;Then, the audio processing components include at least two groups of coils and multiple groups magnet steel.Realize one The humanized audio processing scheme of kind improves so that user's receiving and transmitting voice when using wearable device is more convenient The clarity of receiving and transmitting voice, enhances user experience.
Embodiment eight
Figure 13 is the flow chart of the 8th embodiment of audio-frequency processing method of the present invention, based on the above embodiment, described described Audio processing components are set in audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel, are gone back Include:
S43, the vibrating diaphragm, first coil, the second coil and magnet steel are successively arranged in the audio processing region;
S44, when the audio processing components receive audible signal, pass through the audible signal and trigger second line Circle generates the first active force in the magnetic field that the magnet steel generates, and drives vibrating diaphragm by first active force, generates sound wave letter Number;
S45, when the vibrating diaphragm of the audio processing components receives the acoustic signals, drive the first coil to generate vibration It is dynamic, electromotive force is obtained by the vibration in the magnetic field, and convert the electric signal that quiets down for the electromotive force.
In the present embodiment, firstly, the vibrating diaphragm, first coil, the second coil and magnet steel are in the audio processing area It successively arranges in domain;Then, when the audio processing components receive audible signal, by described in audible signal triggering Second coil generates the first active force in the magnetic field that the magnet steel generates, and drives vibrating diaphragm by first active force, generates Acoustic signals;Finally, driving the first coil to generate when the vibrating diaphragm of the audio processing components receives the acoustic signals Vibration obtains electromotive force by the vibration in the magnetic field, and converts the electric signal that quiets down for the electromotive force.
Figure 14 is the corresponding audio processing components structure chart of audio-frequency processing method of the present invention.When this equipment receives audio letter Number when, audio signal passes through coil 2, in magnet steel magnetic field, generates power, this power takes vibrating diaphragm to, compressed air, generates sound wave.Its In, the principle that quiets down of audio: when this equipment receives sound wave, the strength that sound wave generates is acted on vibrating diaphragm, causes diaphragm oscillations, Band moving winding 1 makees corresponding vibration, and coil 1 moves in magnet steel, generates electromotive force, and voice signal is transformed into electric signal.This equipment Coil 2 and the external processor 5 of coil 1, processor have the signal to coil 2, and the signal of coil 1 is handled respectively.
It is understood that this programme is sounding and the integrated technology that quiets down.It, can also since sound can be with air borne It is propagated with medium, not only draws the sound said from mouth.This programme is to draw from the sound that ear canal comes out to pass on to other side.Institute It is placed on inside ear with device (earpiece/microphone), not outside.It is not only on software and completely isolated outside is made an uproar in structure Sound.Therefore, this programme can clearly pass on the sound of user in noisy environment, and artificial intelligence product can be improved in this programme Phonetic recognization rate.
The beneficial effect of the present embodiment is, by the vibrating diaphragm, first coil, the second coil and magnet steel in the sound It successively arranges in frequency processing region;Then, when the audio processing components receive audible signal, pass through the audible signal It triggers second coil and generates the first active force in the magnetic field that the magnet steel generates, driven and shaken by first active force Film generates acoustic signals;Finally, driving described first when the vibrating diaphragm of the audio processing components receives the acoustic signals Coil generates vibration, obtains electromotive force by the vibration in the magnetic field, and convert the electric signal that quiets down for the electromotive force. A kind of humanized audio processing scheme is realized, so that receiving and transmitting voice is more just when using wearable device by user Victory improves the clarity of receiving and transmitting voice, enhances user experience.
Embodiment nine
The invention also provides a kind of audio processing equipment, which includes:
Memory, processor and it is stored in the computer program that can be run on the memory and on the processor;
The step of computer program realizes method as described in any one of the above embodiments when being executed by the processor.
Specifically, in the present embodiment, firstly, determining the wearing state of the wearable device;Then, in the wearing Under state, grip of wearable device during carrying out voice transmitting-receiving is identified;Subsequently, according to the gripping hand Gesture and ear position region determine the audio processing region of the voice transmitting-receiving in the wearable device;Finally, institute State setting audio processing components in audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnet steel.
Optionally, in the present embodiment, wearable device has a variety of manners, for example, table chain fastens, absorption type Fasten etc., then led to by the way that sensory package is arranged in the one or both ends of fastening to identify the buckling state of wearable device Cross buckling state identification wearing state;
Optionally, in the present embodiment, according to the different use demand of wearable device user and different uses Scene determines the optimal audio transceiver location, for example, when user answers under the grip in conjunction with current grip Used grip is different when used grip is from hearing voice messages when making a phone call, in another example, user's It is different that left hand and right hand is respectively used to used grip when hearing voice messages;
Optionally, in the present embodiment, audio processing region is set in one or more positions of wearable device, it can be with Understand, due to the different use demands and different usage scenarios of user, receives and dispatches position with different optimal audios It sets, therefore, audio processing region is arranged in the one or more positions of corresponding setting;
Optionally, in the present embodiment, one group of audio processing group is at least set in one or more audio processing regions Part, wherein the audio processing components include vibrating diaphragm, coil and magnet steel.
In the present embodiment, firstly, obtaining the connection status of the wearable device buckling component;Then, if the button Seaming element is in connection status, it is determined that the wearable device is in wearing state, if the buckling component, which is in, disconnects shape State, it is determined that the wearable device is in disengaged position.
Optionally, by the way that pressure sensor module or sensing contact component is arranged in the one or both ends of fastening, to know The buckling state of other wearable device, then identifies wearing state by buckling state;
Optionally, by the way that warming sensory package is arranged in the one or both ends of fastening, to be worn according to shell temperature identification Wear the wearing state of formula equipment.
In the present embodiment, firstly, under the wearing state, according to the closed position of buckling component and the pressure of fastening Inductance value determines current wearing position;Then, under the disengaged position, according to the screen trigger region of the wearable device Domain and/or edge pressure sensitivity region, determine current gripping area.
Optionally, under the disengaged position, according to preset one or more screen triggerings of the wearable device Region determines current gripping area;
Optionally, it under the disengaged position, according to the both sides of the edge pressure sensitivity region of the wearable device, determines current Gripping area.
In the present embodiment, firstly, obtaining the operational order of the voice transmitting-receiving;Then, known according to the operational order Grip of the not described wearable device under current state, wherein the grip includes under wearing state One grip, and the second grip in a disengaged condition.
Optionally, grip of the wearable device under current state is identified according to the operational order, wherein The grip includes the first grip under wearing state, and the second grip in a disengaged condition, more Specifically, in buckling state, it is also possible to it is in disengaged position, due in this state, under, wearable device is still with the side of gripping Formula carries out voice transmitting-receiving and the gesture under such state is therefore attributed to second grip of the present embodiment.
In the present embodiment, firstly, during the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained; Then, first audio processing parameters of the gesture voice transmitting-receiving under first grip are adjusted according to the kinematic parameter, Alternatively, adjusting second audio processing parameters of the gesture voice transmitting-receiving under second grip according to the kinematic parameter.
Optionally, during the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained, wherein the movement Parameter includes the parameter sensing of three-axis gyroscope, acceleration sensing parameter etc.;
Optionally, first audio of the gesture voice transmitting-receiving under first grip is adjusted according to the kinematic parameter Processing parameter, alternatively, adjusting second audio of the gesture voice transmitting-receiving under second grip according to the kinematic parameter Processing parameter, wherein voice when the first audio processing parameters and the second audio processing parameters include voice transmitting-receiving plays ginseng Several and voice collecting parameter.
In the present embodiment, firstly, by the ultrasonic module and/or image recognition component of the wearable device, identification The ear position region of voice transmitting-receiving person;Then, in the first grip state, the wearable device and institute are determined The associated region in ear position region is stated, as the audio processing region, alternatively, in the second grip state In, wearable device region associated with the ear position region is determined, as the audio processing region.
Optionally, by the ultrasonic module of the wearable device and/or image recognition component, voice transmitting-receiving person is identified Ear position region, wherein the ear position region of identification includes auricle region and/or ear canal region.
In the present embodiment, firstly, one or more groups of sounds are arranged in one or more audio processing regions Frequency processing component;Then, the audio processing components include at least two groups of coils and multiple groups magnet steel.
Optionally, one or more groups of audio processing components are set in one or more audio processing regions, Wherein, one or more groups of audio processing components of different zones share the one or more of above-mentioned coil and above-mentioned magnet steel Audio component.
In the present embodiment, firstly, the vibrating diaphragm, first coil, the second coil and magnet steel are in the audio processing area It successively arranges in domain;Then, when the audio processing components receive audible signal, by described in audible signal triggering Second coil generates the first active force in the magnetic field that the magnet steel generates, and drives vibrating diaphragm by first active force, generates Acoustic signals;Finally, driving the first coil to generate when the vibrating diaphragm of the audio processing components receives the acoustic signals Vibration obtains electromotive force by the vibration in the magnetic field, and converts the electric signal that quiets down for the electromotive force.
It is understood that this programme is sounding and the integrated technology that quiets down.It, can also since sound can be with air borne It is propagated with medium, not only draws the sound said from mouth.This programme is to draw from the sound that ear canal comes out to pass on to other side.Institute It is placed on inside ear with device (earpiece/microphone), not outside.It is not only on software and completely isolated outside is made an uproar in structure Sound.Therefore, this programme can clearly pass on the sound of user in noisy environment, and artificial intelligence product can be improved in this programme Phonetic recognization rate.
The beneficial effect of the present embodiment is, passes through the wearing state of the determination wearable device;Then, it is worn described It wears under state, identifies grip of wearable device during carrying out voice transmitting-receiving;Subsequently, according to the gripping Gesture and ear position region determine the audio processing region of the voice transmitting-receiving in the wearable device;Finally, Audio processing components are set in the audio processing region, wherein the audio processing components include vibrating diaphragm, coil and magnetic Steel.A kind of audio processing scheme of hommization is realized, so that user's receiving and transmitting voice when using wearable device is more convenient, The clarity for improving receiving and transmitting voice, enhances user experience.
Implement Folding display method, equipment and computer readable storage medium of the invention, it is described wearable by determination The wearing state of equipment;Then, under the wearing state, identify the wearable device carry out voice transmitting-receiving during Grip;Subsequently, according to the grip and ear position region, institute's predicate is determined in the wearable device The audio processing region of sound transmitting-receiving;Finally, audio processing components are arranged in the audio processing region, wherein the audio Processing component includes vibrating diaphragm, coil and magnet steel.The audio processing scheme for realizing a kind of hommization, so that user wears in use Receiving and transmitting voice is more convenient when wearing formula equipment, improves the clarity of receiving and transmitting voice, enhances user experience.
It should be noted that, in this document, the terms "include", "comprise" or its any other variant are intended to non-row His property includes, so that the process, method, article or the device that include a series of elements not only include those elements, and And further include other elements that are not explicitly listed, or further include for this process, method, article or device institute it is intrinsic Element.In the absence of more restrictions, the element limited by sentence "including a ...", it is not excluded that including being somebody's turn to do There is also other identical elements in the process, method of element, article or device.
The serial number of the above embodiments of the invention is only for description, does not represent the advantages or disadvantages of the embodiments.
Through the above description of the embodiments, those skilled in the art can be understood that above-described embodiment side Method can be realized by means of software and necessary general hardware platform, naturally it is also possible to by hardware, but in many cases The former is more preferably embodiment.Based on this understanding, technical solution of the present invention substantially in other words does the prior art The part contributed out can be embodied in the form of software products, which is stored in a storage medium In (such as ROM/RAM, magnetic disk, CD), including some instructions are used so that a terminal (can be mobile phone, computer, service Device, air conditioner or network equipment etc.) execute method described in each embodiment of the present invention.
The embodiment of the present invention is described with above attached drawing, but the invention is not limited to above-mentioned specific Embodiment, the above mentioned embodiment is only schematical, rather than restrictive, those skilled in the art Under the inspiration of the present invention, without breaking away from the scope protected by the purposes and claims of the present invention, it can also make very much Form, all of these belong to the protection of the present invention.

Claims (10)

1. a kind of audio-frequency processing method, which is characterized in that the described method includes:
Determine the wearing state of the wearable device;
Under the wearing state, grip of wearable device during carrying out voice transmitting-receiving is identified;
According to the grip and ear position region, the audio of the voice transmitting-receiving is determined in the wearable device Processing region;
Audio processing components are set in the audio processing region, wherein the audio processing components include vibrating diaphragm, coil with And magnet steel.
2. audio-frequency processing method according to claim 1, which is characterized in that the wearing of the determination wearable device State, comprising:
Obtain the connection status of the wearable device buckling component;
If the buckling component is in connection status, it is determined that the wearable device is in wearing state, if the fastening group Part is in an off state, it is determined that the wearable device is in disengaged position.
3. audio-frequency processing method according to claim 2, which is characterized in that the wearing of the determination wearable device State, further includes:
Under the wearing state, according to the closed position of buckling component and the pressure sensitivity value of fastening, current wearing position is determined It sets;
Under the disengaged position, according to the screen trigger area of the wearable device and/or edge pressure sensitivity region, determination is worked as Preceding gripping area.
4. audio-frequency processing method according to claim 3, which is characterized in that it is described under the wearing state, identify institute State grip of wearable device during carrying out voice transmitting-receiving, comprising:
Obtain the operational order of the voice transmitting-receiving;
Grip of the wearable device under current state is identified according to the operational order, wherein the gripping hand Gesture includes the first grip under wearing state, and the second grip in a disengaged condition.
5. audio-frequency processing method according to claim 4, which is characterized in that it is described under the wearing state, identify institute State grip of wearable device during carrying out voice transmitting-receiving, further includes:
During the voice transmitting-receiving, the kinematic parameter of the wearable device is obtained;
First audio processing parameters of the gesture voice transmitting-receiving under first grip are adjusted according to the kinematic parameter, or Person adjusts second audio processing parameters of the gesture voice transmitting-receiving under second grip according to the kinematic parameter.
6. audio-frequency processing method according to claim 5, which is characterized in that described according to the grip and ear The band of position determines the audio processing region of the voice transmitting-receiving in the wearable device, comprising:
By the ultrasonic module and/or image recognition component of the wearable device, the ear position area of voice transmitting-receiving person is identified Domain;
In the first grip state, wearable device area associated with the ear position region is determined Domain, as the audio processing region, alternatively, determining the wearable device and institute in the second grip state The associated region in ear position region is stated, as the audio processing region.
7. audio-frequency processing method according to claim 6, which is characterized in that described to be arranged in the audio processing region Audio processing components, wherein the audio processing components include vibrating diaphragm, coil and magnet steel, comprising:
One or more groups of audio processing components are set in one or more audio processing regions;
The audio processing components include at least two groups of coils and multiple groups magnet steel.
8. audio-frequency processing method according to claim 7, which is characterized in that described to be arranged in the audio processing region Audio processing components, wherein the audio processing components include vibrating diaphragm, coil and magnet steel, further includes:
The vibrating diaphragm, first coil, the second coil and magnet steel are successively arranged in the audio processing region;
When the audio processing components receive audible signal, second coil is triggered described by the audible signal The first active force is generated in the magnetic field that magnet steel generates, vibrating diaphragm is driven by first active force, generates acoustic signals;
When the vibrating diaphragm of the audio processing components receives the acoustic signals, drives the first coil to generate vibration, pass through Vibration in the magnetic field obtains electromotive force, and converts the electric signal that quiets down for the electromotive force.
9. a kind of audio processing equipment, which is characterized in that the equipment includes:
Memory, processor and it is stored in the computer program that can be run on the memory and on the processor;
The step such as method described in any item of the claim 1 to 8 is realized when the computer program is executed by the processor Suddenly.
10. a kind of computer readable storage medium, which is characterized in that be stored at audio on the computer readable storage medium Program is managed, such as audio processing described in any item of the claim 1 to 8 is realized when the audio processing program is executed by processor The step of method.
CN201910360041.4A 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium Active CN110111786B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910360041.4A CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910360041.4A CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN110111786A true CN110111786A (en) 2019-08-09
CN110111786B CN110111786B (en) 2023-11-10

Family

ID=67487736

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910360041.4A Active CN110111786B (en) 2019-04-30 2019-04-30 Audio processing method, equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN110111786B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117255283A (en) * 2023-11-16 2023-12-19 深圳市雅乐电子有限公司 Earphone, control method and control device thereof

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1324033A (en) * 2000-05-15 2001-11-28 张越苏 Wearable radio multimedia miniature computer with Chinese LINUX operation system
JP2005191766A (en) * 2003-12-25 2005-07-14 Seiko Instruments Inc Portable electronic device
CN103809866A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Operation mode switching method and electronic equipment
US20150230022A1 (en) * 2014-02-07 2015-08-13 Samsung Electronics Co., Ltd. Wearable electronic system
CN107409154A (en) * 2015-12-31 2017-11-28 深圳市柔宇科技有限公司 Flexible wearable device
WO2018095044A1 (en) * 2016-11-22 2018-05-31 华为技术有限公司 Audio input/output method and terminal device
WO2019053996A1 (en) * 2017-09-13 2019-03-21 ソニー株式会社 Headphone device
CN109561221A (en) * 2018-12-26 2019-04-02 努比亚技术有限公司 A kind of call control method, equipment and computer readable storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1324033A (en) * 2000-05-15 2001-11-28 张越苏 Wearable radio multimedia miniature computer with Chinese LINUX operation system
JP2005191766A (en) * 2003-12-25 2005-07-14 Seiko Instruments Inc Portable electronic device
CN103809866A (en) * 2012-11-13 2014-05-21 联想(北京)有限公司 Operation mode switching method and electronic equipment
US20150230022A1 (en) * 2014-02-07 2015-08-13 Samsung Electronics Co., Ltd. Wearable electronic system
CN107409154A (en) * 2015-12-31 2017-11-28 深圳市柔宇科技有限公司 Flexible wearable device
WO2018095044A1 (en) * 2016-11-22 2018-05-31 华为技术有限公司 Audio input/output method and terminal device
WO2019053996A1 (en) * 2017-09-13 2019-03-21 ソニー株式会社 Headphone device
CN109561221A (en) * 2018-12-26 2019-04-02 努比亚技术有限公司 A kind of call control method, equipment and computer readable storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王玉;张重阳;郑世宝;薛明;邵长彬;於家健;: "穿戴式多媒体系统", 电视技术, no. 08 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117255283A (en) * 2023-11-16 2023-12-19 深圳市雅乐电子有限公司 Earphone, control method and control device thereof
CN117255283B (en) * 2023-11-16 2024-04-05 深圳市雅乐电子有限公司 Earphone and control method thereof

Also Published As

Publication number Publication date
CN110111786B (en) 2023-11-10

Similar Documents

Publication Publication Date Title
CN109982179A (en) Audio frequency signal output, device, wearable device and storage medium
CN110350935A (en) Audio signal output control method, wearable device and readable storage medium storing program for executing
CN110109544A (en) Method of adjustment, wearable device and the readable storage medium storing program for executing of motor vibration amplitude
CN110362368A (en) Picture customization display methods, relevant device and the storage medium of wearable device
CN109799912A (en) A kind of display control method, equipment and computer readable storage medium
CN110198411A (en) Depth of field control method, equipment and computer readable storage medium during a kind of video capture
CN110086563A (en) A kind of method of controlling rotation, equipment and computer readable storage medium
CN110083331A (en) Wearable device play mode control method, device, wearable device and medium
CN110083289A (en) A kind of button display methods, wearable device and computer readable storage medium
CN110134312A (en) Control method, wearable device and computer readable storage medium
CN110177209A (en) A kind of video parameter regulation method, equipment and computer readable storage medium
CN110083513A (en) A kind of lamp effect control method, equipment and computer readable storage medium
CN110086929A (en) Breath screen display methods, mobile phone, wearable device and computer readable storage medium
CN110111786A (en) A kind of audio-frequency processing method, equipment and computer readable storage medium
CN110098851A (en) A kind of sound control method, wearable device and computer readable storage medium
CN110069102A (en) A kind of display area regulation method, equipment and computer readable storage medium
CN110177208A (en) A kind of association control method of video record, equipment and computer readable storage medium
CN110138965A (en) The method, apparatus and storage medium of a kind of wearable device and its number format
CN110213442A (en) Speech playing method, terminal and computer readable storage medium
CN110069136A (en) A kind of wearing state recognition methods, equipment and computer readable storage medium
CN110083332A (en) Audio switching method, wearable device and computer readable storage medium
CN110187770A (en) A kind of wrist strap acoustic control exchange method, equipment and computer readable storage medium
CN110448020A (en) A kind of wearable device
CN110191282A (en) A kind of acquisition parameters regulation method, equipment and computer readable storage medium
CN110109583A (en) A kind of display and setting method, equipment and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20230922

Address after: No. 3605, 5th Floor, Building 3, Micro World, No. 23 Century Avenue, Liuwu New District, Lhasa, Tibet Autonomous Region, 850000

Applicant after: Tibet Canyu Lunzhu Technology Co.,Ltd.

Address before: 518000 6-8 / F, 10-11 / F, 6 / F, 6-10 / F, C zone, Han's innovation building, No. 9018, Beihuan Avenue, high tech Zone, Nanshan District, Shenzhen City, Guangdong Province

Applicant before: NUBIA TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20240905

Address after: Room 605, Building 2, Eurasia International Phase II, No. 3939 Eurasia Avenue, Chanba Ecological Zone, Xi'an City, Shaanxi Province 710000

Patentee after: Xi'an Zhidian Decibel Information Technology Co.,Ltd.

Country or region after: China

Address before: No. 3605, 5th Floor, Building 3, Micro World, No. 23 Century Avenue, Liuwu New District, Lhasa, Tibet Autonomous Region, 850000

Patentee before: Tibet Canyu Lunzhu Technology Co.,Ltd.

Country or region before: China