[go: up one dir, main page]

CN108305603B - Sound effect processing method and equipment, storage medium, server and sound terminal thereof - Google Patents

Sound effect processing method and equipment, storage medium, server and sound terminal thereof Download PDF

Info

Publication number
CN108305603B
CN108305603B CN201710999163.9A CN201710999163A CN108305603B CN 108305603 B CN108305603 B CN 108305603B CN 201710999163 A CN201710999163 A CN 201710999163A CN 108305603 B CN108305603 B CN 108305603B
Authority
CN
China
Prior art keywords
audio
sample
sound effect
target
data packet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710999163.9A
Other languages
Chinese (zh)
Other versions
CN108305603A (en
Inventor
周大丰
吴丹
王朝勤
杨宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN201710999163.9A priority Critical patent/CN108305603B/en
Publication of CN108305603A publication Critical patent/CN108305603A/en
Application granted granted Critical
Publication of CN108305603B publication Critical patent/CN108305603B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0091Means for obtaining special acoustic effects

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Reverberation, Karaoke And Other Acoustics (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Stereophonic System (AREA)

Abstract

The embodiment of the invention discloses a sound effect processing method and equipment thereof, a storage medium, a server and a sound terminal, wherein the method comprises the following steps: when a playing instruction of a target audio is received, audio data of the target audio is obtained; acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio; synthesizing the audio data by adopting the target sound effect data packet; and outputting the audio data after the synthesis processing. By adopting the invention, the self-adaptive sound effect can be provided based on the information of the audio frequency, the sound effect most suitable for the audio content is built, and the intelligence of sound effect processing is improved.

Description

Sound effect processing method and equipment, storage medium, server and sound terminal thereof
Technical Field
The invention relates to the technical field of internet, in particular to a sound effect processing method and equipment, a storage medium, a server and a sound terminal.
Background
Sound effects refer to effects produced by sound, which are artifacts or accents imposed on the soundtrack to enhance the realism, atmosphere or dramatic message of the scene, for enhancing sound processing of the art or other content of a movie, video game, music or other media.
With the widespread application of smart devices, the requirement of users for listening sensation increases, and in order to enhance the multimedia audio experience, smart devices are generally configured with a variety of sound effect settings, such as digital equalization, reverberation effect, channel expansion, and the like. These sound effect settings provide more setting options to reach different demands of customers, but once set for, all music of broadcast all can use same sound effect to lead to can't provide the most suitable sound effect according to user's listening habits and song style, reduced the intelligence of sound effect processing.
Disclosure of Invention
The embodiment of the invention provides a sound effect processing method and equipment, a storage medium and a terminal thereof, which can provide self-adaptive sound effect based on audio information, build the sound effect most suitable for audio content and improve the intelligence of sound effect processing.
An aspect of an embodiment of the present invention provides a sound effect processing method, which may include:
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
synthesizing the audio data by adopting the target sound effect data packet; and
and outputting the audio data after the synthesis processing.
Optionally, before the receiving of the play instruction for the target audio, the method further includes:
acquiring sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring sample frequency information and sample tone information of the sample audio;
acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
acquiring a sample sound effect label of the sample audio, and adjusting octave characteristic points in a selected frequency range in the digital equalization processing curve based on an equal response curve and the sample sound effect label to obtain digital equalization parameters corresponding to each octave characteristic point;
and storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set.
Optionally, the storing the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, and the sample sound effect tag into a sound effect parameter set includes:
compressing the digital equalization parameters corresponding to each octave characteristic point and the reverberation parameters of the sample audios and storing the compressed digital equalization parameters and the reverberation parameters of the sample audios as sound effect data packets corresponding to the sample sound effect labels;
and storing the sample sound effect label and the sound effect data packet corresponding to the sample sound effect label into the sound effect parameter set.
Optionally, the acquiring the audio data of the target audio includes:
acquiring audio data and an audio tag of the target audio;
the obtaining of the target audio data packet corresponding to the target audio in the audio parameter set includes:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
Optionally, the obtaining of the target sound effect data packet corresponding to the audio tag in the sound effect parameter set includes:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
Optionally, the storing the digital equalization parameter, the reverberation parameter, and the sample sound effect tag corresponding to each octave feature point into a sound effect parameter set includes:
and storing the digital equalization parameters, the reverberation parameters, the sample frequency information, the sample tone color information and the sample sound effect labels corresponding to the octave characteristic points into a sound effect parameter set.
Optionally, the acquiring the audio data of the target audio includes:
acquiring audio data, target frequency information and target tone information of the target audio;
the obtaining of the target audio effect data packet corresponding to the target audio in the audio effect parameter set includes:
matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
and acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
Optionally, when the sound effect processing method is operated on the server side;
the outputting the audio data after the synthesis processing includes:
and sending the audio data after the synthesis processing to an audio terminal so that the audio terminal outputs the audio data after the synthesis processing.
Optionally, when the sound effect processing method is operated on the side of the sound terminal;
when a playing instruction of a target audio is received, acquiring audio data of the target audio, including:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
the acquiring a target audio effect data packet corresponding to the target audio in the audio effect parameter set, where the target audio effect data packet includes a digital equalization parameter of the target audio and a reverberation parameter of the target audio, and includes:
receiving a sound effect parameter set sent by the server;
and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio.
Another aspect of an embodiment of the present invention provides a sound processing device, which may include:
the information acquisition unit is used for acquiring audio data of a target audio when a playing instruction of the target audio is received;
the parameter acquisition unit is used for acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and the data output unit is used for synthesizing the audio data by adopting the target sound effect data packet and sending the audio data after the synthesis processing to the sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
Optionally, the apparatus further comprises:
the system comprises a sample information acquisition unit, a sample processing unit and a sample processing unit, wherein the sample information acquisition unit is used for acquiring a sample audio, acquiring a characteristic frequency response curve of the sample audio and acquiring sample frequency information and sample tone information of the sample audio;
a sample parameter obtaining unit, configured to obtain a digital equalization processing curve of the sample audio and a reverberation parameter of the sample audio based on the characteristic frequency response curve, the sample frequency information, and the sample timbre information;
the sample parameter adjusting unit is used for acquiring a sample sound effect label of the sample audio, and adjusting the octave characteristic points in the selected frequency range in the digital equalization processing curve based on the equal-response curve and the sample sound effect label to obtain digital equalization parameters corresponding to each octave characteristic point;
and the sample information storage unit is used for storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set.
Optionally, the sample information saving unit includes:
a data packet obtaining subunit, configured to compress the digital equalization parameter corresponding to each octave feature point and the reverberation parameter of the sample audio, and store the compressed digital equalization parameter and the reverberation parameter as a sound effect data packet corresponding to the sample sound effect tag;
and the information storage subunit is used for storing the sample sound effect label and the sound effect data packet corresponding to the sample sound effect label into the sound effect parameter set.
Optionally, the information obtaining unit is specifically configured to:
acquiring audio data and an audio tag of the target audio;
the parameter obtaining unit is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
Optionally, the parameter obtaining unit is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
Optionally, the data packet obtaining subunit is specifically configured to compress the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, the sample frequency information, and the sample tone information, and store the compressed digital equalization parameter, the reverberation parameter of the sample audio, the sample frequency information, and the sample tone information as the sound effect data packet corresponding to the sample sound effect tag.
Optionally, the information obtaining unit is specifically configured to:
acquiring audio data, target frequency information and target tone information of the target audio;
the parameter acquisition unit includes:
the similarity obtaining subunit is configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
and the data packet acquisition subunit is used for acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
In another aspect, an embodiment of the present invention provides a computer storage medium, where a plurality of instructions are stored, and the instructions are adapted to be loaded by a processor and execute the following steps:
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
An embodiment of the present invention further provides a server, which may include: a processor and a memory; wherein the memory stores a computer program adapted to be loaded by the processor and to perform the steps of:
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
An embodiment of the present invention further provides a sound processing device, which may include:
the information receiving unit is used for receiving audio data of the target audio sent by the server when a playing instruction of the target audio is received;
the parameter acquisition unit is used for receiving a sound effect parameter set sent by the server and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and the data output unit is used for synthesizing the audio data by adopting the target sound effect data packet and outputting the audio data after synthesis.
Optionally, the information receiving unit is specifically configured to:
receiving audio data and an audio tag of the target audio sent by a server;
the parameter obtaining unit is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
Optionally, the parameter obtaining unit is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
Optionally, the information receiving unit is specifically configured to:
receiving audio data, target frequency information and target tone color information of the target audio sent by a server;
the parameter acquisition unit includes:
the similarity obtaining subunit is configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
and the data packet acquisition subunit is used for acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
An embodiment of the present invention further provides a computer storage medium, where the computer storage medium stores a plurality of instructions, and the instructions are suitable for being loaded by a processor and executing the following steps:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
An embodiment of the present invention further provides an audio terminal, which may include: a processor and a memory; wherein the memory stores a computer program adapted to be loaded by the processor and to perform the steps of:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
In the embodiment of the invention, when a playing instruction of a target audio is received, the audio data of the target audio is obtained, and after a target audio effect data packet corresponding to the target audio is obtained from an audio effect parameter set, the audio data is synthesized by using the target audio effect data packet, and finally the audio data after synthesis is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic flow chart illustrating a sound effect processing method according to an embodiment of the present invention;
FIG. 2 is a flow chart illustrating another sound processing method according to an embodiment of the present invention;
FIG. 3 is an interface diagram of a characteristic frequency response curve according to an embodiment of the present invention;
FIG. 4 is an interface diagram of a digital equalization processing curve according to an embodiment of the present invention;
FIG. 5 is an interface schematic of an equal loudness curve provided by an embodiment of the present invention;
fig. 6 is a schematic interface diagram after dispersion of octave feature points according to an embodiment of the present invention;
FIG. 7 is a flow chart illustrating another sound processing method according to an embodiment of the present invention;
FIG. 8 is a flow chart illustrating another sound processing method according to an embodiment of the present invention;
FIG. 9 is a flowchart illustrating another sound processing method according to an embodiment of the present invention;
FIG. 10 is a flowchart illustrating another sound processing method according to an embodiment of the present invention;
FIG. 11 is a schematic structural diagram of an audio processing device according to an embodiment of the present invention;
FIG. 12 is a schematic structural diagram of another sound effect processing device according to an embodiment of the present invention;
fig. 13 is a schematic structural diagram of a sample information holding unit according to an embodiment of the present invention;
FIG. 14 is a schematic diagram of a parameter obtaining unit according to an embodiment of the present invention;
FIG. 15 is a schematic diagram of an audio processing device according to another embodiment of the present invention;
FIG. 16 is a schematic structural diagram of another parameter obtaining unit according to an embodiment of the present invention;
fig. 17 is a schematic structural diagram of a server according to an embodiment of the present invention;
fig. 18 is a schematic structural diagram of an audio terminal according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The sound effect processing method provided by the embodiment of the invention can be applied to a sound effect enhancement scene, and is characterized in that when a playing instruction of a target sound frequency is received, the sound data of the target sound frequency is obtained, and after a target sound effect data packet corresponding to the target sound frequency is obtained in a sound effect parameter set, the target sound effect data packet is adopted to carry out synthesis processing on the sound data, and finally the audio data after the synthesis processing is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
The sound effect processing method according to the embodiment of the present invention is executed by a computer program, and can be executed on a computer system of von Ruiman system based on a sound effect processing device. The sound effect processing equipment can comprise a server with functions of storage, calculation, sound effect synthesis and the like, and can also comprise sound effect processing terminal equipment such as a sound box, a tablet computer, a Personal Computer (PC), a smart phone, a palm computer and mobile internet equipment (MID).
The sound effect processing method provided by the embodiment of the invention will be described in detail below with reference to fig. 1 to 10.
Referring to fig. 1, a flow chart of a sound effect processing method according to an embodiment of the present invention is shown. As shown in fig. 1, the method of the embodiment of the present invention may include the following steps S101 to S103.
S101, when a playing instruction of a target audio is received, audio data of the target audio is obtained;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, i.e., voice, music, and other sounds, and in the embodiment of the present invention, the audio is music, which can be a single piece of music in a music player, or a piece of music in multimedia such as video, games, and electronic books. The target audio is music selected by the user from a plurality of pieces of music for output. For one audio, various information such as a music title, a singer's name, audio data, an album to which the user belongs, year of publication, total time, audio tags, and the like may be included. Wherein the audio data is an opaque binary stream represented by a series of non-semantic symbols, i.e. the content of the target audio. The audio tags can be different types of music styles such as art rock, punk, metal music or folk songs. Optionally, the audio may further include frequency information and tone information of the audio, where the frequency information and the tone information are spectral characteristics of the audio, that is, frequency domain characteristics of the audio signal.
In specific implementation, when the sound effect processing device receives a playing instruction for a target audio, audio information of the target audio is acquired, and audio data, an audio tag, frequency information, tone information and the like in the audio information are extracted. For example, when the sound effect processing device receives a play instruction of a play target audio "actor", audio data of the "actor" and an audio tag "popular" are acquired.
S102, acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital balance parameter of the target audio and a reverberation parameter of the target audio;
it is understood that the sound effect parameter set may include a plurality of sample audios, a sample sound effect tag of each sample audio, a sound effect data packet corresponding to each sample sound effect tag, sample frequency information of each sample audio, and sample timbre information, where the sound effect data packet may include a digital equalization parameter and a reverberation parameter; optionally, the sound effect parameter set may further include a sample sound effect tag set, where the sample sound effect tag set is a corresponding relationship between a sample sound effect tag and a sample audio tag, the sample sound effect tag includes different types of music such as rock, metalwork, folk song, and disco, and the sample audio tag corresponding to each type of sample sound effect tag is a different style of music under such a type of sample sound effect tag. For example, as shown in table 1, a type of sound effect parameter set, and table 2, a type of sample sound effect label set, if the sample sound effect label is "rock", the corresponding sample sound effect label may include "art rock, punk, Post rock, check, etc.
Specifically, the digital equalization parameter is a signal obtained by adjusting signals of each frequency band in the digital equalization processing curveAnd the digital equalization parameters correspond to the characteristic points of each octave one by one. The octave feature points are frequency bands obtained by dividing the discrete spectrum into individual frequency bands, and the upper limit frequency of each frequency band is doubled (i.e. the frequency band with the frequency interval of 2: 1) than the lower limit frequency, so as to obtain a plurality of octaves, and then the feature points are respectively taken in each octave. For example, if the frequency range of the signal is 63Hz to 16kHz, the signal can be divided into 8 octaves, which are 63Hz to 126Hz, 126Hz to 252Hz, 252Hz to 504Hz, 504Hz to 1.08kHz, 1.008kHz to 2.016kHz, 2.016kHz to 4.032kHz, 4.032kHz to 8.064kHz and 8.064Hz to 16kHz, and then at least one feature point is taken in each octave, and each feature point corresponds to one digital equalization parameter. Similarly, if two frequencies are inserted between the upper and lower limit frequencies of an octave, and the ratio between the 4 frequencies is the same (the ratio between the two adjacent frequencies is 1.33 times), the octave is divided into 3 frequency ranges, which is called 1/3 octaves. According to the calculation method, n times of octaves can be obtained, namely, the octaves are obtained by taking oc as a coefficient ((2) ^1/2) ^ n at the reference frequency f0Before and after each of L ═ f0Oc, then calculating power spectrum by segments and merging, wherein f0Is the reference frequency. The octave can be applied to the fields of 31-segment equalizers, sound pressure analysis, vibration reduction, noise reduction and the like.
Reverberation is generated because a sound-producing object emits sound waves, and the sound waves are reflected by contacting the surface of an obstacle through air, and due to the complexity of the real environment, the sound emitted by a sound source generates various echoes from various directions, and the sound is mixed to form the so-called reverberation. The reverberation parameters may include reverberation intensity, reverberation time, scattering degree, reverberation density, and the like. The sound with short reverberation time is dry and dry, the sound with too long reverberation time is not confused and loses a large amount of details, and the proper reverberation time can beautify the sound, cover the noise of musical instruments and enable the tone fusion to increase the loudness and the consistency of syllables. It should be noted that one target audio corresponds to one reverberation parameter.
In the concrete implementation, sound effect processing equipment acquires an audio label of a target audio, searches a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquires a sound effect data packet corresponding to the target sample sound effect label in a sound effect parameter set; optionally, a target sample sound effect tag to which the audio tag belongs is searched in a sample sound effect tag set, a target sound effect data packet corresponding to the target sample sound effect tag is obtained in the sound effect parameter set, and a digital equalization parameter of a target audio frequency and a reverberation parameter of the target audio frequency in the target sound effect data packet are read; optionally, the target frequency information and the target tone information of the target audio are matched with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, the matching similarity between the target frequency information and the sample tone information of each sound effect data packet and the sample frequency information and the sample tone information of each sound effect data packet is obtained after matching, the sample frequency information with the highest matching similarity and the sample sound effect data packet corresponding to the sample tone information are obtained, and the sample sound effect data packet is used as the target sound effect data packet.
For example, if the audio tag of the target audio is "heavy metal", the target sample audio tag "metallacy" to which the "heavy metal" belongs is determined in the sample audio tag set shown in table 2, and then the audio parameter of "metallacy" is determined to be "B1" in the audio parameter set shown in table 1.
TABLE 1
Sample sound effect label Sound effect parameters Frequency information Timbre information
Rocking and rolling device A1 A2 A3
Metal music B1 B2 B3
Folk song C1 C2 C3
Disco D1 D2 D3
TABLE 2
Figure BDA0001442990240000111
S103, synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
In the concrete implementation, the sound effect processing equipment is used for carrying out synthesis processing such as collection, transformation, filtering, estimation, enhancement, compression, identification and the like on the sound effect parameters and the audio data in the determined target sound effect data packet, so that the audio with the sound effect transformed is obtained, and then the audio is output.
When the sound effect processing equipment is a server, sending the audio subjected to sound effect conversion to a sound terminal so that the sound terminal can output the audio subjected to sound effect conversion; when the sound effect processing equipment is a sound terminal, the audio frequency after the sound effect is changed can be directly output and played.
In a feasible implementation mode, a DSP (digital Signal processing) digital sound effect processing system is arranged in the sound effect processing equipment, and comprises a digital Signal processor DSP, an audio A/D (analog/digital) and D/A (digital/analog), a RAM, a ROM and a peripheral processor, after a codec transmits 16-bit sampling data to the DSP each time, will cause the interrupt key to receive the interrupt, the DSP will store the received data into the input buffer of the system, meanwhile, the audio data stored in the buffer memory are respectively processed (such as transformation, filtering, estimation and the like), and are stored in the output buffer memory of the system after some transformation, the interrupt key output interrupt program can execute the number fetching from the output buffer at regular time, output in an analog mode through the coder and the decoder and then send to the sound terminal or play directly after passing through the power amplifier.
Optionally, dynamic gain processing and noise suppression are used in the synthesis process to ensure that no power overload or clipping distortion is generated.
Optionally, the ultra-high sampling rate of 96kHz is adopted in the synthesis process to ensure high-quality digital-to-analog conversion and high signal-to-noise ratio of over 90 dB.
In the embodiment of the invention, when a playing instruction of a target audio is received, the audio data of the target audio is obtained, and after a target audio effect data packet corresponding to the target audio is obtained from an audio effect parameter set, the audio data is synthesized by using the target audio effect data packet, and finally the audio data after synthesis is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
Referring to fig. 2, a flow chart of another sound effect processing method according to an embodiment of the invention is shown. As shown in fig. 2, the method of the embodiment of the present invention may include the following steps S201 to S207.
S201, collecting sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring sample frequency information and sample tone information of the sample audio;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, namely voice, music and other sounds, and in the embodiment of the present invention, the sample audio is music, and the sample audio is at least one piece of music selected by a user for storage among a plurality of pieces of music. For a sample audio, various information such as sample audio data, frequency information, timbre information, singer's name, album to which it belongs, year of publication, total time length, sample audio tag, etc. may be included. The sample audio tags may be different types of music styles such as art rock, punk, heavy metal, black metal, or folk song, the sample audio data is an opaque binary stream represented by a series of non-semantic symbols, that is, the content of the sample audio, and the sample frequency information and the sample tone color information are spectral characteristics of the sample audio (frequency domain characteristics of the sample audio signal).
The characteristic frequency response is a phenomenon that when a sample audio signal output at a constant voltage is connected to a system, the generated sound pressure increases or attenuates with the change of frequency, and a curve of the associated change relationship between the sound pressure and the frequency is called a characteristic frequency response curve, as shown in fig. 3, the abscissa is the frequency, and the ordinate is the sound pressure level.
In specific implementation, because the frequency range of human ear listening is 20Hz to 20kHz, after sample audio frequency of 20Hz to 20kHz is collected, the spectral characteristics in the appropriate frequency range (for example, 50Hz to 20kHz) in the sample audio frequency are extracted, and the characteristic frequency response curve of the sample audio frequency is drawn according to the extracted spectral characteristics.
S202, acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
it is understood that the principle of the digital equalization process is: the input signal "X" is used to establish the corresponding output signal "Y", where f () includes a function of the frequency "k" corresponding to "X". The functional expression corresponding to "X" is expanded, i.e., Y ═ g (k) ×, where g () varies as the digital equalization parameters are adjusted. In the embodiment of the present invention, the input signal "X" is sample frequency information, the sample tone information, and a characteristic frequency response curve, and the frequency "k" is a frequency value corresponding to the characteristic frequency response curve. Based on the processing principle described above, a digital equalization processing curve may be generated. As shown in fig. 4, the abscissa is frequency, the ordinate is digital equalization parameter, and the digital equalization processing curve is a graph of variation of the digital equalization parameter with frequency.
Meanwhile, based on the characteristic frequency response curve, the sample frequency information and the sample tone color information, reverberation parameters of the sample audio frequency can be obtained. The reverberation is generated because a sound-producing object emits sound waves, and the sound waves are reflected by contacting the surface of an obstacle through air, and due to the complexity of the real environment, the sound emitted by a sound source generates various echoes from various directions, and the sound is mixed to form the so-called reverberation. The first sound of the reverberation is the "direct sound", i.e. the source sound, called dry sound output in the effector, which is the sound emitted from the sound directly reaching the listener's ears, and is also a major component of the sound pressure level. The propagation attenuation of the sound pressure level is inversely proportional to the square of the distance, i.e. the distance is doubled and the sound pressure level is reduced by 6 dB. The subsequent obvious sounds which are relatively separated are called as ' early reflected sounds ', also called as near secondary reflected sounds ', the sounds which are emitted by the sound source and reach the ears of a listener after being reflected for 1-2 times by surrounding interfaces (walls, ceilings and floors), the reflected sounds which reach 50ms later than the direct sounds all belong to the range, the sounds are relatively large and relatively obvious, and the distance relation among the source sounds, the ears and the walls in the space can be reflected. Early reflected sound is characterized by the inability of the ear to distinguish it from direct sound, and can only be perceived as superimposed. Early reflection of sound is therefore beneficial to improve sound pressure level and clarity of sound. Its propagation attenuation is related to the sound absorption properties of the reflective interface. The last sound is a "reverberant sound" which is a multiple reflection sound that is more than 50ms later than the direct sound. For music, although the richness of the music can be increased by the reverberation sound, the clearness of the sound can be reduced while the richness of the music is increased, and if the richness of the music is too small, the sound can be dried and cannot be too large. The magnitude of the reverberant sound is directly related to the sound absorption characteristics of the surrounding interface.
The reverberation parameters may include reverberation intensity, reverberation time, scattering degree, reverberation density, and the like. The sound with short reverberation time is dry and dry, the sound with too long reverberation time is not confused and loses a large amount of details, and the proper reverberation time can beautify the sound, cover the noise of musical instruments and enable the tone fusion to increase the loudness and the consistency of syllables. It should be noted that one sample audio corresponds to one reverberation parameter.
S203, acquiring a sample sound effect label of the sample audio, and adjusting octave characteristic points in a selected frequency range in the digital equalization processing curve based on an equal-response curve and the sample sound effect label to obtain digital equalization parameters corresponding to each octave characteristic point;
specifically, the sample sound effect labels may include different types of music such as rock, metal music, folk song, and disco, and each type of sample sound effect label may include at least one sample audio label, for example, the sample sound effect label "rock" may include "art rock, punk, Post rock, and tomb" sample audio labels. The sample audio effect tag for obtaining the sample audio can be obtained by obtaining the sample audio effect tag for the sample audio based on the corresponding relationship between the sample audio effect tag and the sample audio tag as shown in fig. 2. For example, if the sample audio is "physalis alkekengi" and the sample audio label of the "physalis alkekengi" is "popular metal", then table 2 indicates that "popular metal" belongs to "metal music", and thus the sample sound effect label of the "physalis alkekengi" is "metal music".
The equal loudness curve is a curve describing the relation between sound pressure level and sound wave frequency under equal loudness, and is one of auditory characteristics. I.e. what sound pressure level the pure tone needs to reach at different frequencies in order to obtain a consistent auditory loudness for the listener. Fig. 5 shows a pure tone equal loudness curve with frequency on the abscissa and sound pressure level on the ordinate. The sound pressure is a change of atmospheric pressure caused by disturbance, that is, a residual pressure of atmospheric pressure, and is equivalent to a pressure change caused by superimposing a disturbance on the atmospheric pressure. The sound pressure level is denoted by the symbol SPL in dB, which is defined by taking the common logarithm of the ratio of the effective value p (e) of the sound pressure to be measured to the reference sound pressure p (ref), and multiplying by 20. In the figure, the sound pressure levels of different frequencies corresponding to each curve are different, but the loudness sensed by human ears is the same, and each curve represents the change relationship between the frequency and the sound pressure level under different loudness. From the family of equal loudness curves, it can be seen that the human ear is insensitive to high and low tones when the loudness is small, and becomes increasingly sensitive to high and low tones when the loudness is large.
The digital equalization parameters are gain values of signals after signals of each frequency band in a digital equalization processing curve are adjusted, and the digital equalization parameters correspond to the characteristic points of each octave one by one. The octave feature points are frequency bands obtained by dividing the discrete spectrum into individual frequency bands, and the upper limit frequency of each frequency band is doubled (i.e. the frequency band with the frequency interval of 2: 1) than the lower limit frequency, so as to obtain a plurality of octaves, and then the feature points are respectively taken in each octave. For example, if the frequency range of the signal is 63Hz to 16kHz, the signal can be divided into 8 octaves, 63Hz to 126Hz, 126Hz to 252Hz, 252Hz to 504Hz, 504Hz to 1.08kHz, 1.008kHz to 2.016kHz, 2.016kHz to 4.032kHz, 4.032kHz to 8.064kHz and 8.064Hz to 16kHz, and then the characteristic point is taken in each octave. If the number of the octave feature points is 1/24, dividing each octave into 24 frequency bands, taking each frequency band as a 1/24 octave feature point of the octave, corresponding each 1/24 octave feature point to a digital equalization parameter, and adjusting the digital equalization parameter corresponding to each 1/24 octave feature point, thereby determining a suitable parameter range for each 1/24 octave feature point.
In the specific implementation, a curve of a part of frequency range in the digital equalization processing curve is selected, the selected curve is divided into multiple octaves, then the characteristic points in each octave are adjusted based on the parameter information of the equal-response curve, and the digital equalization parameters of each octave characteristic point are ensured to change in the parameter range indicated by the sample sound effect label in the adjustment process, so that the digital equalization parameter range of each octave characteristic point is determined. For example, a curve in the range of 63Hz to 16kHz in the digital equalization processing curve shown in fig. 4 is divided into 8 octaves, then a feature point is taken for each octave to obtain a discrete octave feature point shown in fig. 6, and then the digital equalization parameter of each discrete octave feature point is adjusted based on the common digital equalization parameter corresponding to the sample sound effect label to determine the digital equalization parameter range of each octave feature point under the sample sound effect label.
S204, storing the digital equalization parameters corresponding to each octave characteristic point, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
it is understood that the digital equalization parameters, the reverberation parameters and the sample sound effect tags may be stored in the form of data packets in the sound effect parameter set, or in the form of a mapping table in the sound effect parameter set.
In a specific implementation manner of the embodiment of the present invention, the storing the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, and the sample audio tag into an audio parameter set may include the following steps, as shown in fig. 7:
s301, compressing the digital equalization parameters corresponding to each octave characteristic point and the reverberation parameters of the sample audio and storing the compressed digital equalization parameters and the reverberation parameters of the sample audio as the sound effect data packet corresponding to the sample sound effect label;
specifically, the data packet is a data storage format, and may include a plurality of fields, and each field may identify different information. And forming an audio effect data packet by compressing the digital equalization parameter and the reverberation parameter into a data field of the audio effect data packet. Each sound effect data packet comprises a plurality of groups of digital equalization parameters and a group of reverberation parameters. The multiple sets of digital equalization parameters include digital equalization parameters corresponding to the feature points of each octave, and each octave may include multiple feature points, where it should be noted that the multiple sets of digital equalization parameters and the reverberation parameters of the sample audio are parameters indicated by the sample audio label, that is, the digital equalization parameters of the sample audio and the reverberation parameters of the sample audio are different under different sample labels.
TABLE 3
Figure BDA0001442990240000161
For example, if the signal frequency range is 63Hz to 16kHz and is divided into 8 octaves, 9 feature points are taken in each octave, the digital equalization parameters corresponding to 9 feature points of the 1 st octave are a11 to a19, the digital equalization parameters corresponding to 9 feature points of the 2 nd octave are a21 to a29 …, the digital equalization parameters corresponding to 9 feature points of the 8 th octave are a81 to a89, and the reverberation parameter of the signal is B, then the digital equalization parameters of the multiple sets of sample audio stored in one sound effect data packet are a11 to a19, a21 to a29 … a81 to a89, and the reverberation parameter of the sample audio is B, as shown in table 3.
S302, storing the sample sound effect label and the sound effect data packet into the sound effect parameter set.
In the concrete implementation, the sample sound effect tag is added to a selected field, such as a header field, of the sound effect data packet, and then the sound effect data packet added with the sample sound effect tag is stored in a sound effect parameter set, or the sample sound effect tag and the sound effect data packet are correspondingly stored in the sound effect parameter set in a form of a mapping relation table.
S205, when a playing instruction of a target audio is received, acquiring audio data of the target audio;
it is to be understood that, in the embodiment of the present invention, the target audio is music, and the target audio is music selected by the user for output among pieces of music.
In a specific implementation manner of the embodiment of the present invention, when the sound effect processing device receives a play instruction for a target audio, audio information of the target audio is obtained, and audio data and an audio tag in the audio information are extracted. For example, when the sound effect processing device receives a play instruction for playing "actor", the audio data of "actor" and the audio tag "popular" are acquired.
In another specific implementation manner of the embodiment of the present invention, when the sound effect processing device receives a play instruction for a target audio, audio data, target frequency information, and target tone color information of the target audio are acquired. The target frequency information and the target timbre information are spectral characteristics of the target audio, that is, frequency domain characteristics of the target audio signal.
S206, acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
in a specific implementation manner of the embodiment of the present invention, the sound effect processing device obtains a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reads a digital equalization parameter of a target audio and a reverberation parameter of the target audio in the target sound effect data packet. The target sound effect data packet corresponding to the audio tag is acquired from the sound effect parameter set, and it can be understood that the target sound effect tag to which the audio tag belongs is searched from the sample sound effect tag set, and the target sound effect data packet corresponding to the target sound effect tag is acquired from the sound effect parameter set. That is to say, the sound effect processing device searches for a sample sound effect tag corresponding to the audio tag, and then can acquire the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the sound effect parameter set based on the corresponding relationship between the sample sound effect tag and the sound effect data packet. The sample audio tag corresponding to the audio tag is searched, and it can be understood that the sample audio tag matched with the keyword is found based on the keyword of the sample audio tag, and then the sample audio tag corresponding to the sample audio tag is obtained from the set sample audio tag set. The sample sound effect tag set may be stored as a subset in the sound effect parameter set, or may be stored as a separate set.
In another specific implementation manner of the embodiment of the present invention, the storing the digital equalization parameter, the reverberation parameter of the sample audio, and the sample audio tag corresponding to each octave feature point into an audio parameter set includes:
storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios, the sample frequency information, the sample tone color information and the sample sound effect labels into a sound effect parameter set; the obtaining of the target sound effect data packet corresponding to the target audio in the sound effect parameter set may include the following steps, as shown in fig. 8:
s401, matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
in specific implementation, when the keywords of the audio tag are unclear or include a plurality of audio tags, the audio data packet may be decompressed by the decompression unit, the decompressed sample frequency information and sample tone information are extracted, each group of sample frequency information and sample tone information in the audio parameter set is sequentially traversed, the target frequency information and the target tone information are respectively matched with the traversed sample frequency information and sample tone information, and matching similarity of each group of matched samples is obtained. Optionally, the matched matching similarity and the sound effect data packet may be cached correspondingly, the caching mode may be to add the matching similarity to a set field of the sound effect data packet, or to cache the matching similarity and the sound effect data packet in a form of a mapping table to a sound effect parameter set.
S402, obtaining sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
And S207, synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
In the concrete implementation, sound effect parameters and sound data in the determined target sound effect data packet are subjected to synthesis processing such as collection, transformation, filtering, estimation, enhancement, compression, identification and the like by sound effect processing equipment, so that the sound frequency after the sound effect is transformed is obtained, and the sound frequency is sent to a sound terminal for output. In a feasible implementation mode, a DSP digital sound effect processing system is arranged in the sound effect processing equipment, and comprises main devices of a digital signal processor DSP, an audio A/D (analog/digital) and a D/A (digital/analog), a RAM, a ROM and a peripheral processor, after a codec transmits 16-bit sampling data to the DSP each time, will cause the interrupt key to receive the interrupt, the DSP will store the received data into the input buffer of the system, meanwhile, the audio data stored in the buffer memory are respectively processed (such as transformation, filtering, estimation and the like), and are stored in the output buffer memory of the system after some transformation, and the interrupt key output interrupt program can execute the acquisition of data from the output buffer at regular time, output the data in an analog mode through the codec, send the data to the sound terminal, and play the data after passing through the power amplifier.
In the embodiment of the invention, firstly, an audio tag set carrying audio parameters is arranged on an audio processing device, when a playing instruction of a target audio is received, audio data of the target audio is obtained, a target audio data packet corresponding to the target audio is obtained in the audio parameter set, then the target audio data packet is adopted to carry out synthesis processing on the audio data, and finally, the audio data after the synthesis processing is sent to a sound terminal for outputting. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, the intelligence of sound effect processing is improved, and the brand-new personalized audio experience is provided.
Referring to fig. 9, a flow chart of another sound effect processing method according to an embodiment of the invention is shown. As shown in fig. 9, the method of the embodiment of the present invention may include the following steps S501 to S503.
S501, receiving audio data of a target audio sent by a server when a playing instruction of the target audio is received;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, i.e., voice, music, and other sounds, and in the embodiment of the present invention, the audio is music, which can be a single piece of music in a music player, or a piece of music in multimedia such as video, games, and electronic books. The target audio is music selected by the user from a plurality of pieces of music for output. For one audio, various information such as a music title, a singer's name, audio data, an album to which the user belongs, year of publication, total time, audio tags, and the like may be included. Wherein the audio data is an opaque binary stream represented by a series of non-semantic symbols, i.e. the content of the target audio. The audio tags can be different types of music styles such as art rock, punk, metal music or folk songs. Optionally, the audio may further include frequency information and tone information of the audio, where the frequency information and the tone information are spectral characteristics of the audio, that is, frequency domain characteristics of the audio signal.
In the concrete implementation, when a server receives a playing instruction aiming at a target audio, the server collects the audio information of the target audio, extracts the audio data, audio tags, frequency information, tone information and other information in the audio information and sends the information to sound effect processing equipment (a sound terminal); optionally, when the sound effect processing device receives a play instruction for the target audio, the sound effect processing device sends an audio information acquisition request for the target audio to the server, so that the server acquires the audio information of the target audio, and receives the acquired audio information fed back by the server.
S502, receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
it can be understood that the sound effect parameter set includes a sample sound effect tag of each sample audio, a sound effect data packet corresponding to each sample sound effect tag, sample frequency information of each sample audio, and sample tone color information, where the sound effect data packet includes a digital equalization parameter of the sample audio and a reverberation parameter of the sample audio; optionally, the sound effect parameter set may further include a sample sound effect tag set, where the sample sound effect tag set is a corresponding relationship between a sample sound effect tag and a sample audio tag, the sample sound effect tag includes different types of music such as rock, metalwork, folk song, and disco, and the sample audio tag corresponding to each type of sample sound effect tag is a different style of music under such a type of sample sound effect tag. For example, as shown in table 1, a type of sound effect parameter set, and table 2, a type of sample sound effect label set, if the sample sound effect label is "rock", the corresponding sample sound effect label may include "art rock, punk, Post rock, check, etc.
In the concrete implementation, the sound effect processing equipment receives a sound effect parameter set sent by the server after the sound effect parameter set is established, stores the sound effect parameter set, and then acquires the digital equalization parameter of the target audio and the reverberation parameter of the target audio from the stored sound effect parameter set.
In a feasible implementation manner, the audio processing device receives audio data and an audio tag of the target audio sent by the server, then obtains a target audio data packet corresponding to the audio tag from the audio parameter set, and reads a digital equalization parameter of the target audio and a reverberation parameter of the target audio from the target audio data packet. Further, the obtaining of the target sound effect data packet corresponding to the target audio in the sound effect parameter set may include searching for a target sample sound effect tag to which the audio tag belongs in a sample sound effect tag set, and obtaining the target sound effect data packet corresponding to the target sample sound effect tag in the sound effect parameter set.
In the concrete implementation, the server sends audio data, audio tags, target frequency information, target tone information and the like to the sound effect processing equipment, the sound effect processing equipment searches for the target sample sound effect tags to which the audio tags belong in the sample sound effect tag set, and the target sound effect data packets corresponding to the target sample sound effect tags are acquired in the sound effect parameter set. That is to say, the sound effect processing device searches for a sample sound effect tag corresponding to the audio tag, and then can acquire the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the sound effect parameter set based on the corresponding relationship between the sample sound effect tag and the sound effect data packet. The sample audio tag corresponding to the audio tag is searched, and it can be understood that the sample audio tag matched with the keyword is found based on the keyword of the sample audio tag, and then the sample audio tag corresponding to the sample audio tag is obtained from the set sample audio tag set. The sample sound effect tag set may be stored as a subset in the sound effect parameter set, or may be stored as a separate set.
In another possible implementation manner, the sound effect processing device receives audio data of the target audio, target frequency information and target tone color information sent by a server; as shown in fig. 10, the obtaining of the target sound effect data packet corresponding to the target audio in the sound effect parameter set includes the following steps:
s601, matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
and S602, obtaining sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
In the specific implementation, when the keyword of the audio tag is unclear or comprises a plurality of audio tags, by decompressing the sound effect data packet, extracting decompressed sample frequency information and sample tone information, sequentially traversing each group of sample frequency information and sample tone information in the sound effect parameter set, respectively matching the target frequency information and the target tone information with the traversed sample frequency information and sample tone information, and after matching, obtaining the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet, obtaining the sample sound effect data packet corresponding to the sample frequency information and the sample tone information with the highest matching similarity, and taking the sample sound effect data packet as a target sound effect data packet, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
S503, synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
In the concrete implementation, the sound effect processing equipment is used for carrying out synthesis processing such as collection, transformation, filtering, estimation, enhancement, compression, identification and the like on the sound effect parameters and the audio data in the determined target sound effect data packet, so that the audio with the sound effect transformed is obtained, and the audio is directly output and played.
In a feasible implementation mode, a DSP digital sound effect processing system is arranged in sound effect processing equipment, and comprises main components of a digital signal processor DSP, audio A/D (analog/digital) and D/A (digital/analog), a RAM, a ROM and a peripheral processor, wherein after a coder-decoder transmits 16-bit sampling data to the DSP each time, an interrupt key receiving interrupt is triggered, the DSP stores the received data into an input buffer memory of the system, simultaneously carries out respective processing (such as conversion, filtering, estimation and the like) on the audio data stored into the buffer memory, and stores the processed data into an output buffer memory of the system after some conversion, and an interrupt key output interrupt program can carry out timing acquisition from the output buffer memory, outputs the processed data in an analog mode through the coder-decoder, and plays the processed data after passing through a power amplifier.
Optionally, dynamic gain processing and noise suppression are used in the synthesis process to ensure that no power overload or clipping distortion is generated.
Optionally, the ultra-high sampling rate of 96kHz is adopted in the synthesis process to ensure high-quality digital-to-analog conversion and high signal-to-noise ratio of over 90 dB.
In the embodiment of the invention, when a server receives a playing instruction of a target audio, audio data of the target audio are acquired and sent to sound effect processing equipment, meanwhile, a sound effect label set carrying sound effect parameters is sent to the sound effect processing equipment, the sound effect processing equipment acquires a target sound effect data packet corresponding to the target audio in the sound effect parameter set, then the target sound effect data packet is adopted to carry out synthesis processing on the audio data, and finally the audio data after the synthesis processing is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
The sound effect processing device according to the embodiment of the present invention will be described in detail with reference to fig. 11 to 18. It should be noted that the sound effect processing apparatus shown in fig. 11 is used for executing the method according to the embodiment of the present invention shown in fig. 1-10, for convenience of description, only the portion related to the embodiment of the present invention is shown, and details of the specific technology are not disclosed, please refer to the embodiment of the present invention shown in fig. 1-10.
Referring to fig. 11, a schematic structural diagram of a sound effect processing device according to an embodiment of the present invention is provided. As shown in fig. 11, the sound effect processing device 1 according to the embodiment of the present invention may include: an information acquisition unit 11, a parameter acquisition unit 12, and a data output unit 13.
An information acquisition unit 11, configured to acquire audio data of a target audio when a play instruction for the target audio is received;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, i.e., voice, music, and other sounds, and in the embodiment of the present invention, the audio is music, which can be a single piece of music in a music player, or a piece of music in multimedia such as video, games, and electronic books. The target audio is music selected by the user from a plurality of pieces of music for output. For an audio, various information such as a music title, a singer's name, an audio stream, an album to which the music belongs, year of publication, total time, an audio tag, and the like may be included. Wherein the audio data is an opaque binary stream represented by a series of non-semantic symbols, i.e. the content of the target audio. The audio tags can be different types of music styles such as art rock, punk, metal music or folk songs. Optionally, the audio may further include frequency information and tone information of the audio, and the frequency information and the tone information are spectral characteristics of the audio, that is, frequency domain characteristics of the audio signal.
In a specific implementation, when the information obtaining unit 11 receives a playing instruction for a target audio, the audio information of the target audio is obtained, and audio data, an audio tag, frequency information, tone color information, and the like in the audio information are extracted. For example, when the sound effect processing apparatus receives a play instruction to play a target audio "actor", the audio data of the "actor" and the audio tag "popular" are acquired.
A parameter obtaining unit 12, configured to obtain a target sound effect data packet corresponding to the target audio in a sound effect parameter set, where the target sound effect data packet includes a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
it can be understood that the sound effect parameter set may include a sample sound effect tag of each sample sound frequency, a sound effect data packet corresponding to each sample sound effect tag, sample frequency information of each sample sound frequency, and sample tone color information, where the sound effect data packet may include a digital equalization parameter and a reverberation parameter, and these information may be stored in the form of a data packet or a mapping relationship table; optionally, the sound effect parameter set may further include a sample sound effect tag set, where the sample sound effect tag set is a corresponding relationship between a sample sound effect tag and a sample audio tag, the sample sound effect tag includes different types of music such as rock, metalwork, folk song, and disco, and the sample audio tag corresponding to each type of sample sound effect tag is a different style of music under such a type of sample sound effect tag. For example, as shown in table 1, a type of sound effect parameter set, and table 2, a type of sample sound effect label set, if the sample sound effect label is "rock", the corresponding sample sound effect label may include "art rock, punk, Post rock, check, etc.
Specifically, the digital equalization parameter is a gain value of a signal after adjusting a signal of each frequency band in a digital equalization processing curve, and the digital equalization parameter corresponds to each octave feature point one by one. The octave feature points are frequency bands obtained by dividing the discrete spectrum into individual frequency bands, and the upper limit frequency of each frequency band is doubled (i.e. the frequency band with the frequency interval of 2: 1) than the lower limit frequency, so as to obtain a plurality of octaves, and then the feature points are respectively selected in each frequency band. For example, if the frequency range of the signal is 63Hz to 16kHz, the signal can be divided into 8 octaves, which are 63Hz to 126Hz, 126Hz to 252Hz, 252Hz to 504Hz, 504Hz to 1.08kHz, 1.008kHz to 2.016kHz, 2.016kHz to 4.032kHz, 4.032kHz to 8.064kHz and 8.064Hz to 16kHz, and then a feature point is taken in each octave, and each feature point corresponds to a digital equalization parameter. Similarly, if two frequencies are inserted between the upper and lower limit frequencies of an octave, and the ratio between the 4 frequencies is the same (the ratio between the two adjacent frequencies is 1.33 times), the octave is divided into 3 frequency ranges, which is called 1/3 octaves. According to the calculation method, n times of octaves can be obtained, namely, the octaves are obtained by taking oc as a coefficient ((2) ^1/2) ^ n at the reference frequency f0Before and after each of L ═ f0Oc, then calculating power spectrum by segments and merging, wherein f0Is the reference frequency. The octave can be applied to the fields of 31-segment equalizers, sound pressure analysis, vibration reduction, noise reduction and the like.
Reverberation is generated because a sound-producing object emits sound waves, and the sound waves are reflected by contacting the surface of an obstacle through air, and due to the complexity of the real environment, the sound emitted by a sound source generates various echoes from various directions, and the sound is mixed to form the so-called reverberation. The reverberation parameters may include reverberation intensity, reverberation time, scattering degree, reverberation density, and the like. The sound with short reverberation time is dry and dry, the sound with too long reverberation time is not confused and loses a large amount of details, and the proper reverberation time can beautify the sound, cover the noise of musical instruments and enable the tone fusion to increase the loudness and the consistency of syllables. It should be noted that one target audio corresponds to one reverberation parameter.
In specific implementation, the parameter obtaining unit 12 obtains an audio tag of a target audio, searches a target sample audio tag to which the audio tag belongs in a sample audio tag set, and obtains an audio data packet corresponding to the target sample audio tag in the audio parameter set; optionally, a target sample sound effect tag to which the audio tag belongs is searched in a sample sound effect tag set, a target sound effect data packet corresponding to the target sample sound effect tag is obtained in the sound effect parameter set, and a digital equalization parameter of a target audio frequency and a reverberation parameter of the target audio frequency in the target sound effect data packet are read; optionally, the target frequency information and the target tone information of the target audio are matched with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, the matching similarity between the target frequency information and the sample tone information of each sound effect data packet and the sample frequency information and the sample tone information of each sound effect data packet is obtained after matching, the sample frequency information with the highest matching similarity and the sample sound effect data packet corresponding to the sample tone information are obtained, and the sample sound effect data packet is used as the target sound effect data packet.
For example, if the audio tag of the target audio is "heavy metal", the target sample audio tag "metallacy" to which the "heavy metal" belongs is determined in the sample audio tag set shown in table 2, and then the audio parameter of "metallacy" is determined to be "B1" in the audio parameter set shown in table 1.
And the data output unit 13 is configured to perform synthesis processing on the audio data by using the target sound effect data packet, and output the audio data after the synthesis processing.
In the specific implementation, the data output unit 13 performs synthesis processing such as collection, conversion, filtering, estimation, enhancement, compression, and identification on the sound effect parameters and the audio data in the determined target sound effect data packet, so as to obtain the audio with the sound effect converted, and then outputs the audio.
When the sound effect processing equipment is a server, the data output unit 13 sends the audio with the changed sound effect to a sound terminal so that the sound terminal can output the audio with the changed sound effect; when the sound effect processing equipment is a sound terminal, the data output unit 13 can directly output and play the audio with the changed sound effect.
In a feasible implementation, a DSP digital sound processing system is provided in the data output unit 13, which includes a digital signal processor DSP, audio a/D (analog/digital) and D/a (digital/analog), RAM, ROM and a peripheral processor, after the codec transmits 16-bit sampling data to the DSP each time, will cause the interrupt key to receive the interrupt, the DSP will store the received data into the input buffer of the system, meanwhile, the audio data stored in the buffer memory are respectively processed (such as transformation, filtering, estimation and the like), and are stored in the output buffer memory of the system after some transformation, the interrupt key output interrupt program can execute the number fetching from the output buffer at regular time, output in an analog mode through the coder and the decoder and then send to the sound terminal or play directly after passing through the power amplifier.
Optionally, the data output unit 13 employs dynamic gain processing and noise suppression during the synthesis process to ensure that no power overload or clipping distortion is generated.
Optionally, the data output unit 13 may use a sampling rate of 96kHz in the synthesis process, which is very high and can ensure high-quality digital-to-analog conversion and a high signal-to-noise ratio of over 90 dB.
In the embodiment of the invention, when a playing instruction of a target audio is received, the audio data of the target audio is obtained, and after a target audio effect data packet corresponding to a target audio label is obtained from an audio effect parameter set, the audio data is synthesized by using the target audio effect data packet, and finally the audio data after synthesis is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
Referring to fig. 12, a schematic structural diagram of another sound effect processing device according to an embodiment of the present invention is provided. As shown in fig. 12, the sound effect processing device 1 according to the embodiment of the present invention may include: an information acquisition unit 11, a parameter acquisition unit 12, a data output unit 13, a sample information acquisition unit 14, a sample parameter acquisition unit 15, a sample parameter adjustment unit 16, and a sample information holding unit 17.
The sample information acquiring unit 14 is configured to acquire a sample audio, acquire a characteristic frequency response curve of the sample audio, and acquire sample frequency information and sample tone information of the sample audio;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, namely voice, music and other sounds, and in the embodiment of the present invention, the sample audio is music, and the sample audio is at least one piece of music selected by a user for storage among a plurality of pieces of music. For a sample audio, various information such as sample audio data, frequency information, timbre information, singer's name, album to which it belongs, year of publication, total time length, sample audio tag, etc. may be included. The sample audio tags may be different types of music styles such as art rock, punk, heavy metal, black metal, or folk song, the sample audio data is an opaque binary stream represented by a series of non-semantic symbols, that is, the content of the sample audio, and the sample frequency information and the sample tone color information are spectral characteristics of the sample audio (frequency domain characteristics of the sample audio signal).
The characteristic frequency response is a phenomenon that when a sample audio signal output at a constant voltage is connected to a system, the generated sound pressure increases or attenuates with the change of frequency, and a curve of the associated change relationship between the sound pressure and the frequency is called a characteristic frequency response curve, as shown in fig. 3, the abscissa is the frequency, and the ordinate is the sound pressure level. In specific implementation, since the frequency range of the human ear listening is 20Hz to 20kHz, the sample information obtaining unit 14 may collect a sample audio frequency of 20Hz to 20kHz, extract a spectral characteristic in a suitable frequency range (e.g., 50Hz to 20kHz) in the sample audio frequency, and draw a characteristic frequency response curve of the sample audio frequency according to the extracted spectral characteristic.
A sample parameter obtaining unit 15, configured to obtain a digital equalization processing curve of the sample audio and a reverberation parameter of the sample audio based on the characteristic frequency response curve, the sample frequency information, and the sample timbre information;
it is understood that the principle of the digital equalization process is: the input signal "X" is used to establish the corresponding output signal "Y", where f () includes a function of the frequency "k" corresponding to "X". The functional expression corresponding to "X" is expanded, i.e., Y ═ g (k) ×, where g () varies as the digital equalization parameters are adjusted. In the embodiment of the present invention, the input signal "X" is sample frequency information, the sample tone information, and a characteristic frequency response curve, and the frequency "k" is a frequency value corresponding to the characteristic frequency response curve. Based on the processing principle described above, a digital equalization processing curve may be generated. As shown in fig. 4, the abscissa is frequency, the ordinate is digital equalization parameter, and the digital equalization processing curve is a graph of variation of the digital equalization parameter with frequency.
Meanwhile, based on the characteristic frequency response curve, the sample frequency information and the sample tone color information, reverberation parameters of the sample audio frequency can be obtained. The reverberation is generated because a sound-producing object emits sound waves, and the sound waves are reflected by contacting the surface of an obstacle through air, and due to the complexity of the real environment, the sound emitted by a sound source generates various echoes from various directions, and the sound is mixed to form the so-called reverberation. The first sound of the reverberation is the "direct sound", i.e. the source sound, called dry sound output in the effector, which is the sound emitted from the sound directly reaching the listener's ears, and is also a major component of the sound pressure level. The propagation attenuation of the sound pressure level is inversely proportional to the square of the distance, i.e. the distance is doubled and the sound pressure level is reduced by 6 dB. The subsequent obvious sounds which are relatively separated are called as ' early reflected sounds ', also called as near secondary reflected sounds ', the sounds which are emitted by the sound source and reach the ears of a listener after being reflected for 1-2 times by surrounding interfaces (walls, ceilings and floors), the reflected sounds which reach 50ms later than the direct sounds all belong to the range, the sounds are relatively large and relatively obvious, and the distance relation among the source sounds, the ears and the walls in the space can be reflected. Early reflected sound is characterized by the inability of the ear to distinguish it from direct sound, and can only be perceived as superimposed. Early reflection of sound is therefore beneficial to improve sound pressure level and clarity of sound. Its propagation attenuation is related to the sound absorption properties of the reflective interface. The last sound is a "reverberant sound" which is a multiple reflection sound that is more than 50ms later than the direct sound. For music, although the richness of the music can be increased by the reverberation sound, the clearness of the sound can be reduced while the richness of the music is increased, and if the richness of the music is too small, the sound can be dried and cannot be too large. The magnitude of the reverberant sound is directly related to the sound absorbing properties of the surrounding interface.
The reverberation parameters may include reverberation intensity, reverberation time, scattering degree, reverberation density, and the like. The sound with short reverberation time is dry and dry, the sound with too long reverberation time is not confused and loses a large amount of details, and the proper reverberation time can beautify the sound, cover the noise of musical instruments and enable the tone fusion to increase the loudness and the consistency of syllables. It should be noted that one sample audio corresponds to one reverberation parameter.
A sample parameter adjusting unit 16, configured to obtain a sample sound effect label of the sample audio, and adjust an octave feature point in a selected frequency range in the digital equalization processing curve based on an equal response curve and the sample sound effect label, so as to obtain a digital equalization parameter corresponding to each octave feature point;
specifically, the sample sound effect labels may include different types of music such as rock, metal music, folk song, and disco, and each type of sample sound effect label may include at least one sample audio label, for example, the sample sound effect label "rock" may include "art rock, punk, Post rock, and tomb" sample audio labels. The sample audio effect tag for obtaining the sample audio can be obtained by obtaining the sample audio effect tag for the sample audio based on the corresponding relationship between the sample audio effect tag and the sample audio tag as shown in fig. 2. For example, if the sample audio is "physalis alkekengi" and the sample audio label of the "physalis alkekengi" is "popular metal", then table 2 indicates that "popular metal" belongs to "metal music", and thus the sample sound effect label of the "physalis alkekengi" is "metal music".
The equal loudness curve is a curve describing the relation between sound pressure level and sound wave frequency under equal loudness, and is one of auditory characteristics. I.e. what sound pressure level the pure tone needs to reach at different frequencies in order to obtain a consistent auditory loudness for the listener. Fig. 5 shows a pure tone equal loudness curve with frequency on the abscissa and sound pressure level on the ordinate. The sound pressure is a change of atmospheric pressure caused by disturbance, that is, a residual pressure of atmospheric pressure, and is equivalent to a pressure change caused by superimposing a disturbance on the atmospheric pressure. The sound pressure level is denoted by the symbol SPL in dB, which is defined by taking the common logarithm of the ratio of the effective value p (e) of the sound pressure to be measured to the reference sound pressure p (ref), and multiplying by 20. In the figure, the sound pressure levels of different frequencies corresponding to each curve are different, but the loudness sensed by human ears is the same, and each curve represents the change relationship between the frequency and the sound pressure level under different loudness. From the family of equal loudness curves, it can be seen that the human ear is insensitive to high and low tones when the loudness is small, and becomes increasingly sensitive to high and low tones when the loudness is large.
The digital equalization parameters are gain values of signals after signals of each frequency band in a digital equalization processing curve are adjusted, and the digital equalization parameters correspond to the characteristic points of each octave one by one. The octave feature points are frequency bands obtained by dividing the discrete spectrum into individual frequency bands, and the upper limit frequency of each frequency band is doubled (i.e. the frequency band with the frequency interval of 2: 1) than the lower limit frequency, so as to obtain a plurality of octaves, and then the feature points are respectively taken in each octave. For example, if the signal frequency range is 63Hz to 16kHz, the signal frequency range can be divided into 8 octaves, which are 63Hz to 126Hz, 126Hz to 252Hz, 252Hz to 504Hz, 504Hz to 1.08kHz, 1.008kHz to 2.016kHz, 2.016kHz to 4.032kHz, 4.032kHz to 8.064kHz and 8.064Hz to 16kHz, and then a feature point is taken in each octave, if the signal frequency range is 1/24 octave feature points, each octave is divided into 24 frequency bands, each frequency band is taken as a 1/24 octave feature point of the octave, each 1/24 octave feature point corresponds to a digital equalization parameter, and the digital equalization parameter corresponding to each 1/24 octave feature point is adjusted, so that a suitable parameter range is determined.
In a specific implementation, the sample parameter adjusting unit 16 selects a curve of a part of frequency ranges in the digital equalization processing curve, divides the selected curve into multiple octaves, then adjusts the feature points in each octave based on the parameter information of the equal-response curve, and ensures that the digital equalization parameters of each octave feature point change in the parameter range indicated by the sample sound effect label during the adjustment process, thereby determining the digital equalization parameter range of each octave feature point. For example, a curve in the range of 63Hz to 16kHz in the digital equalization processing curve shown in fig. 4 is divided into 8 octaves, then a feature point is taken for each octave to obtain a discrete octave feature point shown in fig. 6, and then the digital equalization parameter of each discrete octave feature point is adjusted based on the common digital equalization parameter corresponding to the sample sound effect label to determine the digital equalization parameter range of each octave feature point under the sample sound effect label.
A sample information storage unit 17, configured to store the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, and the sample audio label in an audio parameter set;
it is understood that the digital equalization parameters, the reverberation parameters and the sample sound effect tags may be stored in the form of data packets in the sound effect parameter set, or in the form of a mapping table in the sound effect parameter set.
Optionally, as shown in fig. 13, the sample information saving unit 17 includes:
a data packet obtaining subunit 171, configured to compress the digital equalization parameter corresponding to each octave feature point and the reverberation parameter of the sample audio, and store the compressed digital equalization parameter and the compressed reverberation parameter as a sound effect data packet corresponding to the sample sound effect tag;
specifically, the data packet is a data storage format, and may include a plurality of fields, and each field may identify different information. And forming an audio effect data packet by compressing the digital equalization parameter and the reverberation parameter into a data field of the audio effect data packet. Each sound effect data packet comprises a plurality of groups of digital equalization parameters and a group of reverberation parameters. The multiple sets of digital equalization parameters include digital equalization parameters corresponding to the feature points of each octave, and each octave may include multiple feature points, where it should be noted that the multiple sets of digital equalization parameters and the reverberation parameters of the sample audio are parameters indicated by the sample audio label, that is, the digital equalization parameters of the sample audio and the reverberation parameters of the sample audio are different under different sample labels.
For example, if the signal frequency range is 63Hz to 16kHz and is divided into 8 octaves, 9 feature points are taken in each octave, the digital equalization parameters corresponding to 9 feature points of the 1 st octave are a11 to a19, the digital equalization parameters corresponding to 9 feature points of the 2 nd octave are a21 to a29 …, the digital equalization parameters corresponding to 9 feature points of the 8 th octave are a81 to a89, and the reverberation parameter of the signal is B, then the digital equalization parameters of the multiple sets of sample audio stored in one sound effect data packet are a11 to a19, a21 to a29 … a81 to a89, and the reverberation parameter of the sample audio is B, as shown in table 3.
And the information storage subunit 172 is configured to store the sample sound effect tag and the sound effect data packet into the sound effect parameter set.
In the concrete implementation, the sample sound effect tag is added to a selected field, such as a header field, of the sound effect data packet, and then the sound effect data packet added with the sample sound effect tag is stored in a sound effect parameter set, or the sample sound effect tag and the sound effect data packet are correspondingly stored in the sound effect parameter set in a form of a mapping relation table.
An information acquisition unit 11, configured to acquire audio data of a target audio when a play instruction for the target audio is received;
it is to be understood that, in the embodiment of the present invention, the target audio is music, and the target audio is music selected by the user for output among pieces of music.
In a specific implementation manner of the embodiment of the present invention, when the information obtaining unit 11 receives a play instruction for a target audio, the audio information of the target audio is obtained, and audio data and an audio tag in the audio information are extracted. For example, when the sound effect processing apparatus receives a play instruction to play a target audio "actor", the audio data of the "actor" and the audio tag "popular" are acquired.
In another specific implementation manner of the embodiment of the present invention, when the information obtaining unit 11 receives a play instruction for a target audio, audio data, target frequency information, and target tone color information of the target audio are obtained. The target frequency information and the target timbre information are spectral characteristics of the target audio, that is, frequency domain characteristics of the target audio signal.
A parameter obtaining unit 12, configured to obtain a target sound effect data packet corresponding to the target audio in a sound effect parameter set, where the target sound effect data packet includes a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
optionally, the information obtaining unit 11 is specifically configured to:
acquiring audio data and an audio tag of the target audio;
the parameter obtaining unit 12 is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading a digital equalization parameter of a target audio and a reverberation parameter of the target audio in the target sound effect data packet.
Further, the parameter obtaining unit 12 is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
In specific implementation, the parameter obtaining unit 12 searches for a sample sound effect tag corresponding to the audio tag, and then obtains the digital equalization parameter of the target audio and the reverberation parameter of the target audio based on the corresponding relationship between the sample sound effect tag and the sound effect data packet in the sound effect parameter set. The sample audio tag corresponding to the audio tag is searched, and it can be understood that the sample audio tag matched with the keyword is found based on the keyword of the sample audio tag, and then the sample audio tag corresponding to the sample audio tag is obtained from the set sample audio tag set. The sample sound effect tag set may be stored as a subset in the sound effect parameter set, or may be stored as a separate set.
Optionally, the sample information storing unit 17 is specifically configured to store the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, the sample frequency information, the sample tone information, and the sample sound effect label into a sound effect parameter set;
further, the information obtaining unit 11 is specifically configured to:
acquiring audio data, target frequency information and target tone information of the target audio;
further, as shown in fig. 14, the parameter obtaining unit 12 includes:
a similarity obtaining subunit 121, configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
in specific implementation, when the keywords of the audio tag are unclear or include a plurality of audio tags, the audio data packet may be decompressed by the decompression unit, the decompressed sample frequency information and sample tone information are extracted, each group of sample frequency information and sample tone information in the audio parameter set is sequentially traversed, the target frequency information and the target tone information are respectively matched with the traversed sample frequency information and sample tone information, and matching similarity of each group of matched samples is obtained. Optionally, the matched matching similarity and the sound effect data packet may be cached correspondingly, the caching mode may be to add the matching similarity to a set field of the sound effect data packet, or to cache the matching similarity and the sound effect data packet in a form of a mapping table to a sound effect parameter set.
And the data packet obtaining subunit 122 is configured to obtain a sample sound effect data packet corresponding to the sample tone information and the sample frequency information with the highest matching similarity, and use the sample sound effect data packet as a target sound effect data packet.
And the data output unit 13 is configured to perform synthesis processing on the audio data by using the target sound effect data packet, and send the audio data after the synthesis processing to the sound terminal, so that the sound terminal outputs the audio data after the synthesis processing.
In specific implementation, the data output unit 13 performs synthesis processing such as collection, conversion, filtering, estimation, enhancement, compression, and identification on the determined sound effect parameters and the determined sound effect data in the target sound effect data packet, so as to obtain the sound frequency after the sound effect conversion, and send the sound frequency to the sound terminal for output. In a feasible implementation, a DSP digital sound processing system is provided in the data output unit 13, which includes a digital signal processor DSP, audio a/D (analog/digital) and D/a (digital/analog), RAM, ROM and a peripheral processor, after the codec transmits 16-bit sampling data to the DSP each time, will cause the interrupt key to receive the interrupt, the DSP will store the received data into the input buffer of the system, meanwhile, the audio data stored in the buffer memory are respectively processed (such as transformation, filtering, estimation and the like), and are stored in the output buffer memory of the system after some transformation, and the interrupt key output interrupt program can execute the acquisition of data from the output buffer at regular time, output the data in an analog mode through the codec, send the data to the sound terminal, and play the data after passing through the power amplifier.
In the embodiment of the invention, firstly, an audio tag set carrying audio parameters is arranged on an audio processing device, when a playing instruction of a target audio is received, audio data of the target audio is obtained, a target audio data packet corresponding to the target audio is obtained in the audio parameter set, then the target audio data packet is adopted to carry out synthesis processing on the audio data, and finally, the audio data after the synthesis processing is sent to a sound terminal for outputting. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, the intelligence of sound effect processing is improved, and the brand-new personalized audio experience is provided.
Referring to fig. 15, a schematic structural diagram of another sound effect processing device according to an embodiment of the present invention is provided. As shown in fig. 15, the sound effect processing device 2 according to the embodiment of the present invention may include: an information receiving unit 21, a set transmitting unit 22, and a data output unit 23.
An information receiving unit 21, configured to receive audio data of a target audio sent by a server when a play instruction for the target audio is received;
it is understood that audio is an important medium in multimedia, in the form of sound signals. As a carrier of information, audio can be divided into three types, i.e., voice, music, and other sounds, and in the embodiment of the present invention, the audio is music, which can be a single piece of music in a music player, or a piece of music in multimedia such as video, games, and electronic books. The target audio is music selected by the user from a plurality of pieces of music for output. For one audio, various information such as a music title, a singer's name, audio data, an album to which the user belongs, year of publication, total time, audio tags, and the like may be included. Wherein the audio data is an opaque binary stream represented by a series of non-semantic symbols, i.e. the content of the target audio. The audio tags can be different types of music styles such as art rock, punk, metal music or folk songs. Optionally, the audio may further include frequency information and tone information of the audio, where the frequency information and the tone information are spectral characteristics of the audio, that is, frequency domain characteristics of the audio signal.
In a specific implementation, when the server receives a playing instruction for a target audio, the server collects audio information of the target audio, extracts audio data in the audio information, and sends the audio data to the information receiving unit 21; optionally, when the information receiving unit 21 receives a playing instruction for the target audio, it sends an audio information obtaining request of the target audio to the server, so that the server collects the audio information of the target audio, and receives the collected audio information fed back by the server.
Optionally, the information receiving unit 21 is specifically configured to receive the audio data and the audio tag of the target audio sent by the server.
Optionally, the information receiving unit 21 is specifically configured to receive the audio data of the target audio, the target frequency information, and the target tone color information sent by the server.
A parameter obtaining unit 22, configured to receive a sound effect parameter set sent by the server, and obtain a target sound effect data packet corresponding to the target audio in the sound effect parameter set, where the target sound effect data packet includes a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
it can be understood that the sound effect parameter set includes a sample sound effect tag of each sample audio, a sound effect data packet corresponding to each sample sound effect tag, sample frequency information of each sample audio, and sample tone color information, where the sound effect data packet includes a digital equalization parameter of the sample audio and a reverberation parameter of the sample audio; optionally, the sound effect parameter set may further include a sample sound effect tag set, where the sample sound effect tag set is a corresponding relationship between a sample sound effect tag and a sample audio tag, the sample sound effect tag includes different types of music such as rock, metalwork, folk song, and disco, and the sample audio tag corresponding to each type of sample sound effect tag is a different style of music under such a type of sample sound effect tag. For example, as shown in table 1, a type of sound effect parameter set, and table 2, a type of sample sound effect label set, if the sample sound effect label is "rock", the corresponding sample sound effect label may include "art rock, punk, Post rock, check, etc.
In a specific implementation, the parameter obtaining unit 22 receives a sound effect parameter set sent by the server after the sound effect parameter set is established, stores the sound effect parameter set, and then obtains the digital equalization parameter of the target audio and the reverberation parameter of the target audio from the stored sound effect parameter set.
Optionally, the parameter obtaining unit 22 is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
Further, the parameter obtaining unit 22 is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
In the concrete implementation, the server sends the audio data, the audio tag, the target frequency information, the target tone information and the like to the parameter obtaining unit 22, the parameter obtaining unit 22 searches the target sample audio tag to which the audio tag belongs in the sample audio tag set, and obtains the target audio data packet corresponding to the target sample audio tag in the audio parameter set. That is to say, the parameter obtaining unit 22 searches for a sample sound effect tag corresponding to the audio tag, and then obtains the digital equalization parameter of the target audio and the reverberation parameter of the target audio based on the corresponding relationship between the sample sound effect tag and the sound effect data packet in the sound effect parameter set. The sample audio tag corresponding to the audio tag is searched, and it can be understood that the sample audio tag matched with the keyword is found based on the keyword of the sample audio tag, and then the sample audio tag corresponding to the sample audio tag is obtained from the set sample audio tag set. The sample sound effect tag set may be stored as a subset in the sound effect parameter set, or may be stored as a separate set.
Optionally, as shown in fig. 16, the parameter obtaining unit 22 includes:
a similarity obtaining subunit 221, configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
the data packet obtaining subunit 222 is configured to obtain a sample sound effect data packet corresponding to the sample tone information and the sample frequency information with the highest matching similarity, and use the sample sound effect data packet as a target sound effect data packet.
In the specific implementation, when the keyword of the audio tag is unclear or comprises a plurality of audio tags, by decompressing the sound effect data packet, extracting decompressed sample frequency information and sample tone information, sequentially traversing each group of sample frequency information and sample tone information in the sound effect parameter set, respectively matching the target frequency information and the target tone information with the traversed sample frequency information and sample tone information, and after matching, obtaining the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet, obtaining the sample sound effect data packet corresponding to the sample frequency information and the sample tone information with the highest matching similarity, and taking the sample sound effect data packet as a target sound effect data packet, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
And a data output unit 23, configured to perform synthesis processing on the audio data by using the target sound effect data packet, and output the audio data after the synthesis processing.
In the concrete implementation, the sound effect processing equipment is used for carrying out synthesis processing such as collection, transformation, filtering, estimation, enhancement, compression, identification and the like on the sound effect parameters and the audio data in the determined target sound effect data packet, so that the audio with the sound effect transformed is obtained, and the audio is directly output and played.
In a feasible implementation, a DSP digital sound processing system is provided in the data output unit 23, and includes main devices including a digital signal processor DSP, audio a/D (analog/digital) and D/a (digital/analog), RAM, ROM, and a peripheral processor, after a codec transmits 16-bit sampling data to the DSP each time, an interrupt key reception interrupt is triggered, the DSP stores the received data in an input buffer of the system, and simultaneously performs respective processing (such as conversion, filtering, estimation, etc.) on the audio data already stored in the buffer, and stores the processed data in an output buffer of the system after some conversion, and an interrupt program of the interrupt key output periodically performs data fetching from the output buffer, outputs the processed data in an analog manner through the codec, and plays the processed data after passing through a power amplifier.
Optionally, the data output unit 23 employs dynamic gain processing and noise suppression during the synthesis processing to ensure that no power overload or clipping distortion is generated.
Optionally, the data output unit 23 may use a sampling rate of 96kHz in the synthesis process, which is very high and can ensure high-quality digital-to-analog conversion and a high signal-to-noise ratio of over 90 dB.
In the embodiment of the invention, when a server receives a playing instruction of a target audio, audio data of the target audio are acquired and sent to sound effect processing equipment, meanwhile, a sound effect label set carrying sound effect parameters is sent to the sound effect processing equipment, the sound effect processing equipment acquires a target sound effect data packet corresponding to the target audio in the sound effect parameter set, then the target sound effect data packet is adopted to carry out synthesis processing on the audio data, and finally the audio data after the synthesis processing is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
An embodiment of the present invention further provides a computer storage medium, where the computer storage medium may store a plurality of instructions, where the instructions are suitable for being loaded by a processor and executing the method steps in the embodiments shown in fig. 1 to 10, and a specific execution process may refer to specific descriptions of the embodiments shown in fig. 1 to 10, which are not described herein again.
Fig. 17 is a schematic structural diagram of a server according to an embodiment of the present invention. As shown in fig. 17, the server 1000 may include: at least one processor 1001, such as a CPU, at least one network interface 1004, a user interface 1003, memory 1005, at least one communication bus 1002. Wherein a communication bus 1002 is used to enable connective communication between these components. The user interface 1003 may include a Display screen (Display) and a Keyboard (Keyboard), and the optional user interface 1003 may also include a standard wired interface and a standard wireless interface. The network interface 1004 may optionally include a standard wired interface, a wireless interface (e.g., WI-FI interface). The memory 1005 may be a high-speed RAM memory or a non-volatile memory (non-volatile memory), such as at least one disk memory. The memory 1005 may optionally be at least one memory device located remotely from the processor 1001. As shown in fig. 17, a memory 1005, which is a kind of computer storage medium, may include therein an operating system, a network communication module, a user interface module, and a sound effect processing application program.
In the server 1000 shown in fig. 17, the user interface 1003 is mainly used as an interface for providing input for a user, and acquiring data input by the user; the network interface 1004 is mainly used for data communication with the user terminal; the processor 1001 may be configured to call the sound effect processing application stored in the memory 1005, and specifically perform the following operations:
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
In one embodiment, before executing the step of receiving the playing instruction of the target audio, the processor 1001 further performs the following operations:
acquiring sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring frequency information and tone information of the sample audio;
acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
acquiring a sample sound effect label of the sample audio, and adjusting octave characteristic points in a selected frequency range in the digital equalization processing curve based on an equal response curve and the sample sound effect label to obtain digital equalization parameters corresponding to each octave characteristic point;
and storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set.
In one embodiment, when the processor 1001 stores the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, and the sample audio effect tag into the audio effect parameter set, the following operations are specifically performed:
compressing the digital equalization parameters corresponding to each octave characteristic point and the reverberation parameters of the sample audios and storing the compressed digital equalization parameters and the reverberation parameters of the sample audios as sound effect data packets corresponding to the sample sound effect labels;
and storing the sample sound effect label and the sound effect data packet corresponding to the sample sound effect label into the sound effect parameter set.
In an embodiment, when the processor 1001 executes the audio data of the target audio, the following steps are specifically executed:
acquiring audio data and an audio tag of the target audio;
further, when the processor 1001 acquires a target audio effect data packet corresponding to the target audio from the audio effect parameter set, the following operations are specifically performed:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
In an embodiment, when the processor 1001 acquires the target sound effect data packet corresponding to the audio tag from the sound effect parameter set, the following operations are specifically performed:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
In one embodiment, when the processor 1001 stores the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, and the sample audio effect tag into the audio effect parameter set, the following operations are specifically performed:
and storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios, the sample frequency information, the sample tone color information and the sample sound effect labels into a sound effect parameter set.
In one embodiment, when the processor 1001 executes the following operation of acquiring the audio data of the target audio:
acquiring audio data, target frequency information and target tone information of the target audio;
further, when the processor 1001 acquires a target sound effect data packet corresponding to the target audio from the sound effect parameter set, specifically perform the following operations:
matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
and acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
In the embodiment of the invention, when a playing instruction of a target audio is received, the audio data of the target audio is obtained, a target audio data packet corresponding to the target audio is obtained from an audio parameter set, then the audio data is synthesized by using the target audio data packet, and finally the audio data after synthesis is sent to a sound terminal for output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
Fig. 18 is a schematic structural diagram of an audio terminal according to an embodiment of the present invention. As shown in fig. 18, the acoustic terminal 2000 may include: at least one processor 2001, e.g., a CPU, at least one network interface 2004, a user interface 2003, a memory 2005, at least one communication bus 2002. The communication bus 2002 is used to implement connection communication between these components. The user interface 2003 may include a Display (Display) and a Keyboard (Keyboard), and the optional user interface 2003 may further include a standard wired interface and a standard wireless interface. The network interface 1004 may optionally include a standard wired interface, a wireless interface (e.g., WI-FI interface). Memory 2005 may be a high-speed RAM memory or a non-volatile memory (e.g., at least one disk memory). The memory 2005 may optionally also be at least one memory device located remotely from the aforementioned processor 2001. As shown in fig. 18, the memory 2005, which is a kind of computer storage medium, may include therein an operating system, a network communication module, a user interface module, and a sound effect processing application program.
In the audio terminal 2000 shown in fig. 18, the user interface 2003 is mainly used as an interface for providing input for a user to acquire data input by the user; the network interface 2004 is mainly used for data communication with the user terminal; the processor 2001 may be configured to call the sound effect processing application stored in the memory 2005, and specifically perform the following operations:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
In one embodiment, when the processor 2001 executes the audio data of the target audio sent by the receiving server, the following steps are specifically executed:
receiving audio data and an audio tag of the target audio sent by a server;
further, when the processor 2001 executes to acquire the target sound effect data packet corresponding to the target audio from the sound effect parameter set, the following operations are specifically executed:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
In an embodiment, when the processor 2001 acquires the target sound effect data packet corresponding to the audio tag from the sound effect parameter set, the following operations are specifically performed:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
In one embodiment, the processor 2001 specifically performs the following operations when executing the audio data of the target audio sent by the receiving server:
receiving audio data, target frequency information and target tone color information of the target audio sent by a server;
further, when the processor 2001 executes to acquire the target sound effect data packet corresponding to the target audio from the sound effect parameter set, the following operations are specifically executed:
matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
and acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
In the embodiment of the invention, when a server receives a playing instruction of a target audio, audio data of the target audio are acquired and sent to sound effect processing equipment, meanwhile, a sound effect label set carrying sound effect parameters is sent to the sound effect processing equipment, the sound effect processing equipment acquires a target sound effect data packet corresponding to the target audio in the sound effect parameter set, then the target sound effect data packet is adopted to carry out synthesis processing on the audio data, and finally the audio data after the synthesis processing is output. The self-adaptive sound effect is provided through the information based on the audio frequency, the sound effect most suitable for the audio content can be built, the sound effect processing mode is enriched, and the intelligence of sound effect processing is improved.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), or the like.
The above disclosure is only for the purpose of illustrating the preferred embodiments of the present invention, and it is therefore to be understood that the invention is not limited by the scope of the appended claims.

Claims (22)

1. A sound effect processing method is characterized by comprising the following steps:
acquiring sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring sample frequency information and sample tone information of the sample audio;
acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
acquiring a sample sound effect label of the sample audio, determining a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusting the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensuring that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label so as to obtain the digital equalization parameters corresponding to each octave characteristic point;
storing the digital equalization parameters corresponding to each octave characteristic point, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
synthesizing the audio data by adopting the target sound effect data packet; and
and outputting the audio data after the synthesis processing.
2. The method of claim 1, wherein the storing the digital equalization parameters corresponding to each octave feature point, the reverberation parameters of the sample audio, and the sample audio effect label into an audio effect parameter set comprises:
compressing the digital equalization parameters corresponding to each octave characteristic point and the reverberation parameters of the sample audios and storing the compressed digital equalization parameters and the reverberation parameters of the sample audios as sound effect data packets corresponding to the sample sound effect labels;
and storing the sample sound effect label and the sound effect data packet corresponding to the sample sound effect label into the sound effect parameter set.
3. The method of claim 1, wherein the obtaining audio data of the target audio comprises:
acquiring audio data and an audio tag of the target audio;
the obtaining of the target audio data packet corresponding to the target audio in the audio parameter set includes:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
4. The method according to claim 3, wherein the obtaining of the target sound effect data packet corresponding to the audio tag from the sound effect parameter set comprises:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
5. The method of claim 1, wherein the storing the digital equalization parameters corresponding to each octave feature point, the reverberation parameters of the sample audio, and the sample audio effect label into an audio effect parameter set comprises:
and storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios, the sample frequency information, the sample tone color information and the sample sound effect labels into a sound effect parameter set.
6. The method of claim 5, wherein the obtaining audio data of the target audio comprises:
acquiring audio data, target frequency information and target tone information of the target audio;
the obtaining of the target audio effect data packet corresponding to the target audio in the audio effect parameter set includes:
matching the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and acquiring the matching similarity of the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet after matching;
and acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
7. The method according to claim 1, wherein when the sound effect processing method is running on the server side;
the outputting the audio data after the synthesis processing includes:
and sending the audio data after the synthesis processing to an audio terminal so that the audio terminal outputs the audio data after the synthesis processing.
8. The method according to claim 1, characterized in that when the sound effect processing method is operated on the side of a sound terminal;
when a playing instruction of a target audio is received, acquiring audio data of the target audio, including:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
the acquiring a target audio effect data packet corresponding to the target audio in the audio effect parameter set, where the target audio effect data packet includes a digital equalization parameter of the target audio and a reverberation parameter of the target audio, and includes:
receiving a sound effect parameter set sent by the server;
and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio.
9. An audio effect processing apparatus, comprising:
the system comprises a sample information acquisition unit, a sample processing unit and a sample processing unit, wherein the sample information acquisition unit is used for acquiring a sample audio, acquiring a characteristic frequency response curve of the sample audio and acquiring sample frequency information and sample tone information of the sample audio;
a sample parameter obtaining unit, configured to obtain a digital equalization processing curve of the sample audio and a reverberation parameter of the sample audio based on the characteristic frequency response curve, the sample frequency information, and the sample timbre information;
the sample parameter adjusting unit is used for acquiring a sample sound effect label of the sample audio, determining a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusting the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensuring that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label so as to obtain the digital equalization parameters corresponding to each octave characteristic point;
the sample information storage unit is used for storing the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
the information acquisition unit is used for acquiring audio data of a target audio when a playing instruction of the target audio is received;
the parameter acquisition unit is used for acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and the data output unit is used for synthesizing the audio data by adopting the target sound effect data packet and sending the audio data after the synthesis processing to the sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
10. The apparatus of claim 9, wherein the sample information holding unit comprises:
a data packet obtaining subunit, configured to compress the digital equalization parameter corresponding to each octave feature point and the reverberation parameter of the sample audio, and store the compressed digital equalization parameter and the reverberation parameter as a sound effect data packet corresponding to the sample sound effect tag;
and the information storage subunit is used for storing the sample sound effect label and the sound effect data packet corresponding to the sample sound effect label into the sound effect parameter set.
11. The device according to claim 9, wherein the information obtaining unit is specifically configured to:
acquiring audio data and an audio tag of the target audio;
the parameter obtaining unit is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
12. The device according to claim 11, wherein the parameter obtaining unit is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
13. The apparatus according to claim 9, wherein the sample information saving unit is specifically configured to save the digital equalization parameter corresponding to each octave feature point, the reverberation parameter of the sample audio, the sample frequency information, the sample timbre information, and the sample audio label into an audio parameter set.
14. The device according to claim 13, wherein the information obtaining unit is specifically configured to:
acquiring audio data, target frequency information and target tone information of the target audio;
the parameter acquisition unit includes:
the similarity obtaining subunit is configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
and the data packet acquisition subunit is used for acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
15. A computer storage medium having stored thereon a plurality of instructions adapted to be loaded by a processor and to perform the steps of:
acquiring sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring sample frequency information and sample tone information of the sample audio;
acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
acquiring a sample sound effect label of the sample audio, determining a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusting the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensuring that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label so as to obtain the digital equalization parameters corresponding to each octave characteristic point;
storing the digital equalization parameters corresponding to each octave characteristic point, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
16. A server, comprising: a processor and a memory; wherein the memory stores a computer program adapted to be loaded by the processor and to perform the steps of:
acquiring sample audio, acquiring a characteristic frequency response curve of the sample audio, and acquiring sample frequency information and sample tone information of the sample audio;
acquiring a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone color information;
acquiring a sample sound effect label of the sample audio, determining a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusting the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensuring that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label so as to obtain the digital equalization parameters corresponding to each octave characteristic point;
storing the digital equalization parameters corresponding to each octave characteristic point, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
when a playing instruction of a target audio is received, audio data of the target audio is obtained;
acquiring a target sound effect data packet corresponding to the target audio in a sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio;
and synthesizing the audio data by adopting the target sound effect data packet, and sending the audio data after the synthesis processing to a sound terminal so that the sound terminal outputs the audio data after the synthesis processing.
17. An audio effect processing apparatus, comprising:
the information receiving unit is used for receiving audio data of the target audio sent by the server when a playing instruction of the target audio is received;
the parameter acquisition unit is used for receiving a sound effect parameter set sent by the server and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio; the sound effect parameter set is obtained by the server according to the acquired sample audio, wherein the acquisition process comprises the following steps: the server acquires a characteristic frequency response curve of the sample audio, acquires sample frequency information and sample tone information of the sample audio, acquires a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone information of the sample audio, acquires a sample sound effect label of the sample audio, determines a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusts the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensures that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label to obtain the digital equalization parameters corresponding to each octave characteristic point, the server stores the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
and the data output unit is used for synthesizing the audio data by adopting the target sound effect data packet and outputting the audio data after the synthesis processing.
18. The device according to claim 17, wherein the information receiving unit is specifically configured to:
receiving audio data and an audio tag of the target audio sent by a server;
the parameter obtaining unit is specifically configured to:
and acquiring a target sound effect data packet corresponding to the audio tag from the sound effect parameter set, and reading the digital equalization parameter of the target audio and the reverberation parameter of the target audio in the target sound effect data packet.
19. The device according to claim 18, wherein the parameter obtaining unit is specifically configured to:
and searching a target sample sound effect label to which the audio label belongs in a sample sound effect label set, and acquiring a target sound effect data packet corresponding to the target sample sound effect label in the sound effect parameter set.
20. The device according to claim 17, wherein the information receiving unit is specifically configured to:
receiving audio data, target frequency information and target tone color information of the target audio sent by a server;
the parameter acquisition unit includes:
the similarity obtaining subunit is configured to match the target frequency information and the target tone information with the sample frequency information and the sample tone information of each sound effect data packet in the sound effect parameter set, and obtain matching similarities between the target frequency information and the target tone information and the sample frequency information and the sample tone information of each sound effect data packet after matching;
and the data packet acquisition subunit is used for acquiring sample frequency information with the highest matching similarity and a sample sound effect data packet corresponding to the sample tone information, and taking the sample sound effect data packet as a target sound effect data packet.
21. A computer storage medium having stored thereon a plurality of instructions adapted to be loaded by a processor and to perform the steps of:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio; the sound effect parameter set is obtained by the server according to the acquired sample audio, wherein the acquisition process comprises the following steps: the server acquires a characteristic frequency response curve of the sample audio, acquires sample frequency information and sample tone information of the sample audio, acquires a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone information of the sample audio, acquires a sample sound effect label of the sample audio, determines a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusts the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensures that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label to obtain the digital equalization parameters corresponding to each octave characteristic point, the server stores the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
and synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
22. An audio terminal, comprising: a processor and a memory; wherein the memory stores a computer program adapted to be loaded by the processor and to perform the steps of:
when a playing instruction of a target audio is received, receiving audio data of the target audio sent by a server;
receiving a sound effect parameter set sent by the server, and acquiring a target sound effect data packet corresponding to the target audio in the sound effect parameter set, wherein the target sound effect data packet comprises a digital equalization parameter of the target audio and a reverberation parameter of the target audio; the sound effect parameter set is obtained by the server according to the acquired sample audio, wherein the acquisition process comprises the following steps: the server acquires a characteristic frequency response curve of the sample audio, acquires sample frequency information and sample tone information of the sample audio, acquires a digital equalization processing curve of the sample audio and reverberation parameters of the sample audio based on the characteristic frequency response curve, the sample frequency information and the sample tone information of the sample audio, acquires a sample sound effect label of the sample audio, determines a plurality of octaves in a selected frequency range in the digital equalization processing curve and one octave characteristic point in each octave, adjusts the digital equalization parameters of each octave characteristic point based on the equal response curve and the sample sound effect label, and ensures that the digital equalization parameters of each octave characteristic point are in a parameter range indicated by the sample sound effect label to obtain the digital equalization parameters corresponding to each octave characteristic point, the server stores the digital equalization parameters corresponding to the octave characteristic points, the reverberation parameters of the sample audios and the sample sound effect labels into a sound effect parameter set;
and synthesizing the audio data by adopting the target sound effect data packet, and outputting the audio data after the synthesis processing.
CN201710999163.9A 2017-10-20 2017-10-20 Sound effect processing method and equipment, storage medium, server and sound terminal thereof Active CN108305603B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710999163.9A CN108305603B (en) 2017-10-20 2017-10-20 Sound effect processing method and equipment, storage medium, server and sound terminal thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710999163.9A CN108305603B (en) 2017-10-20 2017-10-20 Sound effect processing method and equipment, storage medium, server and sound terminal thereof

Publications (2)

Publication Number Publication Date
CN108305603A CN108305603A (en) 2018-07-20
CN108305603B true CN108305603B (en) 2021-07-27

Family

ID=62870103

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710999163.9A Active CN108305603B (en) 2017-10-20 2017-10-20 Sound effect processing method and equipment, storage medium, server and sound terminal thereof

Country Status (1)

Country Link
CN (1) CN108305603B (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109151702B (en) * 2018-09-21 2021-10-08 歌尔科技有限公司 Sound effect adjusting method of audio equipment, audio equipment and readable storage medium
CN109254752B (en) * 2018-09-25 2022-03-15 Oppo广东移动通信有限公司 3D sound effect processing method and related product
CN111045635B (en) * 2018-10-12 2021-05-07 北京微播视界科技有限公司 Audio processing method and device
CN111048108B (en) * 2018-10-12 2022-06-24 北京微播视界科技有限公司 Audio processing method and device
CN109410972B (en) * 2018-11-02 2023-09-01 广州酷狗计算机科技有限公司 Method, device and storage medium for generating sound effect parameters
CN109686347B (en) * 2018-11-30 2021-04-23 北京达佳互联信息技术有限公司 Sound effect processing method, sound effect processing device, electronic equipment and readable medium
CN109686348A (en) * 2018-12-13 2019-04-26 广州艾美网络科技有限公司 A kind of audio processing system restoring professional audio
CN109448740B (en) * 2018-12-18 2022-05-27 网易(杭州)网络有限公司 Voice sound effect processing method and device and voice system
CN109920397B (en) * 2019-01-31 2021-06-01 李奕君 System and method for making audio function in physics
CN109785820B (en) * 2019-03-01 2022-12-27 腾讯音乐娱乐科技(深圳)有限公司 Processing method, device and equipment
CN110297543B (en) * 2019-06-28 2022-05-20 维沃移动通信有限公司 A kind of audio playback method and terminal device
CN112309352B (en) * 2020-01-15 2024-11-22 北京字节跳动网络技术有限公司 Audio information processing method, device, equipment and medium
CN111326132B (en) * 2020-01-22 2021-10-22 北京达佳互联信息技术有限公司 Audio processing method and device, storage medium and electronic equipment
CN114697804B (en) * 2020-12-28 2024-09-17 深圳Tcl数字技术有限公司 Audio equalization method and device, intelligent terminal and computer readable storage medium
CN113641329B (en) * 2021-08-10 2025-03-25 广州艾美网络科技有限公司 Sound effect configuration method and device, smart speaker, computer equipment and storage medium
CN114449339B (en) * 2022-02-16 2024-04-12 深圳万兴软件有限公司 Background sound effect conversion method and device, computer equipment and storage medium

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5652797A (en) * 1992-10-30 1997-07-29 Yamaha Corporation Sound effect imparting apparatus
US20060000347A1 (en) * 2004-06-17 2006-01-05 Preece Kenneth A Acoustical device and method
CN101060316A (en) * 2006-03-31 2007-10-24 索尼株式会社 Signal processing apparatus, signal processing method, and sound field correction system
CN101155438A (en) * 2006-09-26 2008-04-02 张秀丽 Frequency response adaptive equalization method for audio device
CN102622999A (en) * 2011-01-26 2012-08-01 英华达(南京)科技有限公司 System for automatically adjusting sound effect and method thereof
CN103137136A (en) * 2011-11-22 2013-06-05 雅马哈株式会社 Sound processing device
CN103151055A (en) * 2013-03-05 2013-06-12 广东欧珀移动通信有限公司 Method and system for automatically switching sound effect
US20140334630A1 (en) * 2013-05-13 2014-11-13 Sound In Motion Ltd. Adding audio sound effects to movies
CN105808204A (en) * 2016-03-31 2016-07-27 联想(北京)有限公司 Sound effect adjusting method and electronic equipment
CN106126176A (en) * 2016-06-16 2016-11-16 广东欧珀移动通信有限公司 A kind of audio collocation method and mobile terminal
CN106155623A (en) * 2016-06-16 2016-11-23 广东欧珀移动通信有限公司 A kind of audio collocation method, system and relevant device
CN106658304A (en) * 2017-01-11 2017-05-10 广东小天才科技有限公司 Output control method for wearable device audio and wearable device
CN106878642A (en) * 2017-02-13 2017-06-20 微鲸科技有限公司 Audio automatic equalization system and method
CN107071680A (en) * 2017-04-19 2017-08-18 歌尔科技有限公司 A kind of tuning method and apparatus of acoustic product
CN107249080A (en) * 2017-06-26 2017-10-13 维沃移动通信有限公司 A kind of method, device and mobile terminal for adjusting audio
CN107659637A (en) * 2017-09-21 2018-02-02 广州酷狗计算机科技有限公司 Audio method to set up, device, storage medium and terminal

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5652797A (en) * 1992-10-30 1997-07-29 Yamaha Corporation Sound effect imparting apparatus
US20060000347A1 (en) * 2004-06-17 2006-01-05 Preece Kenneth A Acoustical device and method
CN101060316A (en) * 2006-03-31 2007-10-24 索尼株式会社 Signal processing apparatus, signal processing method, and sound field correction system
CN101155438A (en) * 2006-09-26 2008-04-02 张秀丽 Frequency response adaptive equalization method for audio device
CN102622999A (en) * 2011-01-26 2012-08-01 英华达(南京)科技有限公司 System for automatically adjusting sound effect and method thereof
CN103137136A (en) * 2011-11-22 2013-06-05 雅马哈株式会社 Sound processing device
CN103151055A (en) * 2013-03-05 2013-06-12 广东欧珀移动通信有限公司 Method and system for automatically switching sound effect
US20140334630A1 (en) * 2013-05-13 2014-11-13 Sound In Motion Ltd. Adding audio sound effects to movies
CN105808204A (en) * 2016-03-31 2016-07-27 联想(北京)有限公司 Sound effect adjusting method and electronic equipment
CN106126176A (en) * 2016-06-16 2016-11-16 广东欧珀移动通信有限公司 A kind of audio collocation method and mobile terminal
CN106155623A (en) * 2016-06-16 2016-11-23 广东欧珀移动通信有限公司 A kind of audio collocation method, system and relevant device
CN106658304A (en) * 2017-01-11 2017-05-10 广东小天才科技有限公司 Output control method for wearable device audio and wearable device
CN106878642A (en) * 2017-02-13 2017-06-20 微鲸科技有限公司 Audio automatic equalization system and method
CN107071680A (en) * 2017-04-19 2017-08-18 歌尔科技有限公司 A kind of tuning method and apparatus of acoustic product
CN107249080A (en) * 2017-06-26 2017-10-13 维沃移动通信有限公司 A kind of method, device and mobile terminal for adjusting audio
CN107659637A (en) * 2017-09-21 2018-02-02 广州酷狗计算机科技有限公司 Audio method to set up, device, storage medium and terminal

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"automotive audio equalization";S Cecchi;《AES 36th international conference》;20090601;全文 *
"基于SOC的卡拉ok音效处理器设计及实现";叶文杰;《中国优秀硕士学位论文全文数据库信息科技辑》;20170315;全文 *

Also Published As

Publication number Publication date
CN108305603A (en) 2018-07-20

Similar Documents

Publication Publication Date Title
CN108305603B (en) Sound effect processing method and equipment, storage medium, server and sound terminal thereof
CN106898340B (en) Song synthesis method and terminal
US9918174B2 (en) Wireless exchange of data between devices in live events
WO2021103314A1 (en) Listening scene constructing method and related device
KR20190084809A (en) Electronic Device and the Method for Editing Caption by the Device
CN111640411A (en) Audio synthesis method, device and computer readable storage medium
WO2019233361A1 (en) Method and device for adjusting volume of music
WO2020228226A1 (en) Instrumental music detection method and apparatus, and storage medium
CN117079623A (en) Audio noise reduction model training method, singing work processing equipment and medium
WO2011122522A1 (en) Ambient expression selection system, ambient expression selection method, and program
CN115273826A (en) Singing voice recognition model training method, singing voice recognition method and related device
CN102576531B (en) Method and apparatus for processing multi-channel audio signals
CN112420006B (en) Method and device for operating simulated musical instrument assembly, storage medium and computer equipment
CN1717956B (en) Apparatus and method for adapting audio signal according to user's preference
CN105632523A (en) Method and device for regulating sound volume output value of audio data, and terminal
CN113348508B (en) Electronic device, method and computer program
CN105487788B (en) A kind of music information real time acquiring method and device
WO2023030017A1 (en) Audio data processing method and apparatus, device and medium
CN111868821A (en) Method and apparatus for extracting pitch-independent timbre properties from media signals
CN117061945A (en) Terminal device, sound adjustment method, and storage medium
CN106847249B (en) Pronunciation processing method and system
US12219188B2 (en) Cheering support method, cheering support apparatus, and program
CN110289010B (en) Sound collection method, device, equipment and computer storage medium
JP7493412B2 (en) Audio processing device, audio processing system and program
CN111739496B (en) Audio processing method, device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant