[go: up one dir, main page]

CN115604475A - Multi-mode information source joint coding method - Google Patents

Multi-mode information source joint coding method Download PDF

Info

Publication number
CN115604475A
CN115604475A CN202210969884.6A CN202210969884A CN115604475A CN 115604475 A CN115604475 A CN 115604475A CN 202210969884 A CN202210969884 A CN 202210969884A CN 115604475 A CN115604475 A CN 115604475A
Authority
CN
China
Prior art keywords
modal
knowledge base
sources
information sources
different
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210969884.6A
Other languages
Chinese (zh)
Other versions
CN115604475B (en
Inventor
宋晓丹
李甫
高大化
谢雪梅
石光明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN202210969884.6A priority Critical patent/CN115604475B/en
Publication of CN115604475A publication Critical patent/CN115604475A/en
Priority to PCT/CN2023/098536 priority patent/WO2024032119A1/en
Application granted granted Critical
Publication of CN115604475B publication Critical patent/CN115604475B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/91Entropy coding, e.g. variable length coding [VLC] or arithmetic coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/40Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

一种多模态信源联合编码方法,先将多个模态信源经过相应第一编码器提取特征去除每个模态信号内部冗余,得到对应的特征图;然后将多组特征图连接起来输入第二编码器,解耦为共性特征图和个性特征图;共性特征图表示不同模态信源之间的共同部分,个性特征图表示每个模态信源所独有的特征;最后将多个模态信源的个性特征图和共性特征图经过相应解码器解码并重建相应的模态信源,即分别经过熵编码,转换为二进制码流进行存储或者传输;在解码端二进制码流进行熵解码后分别经过相应的解码器恢复得到相应的模态信源;本发明利用不同信源之间的相关性,减少相关信息的重复传输降低传输带宽,降低存储空间;解码端恢复出不同模态信源,具有模态可伸缩性。

Figure 202210969884

A multi-modal information source joint encoding method. Firstly, multiple modal information sources are extracted through the corresponding first encoder to remove the internal redundancy of each modal signal, and the corresponding feature map is obtained; then multiple groups of feature maps are connected. Get up and input the second encoder, decoupled into a common feature map and a personality feature map; the common feature map represents the common part between different modal information sources, and the individual characteristic map represents the unique characteristics of each modal information source; finally The individual characteristic map and common feature map of multiple modal information sources are decoded by the corresponding decoder and the corresponding modal information sources are reconstructed, that is, they are converted into binary code streams for storage or transmission through entropy encoding respectively; the binary code at the decoding end After the stream is entropy decoded, the corresponding modal information source is restored by the corresponding decoder; the present invention utilizes the correlation between different information sources to reduce the repeated transmission of related information, reduce the transmission bandwidth, and reduce the storage space; the decoding end restores the Different modality sources, with modality scalability.

Figure 202210969884

Description

一种多模态信源联合编码方法A Joint Coding Method for Multimodal Information Sources

技术领域technical field

本发明涉及信源编码技术领域,具体涉及一种多模态信源联合编码方法。The invention relates to the technical field of information source coding, in particular to a multi-mode information source joint coding method.

背景技术Background technique

信源编码作为一种基础技术,被广泛应用于各个领域。信源编码是信息时代多媒体技术和互联网技术结合的产物,旨在允许一定失真或者不允许失真前提下,用最少的比特表示信源。高效率的信源编码技术可以在有限带宽下大大提升解码后的信源质量,降低存储空间。例如,根据输入的不同目前有文本压缩、图像压缩(如PNG,BMP,JPEG,BPG,WEBP等压缩标准)、视频压缩(如H.264/AVC, H.265/HEVC,H.266/VVC,VP9,AV1,AVS1,AVS2,AVS3等)、音频编码(如AAC等)等等,这些标准有一个共同的特点,只针对单一种类的输入,例如文本压缩只针对文本输入,图像压缩只针对图像,视频压缩针对图像或者视频,音频编码只针对音频输入,无法对其他形式的进行处理,即便处理也需要经过前处理,而且效率低下。例如,视频压缩编码标准无法直观对文本进行压缩,尽管可以通过预处理将文本组织成视频形式,但是其内容与正常的视频差异大,无实际物理意义,视频编解码标准中的技术并不针对这种非正常信号设计,因此即便强制编码也会效率低下。As a basic technology, source coding is widely used in various fields. Source coding is the product of the combination of multimedia technology and Internet technology in the information age. It aims to express the source with the fewest bits under the premise of allowing certain distortion or not allowing distortion. High-efficiency source coding technology can greatly improve the quality of decoded source and reduce storage space under limited bandwidth. For example, depending on the input, there are currently text compression, image compression (such as PNG, BMP, JPEG, BPG, WEBP and other compression standards), video compression (such as H.264/AVC, H.265/HEVC, H.266/VVC , VP9, AV1, AVS1, AVS2, AVS3, etc.), audio coding (such as AAC, etc.), etc., these standards have a common feature, only for a single type of input, for example, text compression is only for text input, image compression is only for Image and video compression is for images or videos, and audio encoding is only for audio input, and cannot process other forms, even if the processing requires pre-processing, and the efficiency is low. For example, video compression coding standards cannot intuitively compress text. Although text can be organized into video form through preprocessing, its content is very different from normal video and has no actual physical meaning. The technology in video codec standards does not target This is an unusual signal design, so even mandatory encoding would be inefficient.

实际中,经常综合几种模态的数据一起进行某种表达,例如,电视剧电影等最常见的模态包含视频、音频和字幕三种模态,按照以上标准,目前的方案几乎都是对三种模态分别编码,但是实际上该三种模态信号之间存在关联,即存在一定程度的冗余,而现有的独立编码方法无法消除此类冗余,因此是对带宽或者存储空间的一种浪费。因此需要一种能够对多种模态的信号进行联合编码的方法,以去除不同模态信号之间的相关性,降低冗余,从而达到减少带宽和节省存储空间的目的。In practice, data from several modalities are often combined for a certain expression. For example, the most common modalities such as TV dramas and movies include three modalities: video, audio, and subtitles. According to the above standards, the current solutions are almost all three The three modes are encoded separately, but in fact there is a correlation between the three mode signals, that is, there is a certain degree of redundancy, and the existing independent encoding method cannot eliminate such redundancy, so it is a limitation on bandwidth or storage space. a waste. Therefore, a method capable of jointly encoding signals of multiple modalities is needed to remove correlations between signals of different modalities and reduce redundancy, thereby achieving the purpose of reducing bandwidth and saving storage space.

发明内容Contents of the invention

为了克服上述现有技术的缺点,本发明的目的在于提供了一种多模态信源联合编码方法,通过在编码压缩过程中利用不同信源之间的相关性,减少相关信息的重复传输从而降低传输带宽,降低存储空间;解码端根据需要恢复出不同模态信源,即具有模态可伸缩性。In order to overcome the shortcomings of the above-mentioned prior art, the purpose of the present invention is to provide a multi-modal information source joint coding method, which reduces the repeated transmission of related information by utilizing the correlation between different information sources in the encoding and compression process. Reduce the transmission bandwidth and storage space; the decoding end recovers different modal information sources according to the needs, that is, it has modal scalability.

为了达到上述目的,本发明采取的技术方案为:In order to achieve the above object, the technical scheme that the present invention takes is:

一种多模态信源联合编码方法,包括以下步骤:A multimodal information source joint coding method, comprising the following steps:

1)将多个模态信源经过相应第一编码器以提取特征去除每个模态信号内部冗余,得到对应的特征图;1) passing multiple modal information sources through corresponding first encoders to extract features and remove internal redundancy of each modal signal to obtain corresponding feature maps;

2)为了去除不同模态信号之间的相关性,将多组特征图连接起来输入第二编码器,解耦为共性特征图和个性特征图;共性特征图表示不同模态信源之间的共同部分,个性特征图表示每个模态信源所独有的特征;2) In order to remove the correlation between different modal signals, multiple sets of feature maps are connected and input to the second encoder, and decoupled into common feature maps and individual feature maps; the common feature maps represent the differences between different modal sources. The common part, the personality map represents the characteristics unique to each modality source;

3)将多个模态信源的个性特征图和共性特征图经过相应解码器解码并重建相应的模态信源,即分别经过熵编码,转换为二进制码流进行存储或者传输;在解码端二进制码流进行熵解码后分别经过相应的解码器恢复得到相应的模态信源。3) Decode the individual feature maps and common feature maps of multiple modal information sources through the corresponding decoder and reconstruct the corresponding modal information sources, that is, respectively undergo entropy encoding and convert them into binary code streams for storage or transmission; at the decoding end After the binary code stream is entropy decoded, the corresponding modal information sources are recovered through corresponding decoders respectively.

引入了知识库,对多模态信源进行联合编码;知识库是多模态或单模态,多模态知识库是指知识库中存储包含多种不同形式来自不同模态信源的信息;单个或多个模态信源经过“模态解析”获得检索知识库的索引,“模态解析”为了获取知识库节点实体以进行查询和推理。The knowledge base is introduced to jointly encode multi-modal information sources; the knowledge base is multi-modal or single-modal, and the multi-modal knowledge base refers to the storage of information in the knowledge base that contains a variety of different forms from different modal sources. ; Single or multiple modal information sources obtain the index of the retrieval knowledge base through "modal analysis", and "modal analysis" is used to obtain the node entity of the knowledge base for query and reasoning.

所述的多模态知识库一种表现形式中有文本和图像,以节点和边表示,每个节点表示一个实体或者表示文本或者表示图像,每条边表示不同节点之间的关系。One form of representation of the multimodal knowledge base includes text and images, represented by nodes and edges, each node represents an entity or text or image, and each edge represents the relationship between different nodes.

本发明的有益效果为:本发明提出了一种多模态信源联合编码方法,将每个模态信源表征为共性特征和个性特征,不同模态信源之间的共性特征相同,进而实现多个模态信源的联合编码。相比于多个模态信源独立编码,本发明通过在编码压缩过程中利用不同信源之间的相关性,减少相关信息的重复传输从而降低传输带宽和存储空间。同时,解码端可以根据需要恢复出不同模态信源,即具有模态可伸缩性的优点。The beneficial effects of the present invention are: the present invention proposes a multi-modal information source joint coding method, which characterizes each modal information source as common features and individual features, and the common features between different modal information sources are the same, and then Realize the joint coding of multiple modal sources. Compared with independent coding of multiple modal information sources, the present invention utilizes the correlation between different information sources in the process of encoding and compression to reduce repeated transmission of related information, thereby reducing transmission bandwidth and storage space. At the same time, the decoding end can recover information sources of different modalities according to needs, that is, it has the advantage of modal scalability.

本发明在以上多模态联合编码方法基础上,引入了知识库(其中存在与待编码信源的强相关已知信息),增加了先验知识,显式关联了不同模态的信源,在编码过程利用知识库中的先验知识对多模态编码过程进行指导。因而,相比没有知识库的多模态联合编码,能够进一步节省存储空间,降低带宽。On the basis of the above multimodal joint coding method, the present invention introduces a knowledge base (where there is known information strongly related to the information source to be encoded), increases prior knowledge, and explicitly associates information sources of different modes, In the encoding process, the prior knowledge in the knowledge base is used to guide the multimodal encoding process. Therefore, compared with multi-modal joint coding without knowledge base, it can further save storage space and reduce bandwidth.

附图说明Description of drawings

图1为本发明实施例1一种多模态信源联合编码方法的流程图。FIG. 1 is a flow chart of a joint encoding method for multi-modal information sources according to Embodiment 1 of the present invention.

图2为本发明实施例2一种知识库辅助的多模态信源联合编码方法的流程图。FIG. 2 is a flow chart of a knowledge base-assisted joint encoding method for multi-modal information sources according to Embodiment 2 of the present invention.

图3为本发明实施例2中图像和文本多模态知识库。Fig. 3 is an image and text multimodal knowledge base in Embodiment 2 of the present invention.

图4为本发明实施例3一种知识库辅助的多模态信源联合编码方法的流程图。FIG. 4 is a flowchart of a knowledge base-assisted joint encoding method for multi-modal information sources according to Embodiment 3 of the present invention.

具体实施方式detailed description

下面结合附图和实施例对本发明做详细描述。The present invention will be described in detail below in conjunction with the accompanying drawings and embodiments.

实施例1,实施例1给出了给出了两种信源作为输入的例子,一种多模态信源联合编码方法,包括以下步骤:Embodiment 1, embodiment 1 provides the example that two kinds of information sources are given as input, a kind of multimodal information source joint coding method, comprises the following steps:

1)给定两个模态信源“模态1”和“模态2”,分别记为src1和 src2,两个模态信号分别经过第一编码器A和第一编码器B以提取特征去除每个模态信号内部冗余,得到特征图feat1和特征图feat2,其中第一编码器A和第二编码器B无特殊限制可以是神经网络中的卷积神经网络CNN,也可以是时序循环神经网络RNN;特征图feat1和特征图feat2可以是一维向量,也可以是二维矩阵甚至更高维度的张量;1) Given two modal signal sources "modal 1" and "modal 2", denoted as src 1 and src 2 respectively, the two modal signals respectively pass through the first encoder A and the first encoder B to Extract features to remove the internal redundancy of each modal signal, and obtain the feature map feat 1 and feature map feat 2 , wherein the first encoder A and the second encoder B can be the convolutional neural network CNN in the neural network without special restrictions, It can also be a time series recurrent neural network RNN; feature map feat 1 and feature map feat 2 can be one-dimensional vectors, two-dimensional matrices or even higher-dimensional tensors;

2)为了去除不同模态信号之间的相关性,将两组特征图连接起来输入第二编码器C,解耦为共性特征图和个性特征图;共性特征图表示不同模态信源之间的共同部分,通常为语义层面;个性特征图表示每个模态信源所独有的特征;以视频和音频两个模态信源为例,共性特征可能是视频中人物所说的话语,音频中通常也包含该信息;视频的个性特征可以是视频中人物的外表或者人物以外其他如花草等背景信息,音频的个性特征可能包含其他非相关音频,也可以是视频通常难以表达的语气等;2) In order to remove the correlation between different modal signals, the two sets of feature maps are connected and input to the second encoder C, decoupled into common feature maps and individual feature maps; the common feature maps represent the differences between different modal sources The common part of , usually at the semantic level; the personality feature map represents the unique features of each modal source; taking video and audio as examples, the common feature may be the words spoken by the characters in the video, The audio usually also contains this information; the personality characteristics of the video can be the appearance of the characters in the video or other background information other than the characters, such as flowers and plants, and the personality characteristics of the audio may include other irrelevant audio, or it can be the tone that is usually difficult to express in the video, etc. ;

本实施例进行共性和个性特征解耦,输出模态1的个性特征 feati1,两种模态的共性特征featc和模态2的个性特征feati2,第二编码器C中可能包含量化过程以实现有损编码,其结构无特殊要求,既可以是CNN,RNN还可以包含hyper prior模型;另外,需要说明的是,feati1,featc和feati2三类特征内部特性不一定相同,如feati1内部可能包含边信息featis1以及特征featii1,其中的边信息featis1用以辅助featii1生成,featc和feati2同理;In this embodiment, the common and individual features are decoupled, and the individual features feati 1 of modality 1, the common features featc of the two modalities and the individual features feati 2 of modality 2 are output, and the second encoder C may include a quantization process to Realize lossy coding, its structure has no special requirements, it can be CNN, RNN can also include hyper prior model; In addition, it should be noted that the internal characteristics of feati 1 , featc and feati 2 are not necessarily the same, such as feati 1 The interior may contain side information featis 1 and feature featii 1. The side information featis 1 is used to assist the generation of featii 1 , and featc and feati 2 are the same;

3)featc,feati1和feati2三类特征分别经过熵编码,转换为二进制码流进行存储或者传输;在解码端二进制码流进行熵解码后恢复得到feati1,featc和feati2;之后feati1和featc共同输入解码器A,以恢复模态1,标记为

Figure RE-RE-RE-GDA0003905068130000051
featci1和featc共同输入解码器B,以恢复模态2,记为
Figure RE-RE-RE-GDA0003905068130000052
3) featc, feati 1 and feati 2 three types of features are respectively entropy coded and converted into binary code streams for storage or transmission; the binary code streams at the decoding end are restored to obtain feati 1 , featc and feati 2 after entropy decoding; after that feati 1 and featc are jointly input to decoder A to recover modality 1, labeled as
Figure RE-RE-RE-GDA0003905068130000051
Featci 1 and featc are jointly input to decoder B to recover modality 2, denoted as
Figure RE-RE-RE-GDA0003905068130000052

以上为测试时的流程,在训练过程中只需要有成对的多模态数据就可以进行训练,训练过程多个模态的编码器和解码器一起进行端到端训练,损失函数

Figure RE-RE-RE-GDA0003905068130000053
设计为以下形式:The above is the test process. In the training process, only paired multi-modal data can be used for training. During the training process, the encoder and decoder of multiple modalities perform end-to-end training together. The loss function
Figure RE-RE-RE-GDA0003905068130000053
designed to be of the form:

Figure RE-RE-RE-GDA0003905068130000054
Figure RE-RE-RE-GDA0003905068130000054

其中的quality1(·,·)和quality2(·,·)分别用于衡量编码造成的模态1和模态2的质量损失。例如对于对于视频或者图像可以用PSNR (峰值信噪比),MS-SSIM(多尺度-结构相似性)或者感知损失等进行衡量;

Figure RE-RE-RE-GDA0003905068130000061
Figure RE-RE-RE-GDA0003905068130000062
用于衡量
Figure RE-RE-RE-GDA0003905068130000063
Figure RE-RE-RE-GDA0003905068130000066
转换为二进制码流所消耗的比特数量,通常可以通过估计获得。例如上述描述中,可以假设featc,feati1和feati2三类特征服从高斯分布,用featis1中的部分特征表示高斯分布的均值,另外部分特征表示方差,即编码器采用变分自编码器VAE结构,则码率
Figure RE-RE-RE-GDA0003905068130000064
Figure RE-RE-RE-GDA0003905068130000065
可以用香农熵估计得到;公式中的λ1,λ,λ3属于超参数,λ1控制模态1 和模态2重建质量之间的折中,即当更希望模态1的信源失真更小时λ1可以设置比较小,反之亦然;λ3在模态1和模态2之间进行码率分配,即两种模态总的带宽或者存储空间要求一定,λ3较大时倾向于模态1码率更大和模态2码率更小,反之亦然;λ用于控制质量和码率之间的折中,通常质量越高所消耗的码率越大,质量越低所消耗的码率越小,即λ用于选取最终的码率点,λ越大则选取的码率点越低,适用于带宽越低的场景,相应的重建质量会越低,反之亦然。Among them, quality 1 ( , ) and quality 2 ( , ) are used to measure the quality loss of mode 1 and mode 2 caused by encoding, respectively. For example, for videos or images, PSNR (peak signal-to-noise ratio), MS-SSIM (multi-scale-structural similarity) or perceptual loss can be used to measure;
Figure RE-RE-RE-GDA0003905068130000061
and
Figure RE-RE-RE-GDA0003905068130000062
used to measure
Figure RE-RE-RE-GDA0003905068130000063
and
Figure RE-RE-RE-GDA0003905068130000066
The number of bits consumed by converting to a binary code stream can usually be estimated. For example, in the above description, it can be assumed that the featc, feati 1 and feati 2 types of features obey the Gaussian distribution, and some features in featis 1 represent the mean value of the Gaussian distribution, and some other features represent the variance, that is, the encoder uses the variational autoencoder VAE structure, the code rate
Figure RE-RE-RE-GDA0003905068130000064
and
Figure RE-RE-RE-GDA0003905068130000065
It can be estimated by Shannon entropy; λ 1 , λ, λ 3 in the formula belong to hyperparameters, and λ 1 controls the compromise between the reconstruction quality of mode 1 and mode 2, that is, when the source distortion of mode 1 is more expected When it is smaller, λ 1 can be set relatively small, and vice versa; λ 3 performs code rate allocation between mode 1 and mode 2, that is, the total bandwidth or storage space requirements of the two modes are certain, and when λ 3 is larger, it tends to Because the code rate of mode 1 is larger and the code rate of mode 2 is smaller, and vice versa; λ is used to control the compromise between quality and code rate. Usually, the higher the quality, the greater the code rate consumed, and the lower the quality. The smaller the code rate consumed, that is, λ is used to select the final code rate point, the larger the λ, the lower the selected code rate point, which is suitable for scenarios with lower bandwidth, and the corresponding reconstruction quality will be lower, and vice versa.

实施例2,参照图2,实施例2在实施例1的基础上引入了知识库,可以更高效的对多模态信源进行联合编码。Embodiment 2, referring to FIG. 2 , embodiment 2 introduces a knowledge base on the basis of embodiment 1, which can perform joint coding on multi-modal information sources more efficiently.

图2中的知识库既可以是多模态也可以是单模态,多模态知识库是指知识库中存储包含不同形式的信息(通常来自不同模态信源);图3以文本和图像为例给出了一种多模态知识库的例子,其中的多模态知识库有文本和图像,以节点和边表示,每个节点表示一个实体或者表示文本或者表示图像,每条边表示不同节点之间的关系,例如 Claude Shannon是World Computer Chess Championship的嘉宾,其中“Claude Shannon”和“WorldComputer Chess Championship”是节点,边“guestOf”表示两者关系。图3右下角给出了Claude Shannon的图像,“Claude Shannon”和其图像使用带方向的边“imageOf”相连;“Deep Thought”参加“World Computer Chess Championship”比赛,通过两个节点分别表示“Deep Thought”和“World Computer Chess Championship”,通过“attend”表示两者之间的关系。The knowledge base in Figure 2 can be either multi-modal or unimodal, and the multi-modal knowledge base means that the knowledge base stores different forms of information (usually from different modal sources); Figure 3 uses text and Image as an example gives an example of a multi-modal knowledge base. The multi-modal knowledge base has text and images, represented by nodes and edges. Each node represents an entity or represents text or represents an image. Each edge Represents the relationship between different nodes. For example, Claude Shannon is a guest of the World Computer Chess Championship, where "Claude Shannon" and "WorldComputer Chess Championship" are nodes, and the edge "guestOf" represents the relationship between the two. The image of Claude Shannon is shown in the lower right corner of Figure 3. "Claude Shannon" and its image are connected by the edge "imageOf" with direction; "Deep Thought" participates in the "World Computer Chess Championship" competition, and the two nodes represent "Deep Thought" and "World Computer Chess Championship", the relationship between the two is represented by "attend".

实施例2在实施例1的基础上引入了知识库,在实施例1的基础上,模态1信源经过“模态1解析”可以获得检索知识库的索引,模态2信源经过“模态2解析”也可以获得检索知识库的索引,两者只有其一也可以,有两种解析可以从知识库中检索到更多的相关信息或者增强鲁棒性,对多模态信源的编码效率提升作用更大。其中“模态 1解析”和“模态2解析”主要为了获取知识库节点实体以进行查询和推理。经过知识库的推理和查询后,相关信息可以经过第三编码器 D进行嵌入编码得到知识库特征,与信源特征通过第二编码器C进行联合编码,去除信源编码与知识库的冗余,从而提升编码效率。相应的在解码过程解码器A和解码器B也需要输入知识库特征对模态1 和模态2信源进行解码。Embodiment 2 introduces the knowledge base on the basis of Embodiment 1. On the basis of Embodiment 1, the information source of mode 1 can obtain the index of the retrieval knowledge base through "analysis of mode 1", and the information source of mode 2 can be obtained through "analysis of mode 1" Mode 2 analysis" can also obtain the index of the retrieval knowledge base, and only one of the two can be used. There are two types of analysis that can retrieve more relevant information from the knowledge base or enhance robustness. For multi-modal information sources The coding efficiency improvement effect is greater. Among them, "modal 1 parsing" and "modal 2 parsing" are mainly for obtaining knowledge base node entities for query and reasoning. After the reasoning and query of the knowledge base, the relevant information can be embedded and encoded by the third encoder D to obtain the features of the knowledge base, and jointly encoded with the source features by the second encoder C to remove the redundancy between the source code and the knowledge base , so as to improve the coding efficiency. Correspondingly, in the decoding process, decoder A and decoder B also need to input knowledge base features to decode the mode 1 and mode 2 sources.

实施例2所引入的知识库的目的在于增加了先验知识;显式关联了不同模态的信源。The purpose of the knowledge base introduced in Embodiment 2 is to increase prior knowledge; to explicitly associate information sources of different modalities.

实施例2的具体流程为:一种多模态信源联合编码方法,包括以下步骤:The specific process of embodiment 2 is: a method for joint encoding of multi-modal information sources, including the following steps:

1)给定两个模态信源“模态1”和“模态2”,分别记为src1和 src2,两个模态信号分别经过第一编码器A和第一编码器B以提取特征去除每个模态信号内部冗余,得到特征图feat1和特征图feat21) Given two modal signal sources "modal 1" and "modal 2", denoted as src 1 and src 2 respectively, the two modal signals respectively pass through the first encoder A and the first encoder B to Extract features to remove the internal redundancy of each modal signal, and obtain feature maps feat 1 and feature maps feat 2 ;

模态1信源经过“模态1解析”获得检索知识库的索引,模态2 信源经过“模态2解析”获得检索知识库的索引,其中“模态1解析”和“模态2解析”主要为了获取知识库节点实体以进行查询和推理;经过知识库的推理和查询后,相关信息经过编码器D进行嵌入编码得到知识库特征;The information source of mode 1 obtains the index of the retrieval knowledge base through "analysis of mode 1", and the information source of mode 2 obtains the index of the retrieval knowledge base through "analysis of mode 2", where "analysis of mode 1" and "analysis of mode 2 "Analysis" is mainly to obtain knowledge base node entities for query and reasoning; after knowledge base reasoning and query, relevant information is embedded and coded by encoder D to obtain knowledge base features;

2)为了去除不同模态信号之间的相关性,将两组特征图连接起来输入第二编码器C,解耦为共性特征图和个性特征图;共性特征图表示不同模态信源之间的共同部分,通常为语义层面;个性特征图表示每个模态信源所独有的特征;以视频和音频两个模态信源为例,共性特征可能是视频中人物所说的话语,音频中通常也包含该信息;视频的个性特征可以是视频中人物的外表或者人物以外其他如花草等北京信息,音频的个性特征可能包含其他非相关音频,也可以是视频通常难以表达的语气等;2) In order to remove the correlation between different modal signals, the two sets of feature maps are connected and input to the second encoder C, decoupled into common feature maps and individual feature maps; the common feature maps represent the differences between different modal sources The common part of , usually at the semantic level; the personality feature map represents the unique features of each modal source; taking video and audio as examples, the common feature may be the words spoken by the characters in the video, The audio usually also contains this information; the personality characteristics of the video can be the appearance of the characters in the video or other Beijing information other than the characters, such as flowers and plants, and the personality characteristics of the audio may include other irrelevant audio, or the tone that is usually difficult to express in the video ;

本实施例进行共性和个性特征解耦,输出模态1的个性特征 feati1,两种模态的共性特征featc和模态2的个性特征feati2,第二编码器C中可能包含量化过程以实现有损编码;In this embodiment, the common and individual features are decoupled, and the individual features feati 1 of modality 1, the common features featc of the two modalities and the individual features feati 2 of modality 2 are output, and the second encoder C may include a quantization process to Implement lossy encoding;

知识库特征与信源特征通过第二编码器C进行联合编码,去除信源编码与知识库的冗余,从而提升编码效率;The knowledge base feature and the information source feature are jointly encoded by the second encoder C to remove the redundancy of the information source encoding and the knowledge base, thereby improving the encoding efficiency;

3)featc,feati1和feati2三类特征分别经过熵编码,转换为二进制码流进行存储或者传输;在解码端二进制码流进行熵解码后恢复得到feati1,featc和feati2;之后feati1和featc共同输入解码器A,以恢复模态1,标记为

Figure RE-RE-RE-GDA0003905068130000091
feati1和featc共同输入解码器B,以恢复模态2,记为
Figure RE-RE-RE-GDA0003905068130000092
3) featc, feati 1 and feati 2 three types of features are respectively entropy coded and converted into binary code streams for storage or transmission; the binary code streams at the decoding end are restored to obtain feati 1 , featc and feati 2 after entropy decoding; after that feati 1 and featc are jointly input to decoder A to recover modality 1, labeled as
Figure RE-RE-RE-GDA0003905068130000091
Feati 1 and featc are jointly input to decoder B to recover modality 2, denoted as
Figure RE-RE-RE-GDA0003905068130000092

在解码过程解码器A和解码器B也需要输入知识库特征对模态1 和模态2信源进行解码。In the decoding process, Decoder A and Decoder B also need to input knowledge base features to decode Modal 1 and Modal 2 sources.

实施例3,参照图4,实施例3给出了引入知识库的一种实施例,知识库所起的作用在于根据“文本”信源中的“Claude Shannon”关键字查询知识库中可以获得其本人的图像,从而无需编码“图像”信源中的Claude Shannon所对应的图像部分,因而可以更高效的对图像进行图像和文本进行编码。Embodiment 3, with reference to Fig. 4, embodiment 3 provides a kind of embodiment that introduces knowledge base, and the effect that knowledge base plays is that can obtain in the knowledge base according to " Claude Shannon " keyword query in " text " information source There is no need to encode the image part corresponding to Claude Shannon in the "image" source, so that the image and text can be encoded more efficiently.

参照图4,本实施例的输入为“文本”和“图像”两种模态信源,分别对应实施例2的图2中的“模态1”和“模态2”,文本信源中的“命名实体识别:BERT”对应“模态1解析”,即可以借鉴自然语言处理领域中的BERT技术对文本中的命名实体解析得到实体名称,如“ClaudeShannon”和“Deep Thought”,输入到知识库中进行查询与推理,经过编码之后生成知识库特征该特征通常是嵌入后的特征向量;图4中未对模态2进行解析,即没有利用图2中的“模态 2解析”。对于主分支,“文本”模态经过文本编码器,如GRU可以编码为文本特征,“图像”模态经过场景图生成技术检测图像中的目标并建立目标之间的关系,该场景图经过卷积网络生成图像特征图,标记为图像特征。之后,文本特征、图像特征连接后,与知识库特征共同作为输入送入第二编码器C进行编码,生成文本个性特征、图像个性特征以及文本和图像的共性特征。图4中未展示将特征进行无损编码生成二进制码流的过程,以及对二进制码流进行解码生成对应特征的部分。除此之外,解析到的“实体名称”也需要编码传输到解码端。With reference to Fig. 4, the input of this embodiment is two kinds of modal information sources of "text" and "image", respectively corresponding to "modal 1" and "modal 2" in Fig. 2 of embodiment 2, in the text information source The "Named Entity Recognition: BERT" corresponds to "Modality 1 Analysis", that is, the BERT technology in the field of natural language processing can be used to analyze the named entities in the text to obtain entity names, such as "Claude Shannon" and "Deep Thought", input to Query and reasoning are performed in the knowledge base, and the knowledge base features are generated after encoding. The features are usually embedded feature vectors; in Figure 4, mode 2 is not analyzed, that is, the "mode 2 analysis" in Figure 2 is not used. For the main branch, the "text" modality passes through a text encoder, such as GRU, which can be encoded into text features, and the "image" modality detects objects in the image and establishes the relationship between them through scene graph generation technology, which is passed through volume The product network generates image feature maps, which are labeled as image features. Afterwards, after the text features and image features are connected, they are sent together with the knowledge base features as input to the second encoder C for encoding to generate text individual features, image individual features, and common features of text and images. Figure 4 does not show the process of losslessly encoding features to generate binary code streams, and decoding the binary code streams to generate corresponding features. In addition, the parsed "entity name" also needs to be encoded and transmitted to the decoding end.

在解码端,文本个性特征、共性特征和知识库特征共同作为输入,经过文本解码器输出文本;图像个性特征、共性特征和知识库特征共同作为输入经过图像解码器输出图像。从图4可以看到,通过引入知识库,编码端无需传输图像中Claude Shannon对应的部分,只需传输解析到的Claude Shannon实体,解码端通过知识库可以获得知识库中的ClaudeShannon对应的图像;此外,也不需要传输编码“Edmonton”和“1989”,可以通过知识库传输推理得到。编码端图像的个性特征主要包含“Feng-hsiung Hsu”的衣着,姿态和位置特征。文本中的个性特征主要包含“Feng-hsiung Hsu”和“first prize”;共性特征包含“ClaudeShannon”和“Deep Thought”等信息。因此,增加知识库后使得编码更加高效。本实施例的训练过程和实施例1类似,损失函数的设计也类似。At the decoding end, the text individual features, common features and knowledge base features are used as input together, and the text is output through the text decoder; the image individual features, common features and knowledge base features are jointly used as input and the image output image is output through the image decoder. It can be seen from Figure 4 that by introducing the knowledge base, the encoding end does not need to transmit the part corresponding to Claude Shannon in the image, but only needs to transmit the parsed Claude Shannon entity, and the decoding end can obtain the image corresponding to Claude Shannon in the knowledge base through the knowledge base; In addition, transfer codes "Edmonton" and "1989" are also not required, which can be inferred through knowledge base transfer. The personality characteristics of the image on the encoding side mainly include the clothing, posture and location characteristics of "Feng-hsiung Hsu". Personality features in the text mainly include "Feng-hsiung Hsu" and "first prize"; common features include information such as "Claude Shannon" and "Deep Thought". Therefore, increasing the knowledge base makes coding more efficient. The training process of this embodiment is similar to that of Embodiment 1, and the design of the loss function is also similar.

Claims (3)

1. A method for multi-modal joint source coding, comprising the steps of:
1) A plurality of modal signal sources pass through corresponding first encoders to extract features to remove internal redundancy of each modal signal, and a corresponding feature map is obtained;
2) In order to remove the correlation among different modal signals, connecting a plurality of groups of feature maps, inputting the feature maps into a second encoder, and decoupling the feature maps into a common feature map and an individual feature map; the common characteristic diagram represents a common part among different modality information sources, and the individual characteristic diagram represents a characteristic unique to each modality information source;
3) The individual characteristic graphs and the common characteristic graphs of the plurality of modal information sources are decoded by corresponding decoders and corresponding modal information sources are rebuilt, namely, the individual characteristic graphs and the common characteristic graphs are respectively subjected to entropy coding and converted into binary code streams to be stored or transmitted; and after entropy decoding is carried out on the binary code stream at the decoding end, recovering the binary code stream through corresponding decoders respectively to obtain corresponding modal information sources.
2. The method of claim 1, wherein: introducing a knowledge base, and carrying out joint coding on the multi-mode information source; the knowledge base is multi-mode or single-mode, and the multi-mode knowledge base stores information containing a plurality of different forms from different mode information sources; the single or multiple modal sources obtain indexes for searching the knowledge base through modal analysis, and the modal analysis is used for obtaining knowledge base node entities for query and reasoning.
3. The method of claim 2, wherein: the multi-modal knowledge base has a representation form of text and images which are represented by nodes and edges, wherein each node represents an entity or represents the text or represents the image, and each edge represents the relationship between different nodes.
CN202210969884.6A 2022-08-12 2022-08-12 Multi-mode information source joint coding method Active CN115604475B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202210969884.6A CN115604475B (en) 2022-08-12 2022-08-12 Multi-mode information source joint coding method
PCT/CN2023/098536 WO2024032119A1 (en) 2022-08-12 2023-06-06 Joint encoding method for multiple modality information sources

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210969884.6A CN115604475B (en) 2022-08-12 2022-08-12 Multi-mode information source joint coding method

Publications (2)

Publication Number Publication Date
CN115604475A true CN115604475A (en) 2023-01-13
CN115604475B CN115604475B (en) 2025-06-10

Family

ID=84843969

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210969884.6A Active CN115604475B (en) 2022-08-12 2022-08-12 Multi-mode information source joint coding method

Country Status (2)

Country Link
CN (1) CN115604475B (en)
WO (1) WO2024032119A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024032119A1 (en) * 2022-08-12 2024-02-15 西安电子科技大学 Joint encoding method for multiple modality information sources
WO2024183179A1 (en) * 2023-03-04 2024-09-12 北京邮电大学 Model-based information traffic service providing method and system, and device and medium
WO2024221743A1 (en) * 2023-04-24 2024-10-31 北京邮电大学 Service information transmission method for intelligent network air interface, and related device
WO2025026013A1 (en) * 2023-08-02 2025-02-06 阿里巴巴(中国)有限公司 Multi-modal task processing method and system, multi-modal dialogue task processing method and system, and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101141644A (en) * 2007-10-17 2008-03-12 清华大学 Coding integration system and method and decoding integration system and method
CN110928961A (en) * 2019-11-14 2020-03-27 出门问问(苏州)信息科技有限公司 Multi-mode entity linking method, equipment and computer readable storage medium
CN112151030A (en) * 2020-09-07 2020-12-29 中国人民解放军军事科学院国防科技创新研究院 Multi-mode-based complex scene voice recognition method and device
US20210281867A1 (en) * 2020-03-03 2021-09-09 Qualcomm Incorporated Video compression using recurrent-based machine learning systems
CN114520915A (en) * 2017-01-12 2022-05-20 索尼公司 Image processing apparatus, image processing method, and computer-readable recording medium
CN114567781A (en) * 2020-11-27 2022-05-31 安徽寒武纪信息科技有限公司 Method, device, electronic equipment and storage medium for coding and decoding video image
US20220210111A1 (en) * 2020-12-29 2022-06-30 Meta Platforms, Inc. Generating Context-Aware Rendering of Media Contents for Assistant Systems

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112119411A (en) * 2018-05-14 2020-12-22 宽腾矽公司 System and method for integrating statistical models of different data modalities
JP6879344B2 (en) * 2019-08-22 2021-06-02 三菱電機株式会社 Decryptor
CN110807122B (en) * 2019-10-18 2022-07-08 浙江大学 Image-text cross-modal feature disentanglement method based on depth mutual information constraint
CN112800292B (en) * 2021-01-15 2022-10-11 南京邮电大学 Cross-modal retrieval method based on modal specific and shared feature learning
CN113591902B (en) * 2021-06-11 2022-09-27 中国科学院自动化研究所 Cross-modal understanding and generation method and device based on multimodal pre-training model
CN114202024A (en) * 2021-12-06 2022-03-18 深圳市安软科技股份有限公司 A training method, system and related equipment for a multimodal decoupling generative model
CN115604475B (en) * 2022-08-12 2025-06-10 西安电子科技大学 Multi-mode information source joint coding method

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101141644A (en) * 2007-10-17 2008-03-12 清华大学 Coding integration system and method and decoding integration system and method
CN114520915A (en) * 2017-01-12 2022-05-20 索尼公司 Image processing apparatus, image processing method, and computer-readable recording medium
CN110928961A (en) * 2019-11-14 2020-03-27 出门问问(苏州)信息科技有限公司 Multi-mode entity linking method, equipment and computer readable storage medium
US20210281867A1 (en) * 2020-03-03 2021-09-09 Qualcomm Incorporated Video compression using recurrent-based machine learning systems
CN112151030A (en) * 2020-09-07 2020-12-29 中国人民解放军军事科学院国防科技创新研究院 Multi-mode-based complex scene voice recognition method and device
CN114567781A (en) * 2020-11-27 2022-05-31 安徽寒武纪信息科技有限公司 Method, device, electronic equipment and storage medium for coding and decoding video image
US20220210111A1 (en) * 2020-12-29 2022-06-30 Meta Platforms, Inc. Generating Context-Aware Rendering of Media Contents for Assistant Systems

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024032119A1 (en) * 2022-08-12 2024-02-15 西安电子科技大学 Joint encoding method for multiple modality information sources
WO2024183179A1 (en) * 2023-03-04 2024-09-12 北京邮电大学 Model-based information traffic service providing method and system, and device and medium
WO2024221743A1 (en) * 2023-04-24 2024-10-31 北京邮电大学 Service information transmission method for intelligent network air interface, and related device
WO2025026013A1 (en) * 2023-08-02 2025-02-06 阿里巴巴(中国)有限公司 Multi-modal task processing method and system, multi-modal dialogue task processing method and system, and device

Also Published As

Publication number Publication date
WO2024032119A1 (en) 2024-02-15
CN115604475B (en) 2025-06-10

Similar Documents

Publication Publication Date Title
CN115604475A (en) Multi-mode information source joint coding method
US11115663B2 (en) Codebook generation for cloud-based video applications
CN113747163B (en) Image coding and decoding method and compression method based on context recombination modeling
CN110290387A (en) A Generative Model-Based Image Compression Method
CN100403801C (en) A context-based adaptive entropy encoding/decoding method
CN114461816B (en) Knowledge-graph-based implementation method of information supplement semantic communication system
CN116233445B (en) Video encoding and decoding processing method and device, computer equipment and storage medium
CN113132735A (en) Video coding method based on video frame generation
CN116437089B (en) A Deep Video Compression Method Based on Key Object
Zhang et al. Learned scalable image compression with bidirectional context disentanglement network
CN117692662A (en) A point cloud encoding and decoding method based on double octree structure
CN113641846A (en) A Cross-modal Retrieval Model Based on Strong Representation Deep Hashing
Saravanan et al. Enhancing efficiency of huffman coding using Lempel Ziv coding for image compression
Gao et al. Cross modal compression with variable rate prompt
CN112866715B (en) Universal video compression coding system supporting man-machine hybrid intelligence
CN103533353A (en) Approximate video encoding system
CN118101961A (en) Quality scalable video coding method based on implicit neural representation
CN117197262B (en) Semantic scalable image coding method, system, device and storage medium
KR20240137005A (en) Data processing methods, devices and media
CN115866266A (en) A multi-bit-rate deep image compression system and method applied in mixed context
CN115379291A (en) Code table updating method, device, equipment and storage medium
Sangeeta et al. Comprehensive Analysis of Flow Incorporated Neural Network-based Lightweight Video Compression Architecture
CN103200408B (en) A kind of video coding-decoding method
KR102072576B1 (en) Apparatus and method for encoding and decoding of data
Luo Compressible and searchable: Ai-native multi-modal retrieval system with learned image compression

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant