[go: up one dir, main page]

CN118101970B - Efficient communication method of monitoring images for ice and snow sports venues based on deep learning - Google Patents

Efficient communication method of monitoring images for ice and snow sports venues based on deep learning Download PDF

Info

Publication number
CN118101970B
CN118101970B CN202410458073.9A CN202410458073A CN118101970B CN 118101970 B CN118101970 B CN 118101970B CN 202410458073 A CN202410458073 A CN 202410458073A CN 118101970 B CN118101970 B CN 118101970B
Authority
CN
China
Prior art keywords
vector
point
snow
optical flow
video frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202410458073.9A
Other languages
Chinese (zh)
Other versions
CN118101970A (en
Inventor
纪沙
鲁承琨
林志东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Normal University
Original Assignee
Harbin Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Normal University filed Critical Harbin Normal University
Priority to CN202410458073.9A priority Critical patent/CN118101970B/en
Publication of CN118101970A publication Critical patent/CN118101970A/en
Application granted granted Critical
Publication of CN118101970B publication Critical patent/CN118101970B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/62Extraction of image or video features relating to a temporal dimension, e.g. time-based feature extraction; Pattern tracking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • H04N19/172Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object the region being a picture, frame or field
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/70Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30221Sports video; Sports image
    • G06T2207/30228Playing field
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30232Surveillance

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

本申请涉及视频监控技术领域,具体涉及基于深度学习的冰雪项目场地监控图像高效通信方法,该方法包括:采集冰雪项目场地监控视频中的各视频帧;基于视频帧各采样点在不同通道下的强度值以及与任一光流向量的位置、距离分布,确定视频帧各采样点的雪地运动赛道能量函数值;基于雪地运动赛道能量函数值以及矢量点在迭代过程中寻找质心的移动位置,确定矢量点的雪地信息密度分裂权重;优化VQ‑LGB压缩算法,得到压缩后的视频帧,输入神经网络实现冰雪项目场地的高效通信。本申请旨在获得更高质量的冰雪项目场地视频压缩数据,提高通信效率。

The present application relates to the field of video surveillance technology, and specifically to an efficient communication method for monitoring images of ice and snow venues based on deep learning, the method comprising: collecting each video frame in the monitoring video of the ice and snow venue; determining the energy function value of the snow sports track of each sampling point in the video frame based on the intensity value of each sampling point in the video frame under different channels and the position and distance distribution with any optical flow vector; determining the snow information density splitting weight of the vector point based on the energy function value of the snow sports track and the moving position of the centroid of the vector point in the iteration process; optimizing the VQ-LGB compression algorithm to obtain the compressed video frame, and inputting the compressed video frame into the neural network to realize efficient communication of the ice and snow venue. The present application aims to obtain higher quality video compression data of the ice and snow venue and improve communication efficiency.

Description

基于深度学习的冰雪项目场地监控图像高效通信方法Efficient communication method of monitoring images for ice and snow sports venues based on deep learning

技术领域Technical Field

本申请涉及视频监控技术领域,具体涉及基于深度学习的冰雪项目场地监控图像高效通信方法。The present application relates to the field of video surveillance technology, and in particular to an efficient communication method for monitoring images of ice and snow project sites based on deep learning.

背景技术Background technique

冰雪项目作为一种体育运动,在日常的场地当中需要对参与运动的人员进行安全监控。由于部分冰雪运动是高速运动,当场地中出现了人员摔倒等情况时,应该及时做出反应,避免摔倒人员遭受碰撞进而受到二次伤害。而冰雪场地过大,对其进行长时间的纯人工检测会造成监控人员疲劳,进而出现监管不力的情况。As a kind of sports, ice and snow events require safety monitoring of participants in daily venues. Since some ice and snow sports are high-speed sports, when people fall in the venue, timely response should be made to prevent the fallen people from being hit and then suffering secondary injuries. However, if the ice and snow venue is too large, long-term pure manual inspection will cause fatigue of the monitoring personnel, resulting in poor supervision.

因此本申请采用深度学习的方法实时采集冰雪项目场地的视频帧,在监控终端进行集中,识别出画面中的摔倒人员并发出提醒。由于冰雪项目场地通常过大,导致实时采集的数据过多,因此需要对采集的视频帧进行通信前后需要进行压缩和解压处理,保证通信的高效。本申请采用VQ-RGB压缩算法,为进一步提高冰雪场地监控的实时性和监控质量,对传统VQ-RGB压缩算法进行改进,使监控视频的通信质量得到提高。Therefore, this application uses a deep learning method to collect video frames of ice and snow venues in real time, collects them at the monitoring terminal, identifies the fallen person in the picture and issues a reminder. Because ice and snow venues are usually too large, resulting in too much data collected in real time, the collected video frames need to be compressed and decompressed before and after communication to ensure efficient communication. This application uses the VQ-RGB compression algorithm to further improve the real-time and monitoring quality of ice and snow venue monitoring, and improves the traditional VQ-RGB compression algorithm to improve the communication quality of the monitoring video.

发明内容Summary of the invention

为了解决上述技术问题,本申请提供基于深度学习的冰雪项目场地监控图像高效通信方法,以解决现有的问题。In order to solve the above technical problems, the present application provides an efficient communication method for monitoring images of ice and snow project sites based on deep learning to solve the existing problems.

本申请的基于深度学习的冰雪项目场地监控图像高效通信方法采用如下技术方案:The deep learning-based efficient communication method for monitoring images of ice and snow sports venues in this application adopts the following technical solutions:

本申请一个实施例提供了基于深度学习的冰雪项目场地监控图像高效通信方法,该方法包括以下步骤:An embodiment of the present application provides an efficient communication method for monitoring images of ice and snow sports venues based on deep learning, the method comprising the following steps:

采集冰雪项目场地监控视频中的各视频帧;Collect each video frame from the monitoring video of the ice and snow event venue;

基于视频帧各采样点在不同通道下的强度值,确定各采样点的雪地色彩突出分割权重;Based on the intensity values of each sampling point in the video frame under different channels, the snow color highlight segmentation weight of each sampling point is determined;

采用稀疏光流算法获取相邻视频帧的光流向量;基于视频帧中各采样点与任一光流向量的位置分布,确定各采样点与任一光流向量的方向吻合度;The sparse optical flow algorithm is used to obtain the optical flow vectors of adjacent video frames; based on the position distribution of each sampling point in the video frame and any optical flow vector, the degree of coincidence between each sampling point and any optical flow vector is determined;

基于相邻视频帧中光流向量的距离分布确定任一光流向量的光流向量非震荡度;基于连续参考帧在相同采样点的雪地色彩突出分割权重、光流向量非震荡度以及方向吻合度,确定视频帧各采样点的雪地运动赛道能量函数值;The optical flow vector non-oscillation degree of any optical flow vector is determined based on the distance distribution of the optical flow vectors in adjacent video frames; the energy function value of the snow sports track at each sampling point of the video frame is determined based on the snow color highlight segmentation weight, optical flow vector non-oscillation degree and direction consistency at the same sampling point of the continuous reference frame;

随机生成视频帧中的矢量点,采用VQ-LGB算法中的LGB算法对视频帧的任意一次迭代过程中,基于雪地运动赛道能量函数值以及矢量点在迭代过程中寻找质心的移动位置,确定矢量点的雪地信息密度分裂权重;Vector points in the video frame are randomly generated, and the snow information density splitting weight of the vector point is determined based on the energy function value of the snow sports track and the moving position of the centroid of the vector point in the iteration process during any iteration of the video frame using the LGB algorithm in the VQ-LGB algorithm;

基于雪地信息密度分裂权重优化VQ-LGB压缩算法,得到压缩后的视频帧,输入神经网络实现冰雪项目场地的高效通信;Based on the snow information density split weight optimization VQ-LGB compression algorithm, the compressed video frames are obtained and input into the neural network to achieve efficient communication in the ice and snow sports venues;

所述基于视频帧各采样点在不同通道下的强度值,确定各采样点的雪地色彩突出分割权重,表达式为:Based on the intensity values of each sampling point in the video frame under different channels, the snow color highlight segmentation weight of each sampling point is determined, and the expression is:

式中,是第m个采样点的雪地色彩突出分割权重,分别是第m个采样点在暗通道、亮通道视频帧中的强度值,分别是第m个采样点在R、G、B通道中的强度值;In the formula, is the snow color highlight segmentation weight of the mth sampling point, , are the intensity values of the mth sampling point in the dark channel and bright channel video frames, respectively. , , are the intensity values of the mth sampling point in the R, G, and B channels respectively;

所述基于视频帧中各采样点与任一光流向量的位置分布,确定各采样点与任一光流向量的方向吻合度,包括:将各采样点与任一光流向量的中点进行连线,将所述连线与任一光流向量的夹角作为各采样点与任一光流向量的方向吻合度;Determining the degree of direction fit between each sampling point and any optical flow vector based on the position distribution of each sampling point and any optical flow vector in the video frame includes: connecting each sampling point with the midpoint of any optical flow vector, and taking the angle between the connecting line and any optical flow vector as the degree of direction fit between each sampling point and any optical flow vector;

所述基于相邻视频帧中光流向量的距离分布确定任一光流向量的光流向量非震荡度,包括:The step of determining the optical flow vector non-oscillation degree of any optical flow vector based on the distance distribution of the optical flow vectors in adjacent video frames includes:

对于每个视频帧中的任一光流向量,将其在相邻前一帧的视频帧中,距离所述任一光流向量的中点最近的中点所在光流向量,记为任一光流向量的关联光流向量;For any optical flow vector in each video frame, the optical flow vector whose midpoint is closest to the midpoint of any optical flow vector in the video frame of the previous adjacent frame is recorded as the associated optical flow vector of any optical flow vector;

将任一光流向量的所有关联光流向量的长度的标准差记为任一光流向量的光流向量非震荡度。The standard deviation of the lengths of all associated optical flow vectors of any optical flow vector is recorded as the optical flow vector non-oscillation degree of any optical flow vector.

优选的,所述基于连续参考帧在相同采样点的雪地色彩突出分割权重、光流向量非震荡度以及方向吻合度,确定视频帧各采样点的雪地运动赛道能量函数值,包括:Preferably, the method of determining the snow sports track energy function value of each sampling point of the video frame based on the snow color highlight segmentation weight, optical flow vector non-oscillation degree and direction consistency at the same sampling point of the continuous reference frame includes:

将任一视频帧相邻的前预设数量个视频帧,记为任一视频帧的参考帧;Recording a preset number of video frames adjacent to any video frame as reference frames of any video frame;

对于参考帧中的任一光流向量,计算任一光流向量的长度与参考帧中各采样点的雪地色彩突出分割权重的乘积,作为分子;For any optical flow vector in the reference frame, the product of the length of any optical flow vector and the snow color highlight segmentation weight of each sampling point in the reference frame is calculated as the numerator;

计算任一光流向量的光流向量非震荡度、参考帧中各采样点与任一光流向量的方向吻合度、参考帧中各采样点与任一光流向量的欧式距离的三项乘积,将所述三项乘积与预设调参因子的和值作为分母;Calculate the product of the non-oscillation degree of the optical flow vector of any optical flow vector, the degree of direction consistency between each sampling point in the reference frame and any optical flow vector, and the Euclidean distance between each sampling point in the reference frame and any optical flow vector, and use the sum of the three products and the preset parameter adjustment factor as the denominator;

计算分子与分母的比值;将所有参考帧中所有的光流向量的所述比值的和值,作为任一视频帧各采样点的雪地运动赛道能量函数值。Calculate the ratio of the numerator to the denominator; and use the sum of the ratios of all optical flow vectors in all reference frames as the energy function value of the snow sports track at each sampling point in any video frame.

优选的,所述基于雪地运动赛道能量函数值以及矢量点在迭代过程中寻找质心的移动位置,确定矢量点的雪地信息密度分裂权重,包括:Preferably, the method of finding the moving position of the centroid based on the energy function value of the snow sports track and the vector point in the iteration process to determine the snow information density splitting weight of the vector point includes:

对于任意一次迭代过程前的任一矢量点,基于任一矢量点及其所有关联采样点确定任一矢量点的传统矢量点分裂系数;所述关联采样点为任一矢量点所在区域内的采样点;For any vector point before any iteration process, a traditional vector point splitting coefficient of any vector point is determined based on any vector point and all its associated sampling points; the associated sampling points are sampling points within the region where any vector point is located;

对于任一矢量点在所述迭代过程中的各次移动,计算各次移动前任一矢量点的所有关联采样点的雪地运动赛道能量函数值的均值,记为任一矢量点各次移动前的矢量点能量值;将所有移动前的所述矢量点能量值组成任一矢量点的能量函数变化向量;For each movement of any vector point in the iterative process, the average of the energy function values of the snow sports track of all associated sampling points of any vector point before each movement is calculated, and recorded as the energy value of the vector point before each movement of any vector point; the energy values of the vector points before all the movements are combined into an energy function change vector of any vector point;

将所述能量函数变化向量的一阶导数向量的均值,记为任一矢量点的雪地运动赛道能量变化倾向;The mean value of the first-order derivative vector of the energy function change vector is recorded as the energy change tendency of the snow sports track at any vector point;

基于任一矢量点在所述迭代过程中的各次移动位置之间的欧式距离确定任一矢量点各次移动的质心移动位置权重;Determine the weight of the centroid movement position of each movement of any vector point based on the Euclidean distance between the movement positions of each vector point in the iterative process;

基于质心移动位置权重、雪地运动赛道能量变化倾向以及传统矢量点分裂系数,确定任一矢量点的雪地信息密度分裂权重。The snow information density splitting weight of any vector point is determined based on the weight of the center of mass movement position, the energy change tendency of the snow sports track and the traditional vector point splitting coefficient.

优选的,所述基于任一矢量点及其所有关联采样点确定任一矢量点的传统矢量点分裂系数,包括:Preferably, the method of determining a traditional vector point splitting coefficient of any vector point based on any vector point and all associated sampling points thereof includes:

将任一矢量点的所有关联采样点灰度值的标准差,记为任一矢量点的失真度;将任一矢量点的关联采样点数量与所述失真度的乘积作为任一矢量点的传统矢量点分裂系数。The standard deviation of the grayscale values of all associated sampling points of any vector point is recorded as the distortion of any vector point; the product of the number of associated sampling points of any vector point and the distortion is taken as the traditional vector point splitting coefficient of any vector point.

优选的,所述基于任一矢量点在所述迭代过程中的各次移动位置之间的欧式距离确定任一矢量点各次移动的质心移动位置权重,表达式为:Preferably, the weight of the centroid movement position of each movement of any vector point is determined based on the Euclidean distance between the movement positions of each vector point in the iterative process, and the expression is:

式中,是第h个矢量点第g次移动的质心移动位置权重,是第h个矢量点第g次移动前的位置,是第h个矢量点G次移动后的位置,之间的欧式距离,是第h个矢量点的所有移动次数的所述欧式距离的最大值。In the formula, is the weight of the center of mass movement of the hth vector point for the gth movement, is the position of the h-th vector point before the g-th movement, is the position of the hth vector point after G moves, yes and The Euclidean distance between It is the maximum value of the Euclidean distance of all movement times of the h-th vector point.

优选的,所述基于质心移动位置权重、雪地运动赛道能量变化倾向以及传统矢量点分裂系数,确定任一矢量点的雪地信息密度分裂权重,包括:Preferably, the snow information density splitting weight of any vector point is determined based on the mass center moving position weight, the energy change tendency of the snow sports track and the traditional vector point splitting coefficient, including:

计算任一矢量点各次移动前的矢量点能量值与任一矢量点各次移动的质心移动位置权重的乘积,计算任一矢量点所有移动次数下的所述乘积的和值;Calculate the product of the vector point energy value before each movement of any vector point and the weight of the center of mass movement position of each movement of any vector point, and calculate the sum of the products under all movement times of any vector point;

获取所有矢量点的雪地运动赛道能量变化倾向中的最大值,计算所述最大值与任一矢量点的雪地运动赛道能量变化倾向的和值;Obtaining a maximum value among the energy change tendencies of the snow sports track of all vector points, and calculating a sum of the maximum value and the energy change tendencies of the snow sports track of any vector point;

将两个和值与任一矢量点的传统矢量点分裂系数三项的乘积,作为任一矢量点的雪地信息密度分裂权重。The product of the two sums and the three traditional vector point splitting coefficients of any vector point is used as the snow information density splitting weight of any vector point.

优选的,所述基于雪地信息密度分裂权重优化VQ-LGB压缩算法,得到压缩后的视频帧,输入神经网络实现冰雪项目场地的高效通信,包括:Preferably, the VQ-LGB compression algorithm is optimized based on the snow information density split weight to obtain compressed video frames, which are input into the neural network to achieve efficient communication of ice and snow project venues, including:

将任一矢量点的雪地信息密度分裂权重代替传统矢量点分裂系数,改进VQ-LGB压缩算法获取压缩后的视频帧;The snow information density splitting weight of any vector point is used to replace the traditional vector point splitting coefficient, and the VQ-LGB compression algorithm is improved to obtain the compressed video frame.

将压缩后的视频帧在数据处理中心进行解压,得到解压后的视频帧;对解压后的视频帧人为设置标签值,标签值为1代表有人员摔倒,0代表没有人员摔倒;The compressed video frames are decompressed in the data processing center to obtain decompressed video frames; a label value is manually set for the decompressed video frames, where a label value of 1 represents that a person has fallen, and a label value of 0 represents that no person has fallen;

将所有解压后的视频帧以及对应标签值输入卷积网络,输出CNN卷积网络决策函数;Input all decompressed video frames and corresponding label values into the convolutional network and output the CNN convolutional network decision function;

将实时传输的解压后的视频帧作为CNN卷积网络决策函数的输入,输出标签值;当输出标签值为1,发出警报;否则不发出警报。The decompressed video frames transmitted in real time are used as the input of the CNN convolutional network decision function, and the label value is output; when the output label value is 1, an alarm is issued; otherwise, no alarm is issued.

本申请至少具有如下有益效果:This application has at least the following beneficial effects:

本申请通过对视频帧的暗通道、亮通道以及三原色通道进行灰度值计算得到雪地色彩突出分割权重,用于准确区分雪地背景和穿着色彩艳丽的运动人员;进一步通过稀疏光流向量与采样点之间形成的夹角获得方向吻合度,通过视频帧某一地区的光流向量的长度标准差,提高背景物品震荡的光流向量与运动员运动造成的光流向量的辨识度,并结合雪地色彩突出分割权重计算出雪地运动赛道能量函数值,精准定位视频帧中运动员运动的位置;进一步根据雪地运动赛道能量函数值以及VQ-LGB算法中矢量点的质心移动操作,获取矢量点能量值和雪地运动赛道能量变化倾向,表征出矢量点附近是否是雪地监控图中的运动员位置以及矢量点过去的搜索路径是否是向雪地监控图中的运动员位置进行搜索,挖掘历史移动数据中的价值信息;进一步,结合质心操作的移动情况、矢量点能量值以及雪地运动赛道能量变化倾向,计算出矢量点的雪地信息密度分裂权重,使VQ-LGB算法中分裂矢量点时倾向于选择视频帧中的运动员运动的位置,使VQ-LGB算法进行视频帧压缩时在视频帧中的运动员位置生成更多的矢量点,而在其它位置生成较少的矢量点,比传统算法更快得到低失真度的压缩结果,能够获得更高质量的冰雪项目场地的监控压缩数据的同时,加快压缩速度,提高通信质量和通信效率。The present application calculates the grayscale values of the dark channel, bright channel and three primary color channels of the video frame to obtain the snow color highlight segmentation weight, which is used to accurately distinguish the snow background from the athletes wearing colorful clothes; further, the direction consistency is obtained through the angle formed between the sparse optical flow vector and the sampling point, and the length standard deviation of the optical flow vector in a certain area of the video frame is used to improve the recognition of the optical flow vector caused by the oscillation of background objects and the optical flow vector caused by the athlete's movement, and the snow color highlight segmentation weight is combined to calculate the snow sports track energy function value, and accurately locate the athlete's movement in the video frame; further, according to the snow sports track energy function value and the center of mass movement operation of the vector point in the VQ-LGB algorithm, the vector point energy value and the snow sports track energy change tendency are obtained to indicate whether the vector point is near The athlete positions in the snow monitoring map and whether the past search paths of the vector points are directed towards the athlete positions in the snow monitoring map are used to mine valuable information in historical mobile data; further, the movement of the centroid operation, the energy value of the vector points and the energy change tendency of the snow sports track are combined to calculate the snow information density splitting weight of the vector points, so that when splitting the vector points in the VQ-LGB algorithm, the athlete movement positions in the video frame are tended to be selected, so that when the VQ-LGB algorithm compresses the video frame, more vector points are generated at the athlete positions in the video frame, and fewer vector points are generated at other positions. The compression results with low distortion are obtained faster than traditional algorithms, and higher-quality monitoring compression data of ice and snow venues can be obtained while accelerating the compression speed, improving communication quality and efficiency.

附图说明BRIEF DESCRIPTION OF THE DRAWINGS

为了更清楚地说明本申请实施例或现有技术中的技术方案和优点,下面将对实施例或现有技术描述中所需要使用的附图作简单的介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它附图。In order to more clearly illustrate the technical solutions and advantages in the embodiments of the present application or the prior art, the drawings required for use in the embodiments or the description of the prior art will be briefly introduced below. Obviously, the drawings described below are only some embodiments of the present application. For ordinary technicians in this field, other drawings can be obtained based on these drawings without paying any creative work.

图1为本申请提供的基于深度学习的冰雪项目场地监控图像高效通信方法的流程图;FIG1 is a flow chart of an efficient communication method for monitoring images of ice and snow sports venues based on deep learning provided by the present application;

图2为光流向量方向吻合度示意图;FIG2 is a schematic diagram of the degree of consistency of the direction of the optical flow vector;

图3为视频帧压缩过程的指标构建流程图。FIG3 is a flowchart of indicator construction for the video frame compression process.

具体实施方式Detailed ways

为了更进一步阐述本申请为达成预定发明目的所采取的技术手段及功效,以下结合附图及较佳实施例,对依据本申请提出的基于深度学习的冰雪项目场地监控图像高效通信方法,其具体实施方式、结构、特征及其功效,详细说明如下。在下述说明中,不同的“一个实施例”或“另一个实施例”指的不一定是同一实施例。此外,一或多个实施例中的特定特征、结构或特点可由任何合适形式组合。In order to further explain the technical means and effects adopted by this application to achieve the predetermined invention purpose, the specific implementation method, structure, features and effects of the efficient communication method of ice and snow project site monitoring images based on deep learning proposed in this application are described in detail below in combination with the accompanying drawings and preferred embodiments. In the following description, different "one embodiment" or "another embodiment" does not necessarily refer to the same embodiment. In addition, specific features, structures or characteristics in one or more embodiments may be combined in any suitable form.

除非另有定义,本文所使用的所有的技术和科学术语与属于本申请的技术领域的技术人员通常理解的含义相同。Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.

下面结合附图具体的说明本申请所提供的基于深度学习的冰雪项目场地监控图像高效通信方法的具体方案。The specific scheme of the efficient communication method for monitoring images of ice and snow project sites based on deep learning provided by this application is described in detail below with reference to the accompanying drawings.

本申请一个实施例提供的基于深度学习的冰雪项目场地监控图像高效通信方法。An embodiment of the present application provides an efficient communication method for monitoring images of ice and snow project sites based on deep learning.

具体的,提供了如下的基于深度学习的冰雪项目场地监控图像高效通信方法,请参阅图1,该方法包括以下步骤:Specifically, the following deep learning-based efficient communication method for monitoring images of ice and snow sports venues is provided. Please refer to FIG1. The method includes the following steps:

步骤S001,采集冰雪项目场地监控视频的各视频帧。Step S001, collecting video frames of the monitoring video of the ice and snow sports venue.

实时采集冰雪项目场地的视频,其中将某个监控地区视频的第n帧记为雪地监控视频帧,视频帧的色彩空间为RGB色彩空间,进一步将雪地监控视频帧为输入,采用RGB到灰度色彩空间转化算法,输出为视频帧。色彩空间转化算法为本领域公知技术,本实施例不再赘述。Collect videos of ice and snow venues in real time, where the nth frame of a video in a certain monitoring area is recorded as the snow monitoring video frame , the color space of the video frame is RGB color space, and the snow monitoring video frame is further As input, the RGB to grayscale color space conversion algorithm is used, and the output is a video frame The color space conversion algorithm is a well-known technology in the art and will not be described in detail in this embodiment.

至此,可以得到冰雪项目场地监控视频的各视频帧。At this point, each video frame of the monitoring video of the ice and snow sports venue can be obtained.

步骤S002,对视频帧进行分析,优化VQ-RGB算法中各迭代次数下任一矢量点的传统矢量点分裂系数。Step S002, analyzing the video frame, and optimizing the traditional vector point splitting coefficient of any vector point at each iteration number in the VQ-RGB algorithm.

传统的VQ-RGB算法对视频帧进行区域分割时,对同一区域的采样点用相同的矢量类型进行表示,以到达压缩视频的目的。但是在雪地监控视频中,大部分视频帧是不包含人员摔倒监控的雪地画面,在传统算法对其进行压缩时,其中的RGB算法部分对区域分割迭代运算的过程中,对雪地消耗了大量的迭代计算资源,进而导致算法压缩速度降低,降低了雪地监控的通信速度。When the traditional VQ-RGB algorithm performs region segmentation on video frames, the sampling points in the same region are represented by the same vector type to achieve the purpose of video compression. However, in snow monitoring videos, most video frames do not contain snow scenes where people fall down. When the traditional algorithm compresses them, the RGB algorithm consumes a lot of iterative computing resources on the snow during the iterative operation of region segmentation, which reduces the algorithm compression speed and the communication speed of snow monitoring.

因此,本实施例首先以第n帧视频帧为输入,分别采用暗通道算法和亮通道算法进行计算,输出分别为暗通道视频帧和亮通道视频帧。暗通道算法和亮通道算法为本领域公知技术不再赘述。Therefore, this embodiment first takes the nth video frame as input, and uses the dark channel algorithm and the bright channel algorithm for calculation, and outputs the dark channel video frame and the bright channel video frame respectively. The dark channel algorithm and the bright channel algorithm are well-known technologies in the art and will not be described in detail.

计算第n帧视频帧中各采样点的雪地色彩突出分割权重:Calculate the snow color highlight segmentation weights for each sampling point in the nth video frame:

式中,是第m个采样点的雪地色彩突出分割权重,分别是第m个采样点在暗通道和亮通道视频帧中的强度值,分别是第m个采样点在R、G、B通道中的强度值。In the formula, is the snow color highlight segmentation weight of the mth sampling point, , are the intensity values of the mth sampling point in the dark channel and bright channel video frames, respectively. , , are the intensity values of the mth sampling point in the R, G, and B channels respectively.

式中,在雪地中的背景多是白色的雪地部分,该部分的采样点其暗通道强度值和亮通道强度值差异较小,因此将第m采样点的暗亮两通道的差值作为权重,差值越大的部分越不可能是背景雪地;式中将第m个采样点的R、G、B通道两两求差值再求和,其值越大代表第m个采样点在RGB通道中的强度值差异越大,代表第m个采样点的色彩越鲜艳,同时除以第m个采样点暗通道强度值,暗通道强度值越小代表第m个采样点的色彩越明亮,在冰雪项目场地中,为保证安全运动人员多穿着鲜艳明亮的衣物,因此色彩越艳丽越不可能是背景雪地。In the formula, the background in the snow is mostly the white snow part. The difference between the dark channel intensity value and the bright channel intensity value of the sampling point in this part is small. Therefore, the difference between the dark and bright channels of the m-th sampling point is used as the weight. The larger the difference, the less likely it is the background snow. In the formula, the R, G, and B channels of the m-th sampling point are subtracted and then summed. The larger the value, the greater the difference in the intensity value of the m-th sampling point in the RGB channel, and the brighter the color of the m-th sampling point. At the same time, it is divided by the dark channel intensity value of the m-th sampling point. The smaller the dark channel intensity value, the brighter the color of the m-th sampling point. In the ice and snow venues, in order to ensure safety, athletes often wear bright and bright clothes. Therefore, the brighter the color, the less likely it is the background snow.

最终获得雪地色彩突出分割权重,其值越大代表对应的采样点越是运动员而非背景,在进行视频帧压缩时该部分的细节越应该得到保留。Finally, the snow color highlight segmentation weight is obtained. The larger the value, the more the corresponding sampling point is an athlete rather than a background, and the more details of this part should be retained when compressing the video frame.

进一步,由于冰雪场地的赛道的滑雪路线通常较为固定,因此可以根据视频帧中不同区域的运动状态变化选择出常有运动人员经过的区域。Furthermore, since the skiing routes of the tracks in the ice and snow venues are usually relatively fixed, the areas where athletes often pass by can be selected according to the changes in the motion states of different areas in the video frames.

首先以第n帧和第n-1帧的视频帧为输入,采用稀疏光流算法,输出为两帧视频帧的稀疏光流向量集合,集合中有多个光流向量,其中第k个光流向量的起点和终点坐标分别记为,代表第n-1帧的位置的采样点在第n帧中运动到了位置的采样点。稀疏光流算法为本领域公知技术不再赘述。First, the nth and n-1th video frames are used as input, and the sparse optical flow algorithm is used to output the sparse optical flow vector set of the two video frames. There are multiple optical flow vectors in the set, and the starting and ending coordinates of the kth optical flow vector are recorded as , , representing the n-1th frame The sampling point of the position moves to The sparse optical flow algorithm is a well-known technology in the art and will not be described in detail.

进一步将第r帧视频帧中第m个采样点到第k个光流向量中点的连线与第k个光流向量的所构成的夹角记为方向吻合度The angle between the line from the mth sampling point to the midpoint of the kth optical flow vector in the rth video frame and the kth optical flow vector is recorded as the direction fit .

如图2所示,图2中的箭头为光流向量,其中,m表示当前帧中的第m个采样点,k表示前一帧中指向采样点m的第k个光流向量,Ia表示采样点m和第k个光流向量的夹角,Ia越小代表第m个采样点越在第k个光流向量的运动路径上,则在对第m个采样点的运动情况进行评估时,第k个光流向量对其影响越大。As shown in Figure 2, the arrows in Figure 2 are optical flow vectors, where m represents the mth sampling point in the current frame, k represents the kth optical flow vector pointing to the sampling point m in the previous frame, and Ia represents the angle between the sampling point m and the kth optical flow vector. The smaller Ia is, the closer the mth sampling point is to the motion path of the kth optical flow vector, and when evaluating the motion of the mth sampling point, the greater the influence of the kth optical flow vector on it.

同时对第r帧视频帧中的第k个光流向量,获取其中点位置,同时获取第r-1帧视频帧中的光流向量中点位置与欧式距离最近的光流向量,将其作为第r帧视频帧中的第k个光流向量的关联光流向量,以此类推在第r-2帧中也能找到对应的关联光流向量。At the same time, for the kth optical flow vector in the rth video frame, get the midpoint position , and at the same time obtain the midpoint position of the optical flow vector in the r-1th frame of the video frame and The optical flow vector with the closest Euclidean distance is used as the associated optical flow vector of the kth optical flow vector in the rth video frame, and so on, the corresponding associated optical flow vector can also be found in the r-2th frame.

进一步对第r帧视频帧中的第k个光流向量获取N个关联光流向量,计算这些光流向量的长度的标准差记为第r帧视频帧中的第k个光流向量的光流向量非震荡度,其值越小代表第k个光流向量越可能是视频帧中来回震荡晃动的物体造成的,如随风的彩旗,雪地中的树;而对于冰雪项目场地中的运动员,其在赛道中的运动并非来回晃动,因此其光流向量非震荡度值较高。Further, N associated optical flow vectors are obtained for the kth optical flow vector in the rth video frame, and the standard deviation of the lengths of these optical flow vectors is calculated and recorded as the optical flow vector non-oscillation degree of the kth optical flow vector in the rth video frame. The smaller the value, the more likely the kth optical flow vector is to be caused by objects that oscillate back and forth in the video frame, such as colorful flags in the wind and trees in the snow. For athletes in ice and snow events, their movements on the track are not swaying back and forth, so their optical flow vector non-oscillation value is higher.

进一步,计算视频帧中各采样点的雪地运动赛道能量函数值:Furthermore, the energy function value of the snow sports track at each sampling point in the video frame is calculated:

式中,是第n帧视频帧的第m个采样点的雪地运动赛道能量函数值,N是参考帧数量,本实施例取值N=50,是第r帧参考帧中的光流向量数量,是第r帧参考帧第k个光流向量的长度,是第r帧参考帧第m个采样点的雪地色彩突出分割权重,是第r帧参考帧中第m个采样点与第k个光流向量的方向吻合度,是第r帧参考帧中第k个光流向量中点到第m个采样点的欧式距离,是第r帧参考帧中第k个光流向量的光流向量非震荡度,为调参因子,防止分母为0,取值为0.1。In the formula, is the snow sports track energy function value of the mth sampling point of the nth video frame, N is the number of reference frames, and in this embodiment, the value N=50, is the number of optical flow vectors in the r-th reference frame, is the length of the kth optical flow vector of the rth reference frame, is the snow color highlight segmentation weight of the mth sampling point in the rth reference frame, is the degree of consistency between the direction of the mth sampling point and the kth optical flow vector in the rth reference frame, is the Euclidean distance from the midpoint of the kth optical flow vector in the rth reference frame to the mth sampling point, is the non-oscillation degree of the optical flow vector of the kth optical flow vector in the rth reference frame, It is a parameter adjustment factor to prevent the denominator from being 0, and its value is 0.1.

式中,稀疏光流向量的长度越长,代表其附近采样点的移动越距离,代表其越有可能是冰雪项目运动员划过的赛道;式中以欧式距离和方向吻合度为权重,以光流向量的长度为其对采样点的影响值加权相加,值越大代表第m个采样点附近的运动越距离,对应的其雪地运动赛道能量函数值越大;式中第m个采样点的雪地色彩突出分割权重越大,其光流向量非震荡度越小,代表第m个采样点越可能是正在运动的运动员,对应的其雪地运动赛道能量函数值越大。In the formula, the longer the length of the sparse optical flow vector is, the farther the movement of the nearby sampling points is, and the more likely it is that it is a track crossed by athletes of ice and snow events; in the formula, the Euclidean distance and the degree of direction fit are used as weights, and the length of the optical flow vector is used to weight the influence value of the sampling point. The larger the value is, the farther the movement is near the m-th sampling point, and the larger the corresponding energy function value of the snow sports track is; in the formula, the larger the snow color highlight segmentation weight of the m-th sampling point is, the smaller the non-oscillation degree of its optical flow vector is, which means that the m-th sampling point is more likely to be an athlete in motion, and the corresponding energy function value of the snow sports track is larger.

最终得到雪地运动赛道能量函数值,值越大代表第m个采样点越是冰雪场地的赛道部分,对应在VQ-LGB算法中对其该部分采样点应该生成更多的矢量点,以保留该部分的细节信息,而对于其它部分生成较少的矢量点,有助于提高监控视频的通信质量和通信速率。Finally, the energy function value of the snow sports track is obtained. The larger the value, the more the m-th sampling point is on the track of the ice and snow venue. Correspondingly, more vector points should be generated for the sampling points of this part in the VQ-LGB algorithm to retain the detailed information of this part, while fewer vector points are generated for other parts, which helps to improve the communication quality and communication rate of the monitoring video.

进一步,对VQ-LGB算法中的LGB算法进行改进,分别计算矢量点的雪地信息密度分裂权重,具体如下:Furthermore, the LGB algorithm in the VQ-LGB algorithm is improved to calculate the snow information density splitting weights of the vector points respectively, as follows:

LGB算法是VQ-LGB压缩算法中的优化迭代算法,在迭代开始时会在视频帧中随机生成多个矢量点,根据采样点到矢量点的欧式距离,将距离最近的矢量点作为采样点的归属矢量点,每个矢量点有多个关联采样点。The LGB algorithm is an optimized iterative algorithm in the VQ-LGB compression algorithm. At the beginning of the iteration, multiple vector points are randomly generated in the video frame. According to the Euclidean distance from the sampling point to the vector point, the vector point with the closest distance is used as the belonging vector point of the sampling point. Each vector point has multiple associated sampling points.

而传统算法在进行迭代时,会计算第h个矢量点的所有关联采样点的质心来移动第h个矢量点,直到所有矢量点都处在其关联采样点的质心位置,此时对第h个矢量点的所有关联采样点的灰度值计算标准差,作为第h个矢量点的失真度,失真度过大时认为矢量点对视频帧的分割效果不细致,此时选择失真度大或者归属采样点多的矢量点进行分裂操作。当某一次迭代操作时,记录第h个矢量点的失真度与其关联采样点数量之积记为第h个矢量点的传统矢量点分裂系数When the traditional algorithm is iterating, it will calculate the centroid of all the associated sampling points of the h-th vector point to move the h-th vector point until all the vector points are at the centroid of their associated sampling points. At this time, the standard deviation of the grayscale values of all the associated sampling points of the h-th vector point is calculated as the distortion of the h-th vector point. If the distortion is too large, it is considered that the vector point does not segment the video frame in detail. At this time, the vector point with a large distortion or a large number of associated sampling points is selected for splitting. During a certain iteration, the product of the distortion of the h-th vector point and the number of its associated sampling points is recorded as the traditional vector point splitting coefficient of the h-th vector point. .

进一步,在从上一次分裂矢量点后到本次分裂矢量点前,第h个矢量点有多次寻找质心的移动,假设从上一次分裂矢量点后到本次分裂矢量点前第h个矢量点经过了G次寻找质心的移动操作,则在第g次移动前计算此时所有关联采样点的雪地运动赛道能量函数值均值,记为矢量点能量值,进而组成第h个矢量点在该次迭代的移动过程中的能量函数变化向量,对该向量计算一阶导数向量,并对一阶导数向量求均值记为雪地运动赛道能量变化倾向,该值正负代表第h个矢量点的在搜索时是向视频帧的高信息区域靠近还是远离,其绝对值大小代表了靠近还是远离的强度。同时将第g次移动前的位置记为Furthermore, from the last split vector point to the current split vector point, the hth vector point has multiple movements to find the center of mass. Assuming that from the last split vector point to the current split vector point, the hth vector point has undergone G movements to find the center of mass, then before the gth movement, the mean value of the energy function of the snow sports track of all associated sampling points is calculated, which is recorded as the vector point energy value , and then form the energy function change vector of the hth vector point in the movement process of this iteration , calculate the first-order derivative vector of the vector, and average the first-order derivative vector as the energy change tendency of the snow sports track , the positive or negative value represents whether the hth vector point is approaching or moving away from the high information area of the video frame during the search, and its absolute value represents the intensity of approaching or moving away. At the same time, the position before the gth movement is recorded as .

计算矢量点的雪地信息密度分裂权重:Calculate the snow information density splitting weight of the vector point:

式中,是第h个矢量点的雪地信息密度分裂权重,第h个矢量点的传统矢量点分裂系数,是第h个矢量点的雪地运动赛道能量变化倾向,是H个矢量点中雪地运动赛道能量变化倾向的最大值,是第h个矢量点在该次迭代寻找质心移动的总次数,是第h个矢量点第g次移动前的矢量点能量值,是第h个矢量点第g次移动的质心移动位置权重;In the formula, is the snow information density splitting weight of the h-th vector point, The traditional vector point splitting coefficient of the h-th vector point, is the energy change tendency of the snow sports track at the hth vector point, is the maximum value of the energy change tendency of the snow sports track among H vector points, is the total number of times the h-th vector point moves in this iteration to find the center of mass, is the energy value of the h-th vector point before the g-th movement, is the weight of the center of mass movement of the hth vector point for the gth movement;

是第h个矢量点第g次移动前的位置,是第h个矢量点G次移动后的位置,之间的欧式距离,是第h个矢量点的所有移动次数的所述欧式距离的最大值。 is the position of the h-th vector point before the g-th movement, is the position of the hth vector point after G moves, yes and The Euclidean distance between It is the maximum value of the Euclidean distance of all movement times of the h-th vector point.

式中,传统矢量点分裂系数越大,第h个矢量点的失真度越大且关联采样点越多,分裂出新矢量点的概率越大;式中先将雪地运动赛道能量变化倾向进行非负化操作,非负化后其值越大代表第h个矢量点寻找质心的平移操作时越是在靠近视频帧中的运动员位置,越是本场景中所需要保留更多信息的位置,对应的第h矢量点分裂出新矢量点的概率越大。In the formula, the larger the traditional vector point splitting coefficient is, the greater the distortion of the h-th vector point is and the more associated sampling points are, the greater the probability of splitting a new vector point is; in the formula, the energy change tendency of the snow sports track is first non-negatively operated. After non-negative operation, the larger the value is, the closer the h-th vector point is to the athlete's position in the video frame when searching for the center of mass translation operation, and the more information needs to be retained in this scene. The corresponding h-th vector point has a greater probability of splitting a new vector point.

式中,以每次计算矢量点能量值时到目前矢量点位置的欧式距离反向归一化作为质心移动位置权重,历史质心移动搜索的位置距离当前矢量点越近权重越大,然后将矢量点能量值加权相加,值越大代表第h个矢量点目前所处位置越是视频帧中的运动员位置,对应的第h矢量点分裂出新矢量点的概率越大。In the formula, the Euclidean distance from the current vector point position to the energy value of each vector point is reversely normalized as the weight of the center of mass movement position. The closer the historical center of mass movement search position is to the current vector point, the greater the weight. Then the vector point energy values are weighted and added. The larger the value, the closer the current position of the h-th vector point is to the player position in the video frame, and the greater the probability of the corresponding h-th vector point splitting into a new vector point.

进一步,以视频帧为输入,采用VQ-LGB压缩算法进行计算,在计算过程中,将第h个矢量点的传统矢量点分裂系数替换为第h个矢量点的雪地信息密度分裂权重,输出为压缩后的视频帧数据。其中,视频帧压缩过程的指标构建流程图如图3所示。Further, the video frame As input, the VQ-LGB compression algorithm is used for calculation. During the calculation process, the traditional vector point splitting coefficient of the hth vector point is Replaced with the snow information density splitting weight of the h-th vector point, and output as a compressed video frame Data. Among them, the indicator construction flow chart of the video frame compression process is shown in Figure 3.

本实施例所述改进方法,相比与传统的VQ-LGB压缩算法,其在视频帧中的运动员位置生成更多的矢量点,而在其它位置生成较少的矢量点,而且能比传统算法更快的得到低失真度的压缩结果,能够获得更高质量的冰雪项目场地的监控视频压缩数据的同时,加快压缩速度,提高通信质量和通信效率。Compared with the traditional VQ-LGB compression algorithm, the improved method described in this embodiment generates more vector points at the athlete positions in the video frame and fewer vector points at other positions, and can obtain low-distortion compression results faster than traditional algorithms. It can obtain higher-quality monitoring video compression data of ice and snow venues while accelerating the compression speed and improving communication quality and efficiency.

进一步将压缩后的视频帧数据传输到冰雪项目的数据处理中心。Further compress the video frame The data is transmitted to the data processing center of the ice and snow project.

步骤S003,基于压缩后的视频帧,使用深度学习算法进行处理,实现冰雪项目场地的高效通信。Step S003, based on the compressed video frames, use a deep learning algorithm to process and achieve efficient communication at the ice and snow venues.

冰雪项目的数据处理中心以压缩后的视频帧数据为输入,采用VQ-LGB压缩算法的解压算法进行解压,输出为解压后的视频帧The data processing center of the ice and snow project uses compressed video frames The data is input, decompressed using the decompression algorithm of the VQ-LGB compression algorithm, and the output is the decompressed video frame .

进一步,获取数据集,数据集为多张解压后的视频帧,其中每帧的标签值设置为1或0,表征解压后的视频帧中是否有人员摔倒,标签值为1代表有人员摔倒,0代表没有人员摔倒。Further, obtain the data set, which is multiple decompressed video frames , where the label value of each frame is set to 1 or 0, indicating whether there is a person falling in the decompressed video frame. The label value of 1 represents that there is a person falling, and 0 represents that there is no person falling.

将数据集作为CNN卷积网络的输入,输出为CNN卷积网络决策函数,其中CNN卷积网络模型具体构建方法为公知技术,不再赘述。The data set is used as the input of the CNN convolutional network, and the output is the CNN convolutional network decision function, wherein the specific construction method of the CNN convolutional network model is a well-known technology and will not be repeated here.

实时传输的解压后的视频帧作为CNN卷积网络决策函数的输入,输出为数值1或0,当输出数值为1时,代表监控冰雪项目场地中有人摔倒,发出警报。Decompressed video frames for real-time transmission As the input of the CNN convolutional network decision function, the output is a value of 1 or 0. When the output value is 1, it means that someone has fallen in the monitored ice and snow venue and an alarm is issued.

至此,实现冰雪项目场地的视频帧的高效通信。At this point, efficient communication of video frames at ice and snow sports venues is achieved.

需要说明的是:上述本申请实施例先后顺序仅仅为了描述,不代表实施例的优劣。且上述对本说明书特定实施例进行了描述。另外,在附图中描绘的过程不一定要求示出的特定顺序或者连续顺序才能实现期望的结果。在某些实施方式中,多任务处理和并行处理也是可以的或者可能是有利的。It should be noted that the above sequence of the embodiments of the present application is for description only and does not represent the advantages and disadvantages of the embodiments. The above is a description of a specific embodiment of this specification. In addition, the processes depicted in the accompanying drawings do not necessarily require the specific order or continuous order shown to achieve the desired results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.

本说明书中的各个实施例均采用递进的方式描述,各个实施例之间相同或相似的部分互相参见即可,每个实施例重点说明的都是与其他实施例的不同之处。The various embodiments in this specification are described in a progressive manner, and the same or similar parts between the various embodiments can be referenced to each other, and each embodiment focuses on the differences from other embodiments.

以上所述实施例仅用以说明本申请的技术方案,而非对其限制;对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的范围,均应包含在本申请的保护范围之内。The embodiments described above are only used to illustrate the technical solutions of the present application, rather than to limit them. Modifications to the technical solutions recorded in the aforementioned embodiments, or equivalent replacement of some of the technical features therein, do not cause the essence of the corresponding technical solutions to deviate from the scope of the technical solutions of the embodiments of the present application, and should all be included in the protection scope of the present application.

Claims (4)

1. The high-efficiency communication method for the ice and snow project site monitoring image based on deep learning is characterized by comprising the following steps of:
Collecting each video frame in the monitoring video of the ice and snow project site;
Determining the snow color highlighting and dividing weight of each sampling point based on the intensity values of each sampling point of the video frame under different channels;
acquiring optical flow vectors of adjacent video frames by adopting a sparse optical flow algorithm; determining the directional fitness of each sampling point and any optical flow vector based on the position distribution of each sampling point and any optical flow vector in the video frame;
Determining an optical flow vector non-concussion degree of any optical flow vector based on a distance distribution of the optical flow vectors in adjacent video frames; determining a snowfield motion track energy function value of each sampling point of a video frame based on the snowfield color salient segmentation weight, the optical flow vector non-concussion degree and the direction fitness of the continuous reference frame at the same sampling point;
Randomly generating vector points in a video frame, adopting an LGB algorithm in a VQ-LGB algorithm to find the moving position of the centroid in the iteration process of the vector points based on the snowfield motion race track energy function value in any iteration process of the video frame, and determining the snowfield information density splitting weight of the vector points;
Optimizing a VQ-LGB compression algorithm based on the snow information density splitting weight to obtain a compressed video frame, and inputting the compressed video frame into a neural network to realize efficient communication of ice and snow project sites;
The determining the snowfield motion race track energy function value of each sampling point of the video frame based on the snowfield color salient segmentation weight, the optical flow vector non-concussion degree and the direction fitness degree of the continuous reference frame at the same sampling point comprises the following steps:
Recording the preset number of video frames adjacent to any video frame as reference frames of any video frame;
For any optical flow vector in the reference frame, calculating the product of the length of any optical flow vector and the snow color salient division weight of each sampling point in the reference frame as a molecule;
Calculating three products of the non-concussion degree of the optical flow vector of any optical flow vector, the directional fitness of each sampling point in the reference frame and any optical flow vector, and the Euclidean distance between each sampling point in the reference frame and any optical flow vector, and taking the sum of the three products and a preset parameter adjusting factor as a denominator;
Calculating the ratio of the numerator to the denominator; taking the sum of the ratio values of all the optical flow vectors in all the reference frames as a snow movement track energy function value of each sampling point of any video frame;
the method for determining the snow information density splitting weight of the vector point based on the snow motion track energy function value and the moving position of the centroid of the vector point in the iterative process comprises the following steps:
for any vector point before any iteration process, determining a traditional vector point splitting coefficient of any vector point based on any vector point and all relevant sampling points thereof; the associated sampling points are sampling points in the area where any vector point is located;
For each movement of any vector point in the iterative process, calculating the average value of the snow motion track energy function values of all the associated sampling points of any vector point before each movement, and recording the average value as the vector point energy value before each movement of any vector point; forming energy function change vectors of any vector point by all the vector point energy values before movement;
The mean value of the first-order reciprocal vector of the energy function change vector is recorded as the energy change trend of the snow sports track at any vector point;
Determining centroid movement position weights of each movement of any vector point based on Euclidean distances between each movement position of any vector point in the iterative process;
determining the snow information density splitting weight of any vector point based on the centroid moving position weight, the snow movement track energy variation tendency and the traditional vector point splitting coefficient;
the determining the traditional vector point splitting coefficient of any vector point based on any vector point and all the associated sampling points comprises the following steps:
the standard deviation of gray values of all relevant sampling points of any vector point is recorded as the distortion degree of any vector point; taking the product of the number of the associated sampling points of any vector point and the distortion degree as a traditional vector point splitting coefficient of any vector point;
The centroid movement position weight of each movement of any vector point is determined based on the Euclidean distance between each movement position of any vector point in the iterative process, and the expression is as follows:
In the method, in the process of the invention, Is the centroid movement position weight of the g-th movement of the h-th vector point,Is the position before the g-th movement of the h-th vector point,Is the position of the h vector point after the G-th shift,Is thatAnd (3) withThe euclidean distance between the two,Is the maximum value of the Euclidean distance of all the moving times of the h vector point;
The determining the snow information density splitting weight of any vector point based on the centroid moving position weight, the snow movement track energy change trend and the traditional vector point splitting coefficient comprises the following steps:
calculating the product of the vector point energy value before each movement of any vector point and the centroid movement position weight of each movement of any vector point, and calculating the sum of the products under all movement times of any vector point;
Obtaining the maximum value in the snow movement track energy change trend of all vector points, and calculating the sum value of the maximum value and the snow movement track energy change trend of any vector point;
Taking the product of the two sum values and the traditional vector point splitting coefficient of any vector point as the snow information density splitting weight of any vector point;
The method for optimizing the VQ-LGB compression algorithm based on the snow information density splitting weight to obtain a compressed video frame, inputting the compressed video frame into a neural network to realize the efficient communication of the ice and snow project site comprises the following steps:
Replacing the traditional vector point splitting coefficient with the snowfield information density splitting weight of any vector point, and improving the VQ-LGB compression algorithm to obtain a compressed video frame;
Decompressing the compressed video frame in a data processing center to obtain a decompressed video frame; manually setting a label value for the decompressed video frame, wherein the label value is 1, which indicates that a person falls down, and 0 indicates that no person falls down;
inputting all decompressed video frames and corresponding tag values into a convolutional network, and outputting a CNN convolutional network decision function;
Taking the decompressed video frames transmitted in real time as the input of a CNN convolutional network decision function, and outputting a tag value; when the output tag value is 1, an alarm is sent out; otherwise, no alarm is issued.
2. The method for efficiently communicating the monitoring image of the ice and snow project field based on the deep learning according to claim 1, wherein the determining the snow color highlighting and dividing weight of each sampling point based on the intensity values of each sampling point of the video frame under different channels is expressed as follows:
In the method, in the process of the invention, The snow color highlighting segmentation weight for the mth sample point,The m-th sample point is the intensity value in the dark channel and bright channel video frames,The intensity values of the mth sample point in the R, G, B channels are respectively.
3. The method for efficient communication of ice and snow project site monitoring images based on deep learning as claimed in claim 1, wherein determining the directional fitness of each sampling point and any optical flow vector based on the position distribution of each sampling point and any optical flow vector in the video frame comprises: connecting each sampling point with the midpoint of any optical flow vector, and taking the included angle between the connecting line and any optical flow vector as the direction coincidence degree of each sampling point and any optical flow vector.
4. The method for efficient communication of ice and snow project site monitor images based on deep learning as set forth in claim 1, wherein determining the optical flow vector non-concussion degree of any one of the optical flow vectors based on the distance distribution of the optical flow vectors in the adjacent video frames comprises:
for any optical flow vector in each video frame, marking the optical flow vector of the midpoint closest to the midpoint of any optical flow vector in the video frame of the adjacent previous frame as the associated optical flow vector of any optical flow vector;
the standard deviation of the lengths of all associated optical flow vectors of any optical flow vector is taken as the optical flow vector non-concussion of any optical flow vector.
CN202410458073.9A 2024-04-17 2024-04-17 Efficient communication method of monitoring images for ice and snow sports venues based on deep learning Active CN118101970B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410458073.9A CN118101970B (en) 2024-04-17 2024-04-17 Efficient communication method of monitoring images for ice and snow sports venues based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410458073.9A CN118101970B (en) 2024-04-17 2024-04-17 Efficient communication method of monitoring images for ice and snow sports venues based on deep learning

Publications (2)

Publication Number Publication Date
CN118101970A CN118101970A (en) 2024-05-28
CN118101970B true CN118101970B (en) 2024-07-26

Family

ID=91165481

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410458073.9A Active CN118101970B (en) 2024-04-17 2024-04-17 Efficient communication method of monitoring images for ice and snow sports venues based on deep learning

Country Status (1)

Country Link
CN (1) CN118101970B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119155447A (en) * 2024-11-18 2024-12-17 湖南云箭智能科技有限公司 Safety monitoring video optimization transmission method and video monitoring terminal

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103955945A (en) * 2014-05-23 2014-07-30 哈尔滨师范大学 Self-adaption color image segmentation method based on binocular parallax and movable outline
CN104966305A (en) * 2015-06-12 2015-10-07 上海交通大学 Foreground detection method based on motion vector division

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000354248A (en) * 1999-04-09 2000-12-19 Matsushita Electric Ind Co Ltd Image coder and its method for detecting motion vector
TWI433545B (en) * 2007-10-30 2014-04-01 Nippon Telegraph & Telephone Image encoding apparatuses and decoding apparatuses
CN103546759A (en) * 2013-10-29 2014-01-29 沈阳工业大学 An Image Compression Coding Method Based on the Combination of Wavelet Packet and Vector Quantization
WO2016187776A1 (en) * 2015-05-25 2016-12-01 北京大学深圳研究生院 Video frame interpolation method and system based on optical flow method
BR112021016896A2 (en) * 2019-03-08 2021-11-03 Huawei Tech Co Ltd Encoder, decoder and corresponding methods for inter-prediction
CN112866697B (en) * 2020-12-31 2022-04-05 杭州海康威视数字技术股份有限公司 Video image coding and decoding method and device, electronic equipment and storage medium
CN112949407B (en) * 2021-02-02 2022-06-14 武汉大学 Remote sensing image building vectorization method based on deep learning and point set optimization
CN117636268A (en) * 2023-12-04 2024-03-01 吉林大学 Unmanned aerial vehicle aerial natural driving data set construction method oriented to ice and snow environment

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103955945A (en) * 2014-05-23 2014-07-30 哈尔滨师范大学 Self-adaption color image segmentation method based on binocular parallax and movable outline
CN104966305A (en) * 2015-06-12 2015-10-07 上海交通大学 Foreground detection method based on motion vector division

Also Published As

Publication number Publication date
CN118101970A (en) 2024-05-28

Similar Documents

Publication Publication Date Title
CN108234870B (en) Image processing method, device, terminal and storage medium
CN118101970B (en) Efficient communication method of monitoring images for ice and snow sports venues based on deep learning
CN113963315B (en) A method and system for real-time video multi-person behavior recognition in complex scenes
JP4381310B2 (en) Media processing system
CN110781838A (en) Multi-modal trajectory prediction method for pedestrian in complex scene
CN107274433A (en) Method for tracking target, device and storage medium based on deep learning
CN107483887A (en) An early warning and detection method for emergencies in video surveillance of smart cities
JP2006508463A (en) Face detection
CN109635661B (en) A far-field wireless charging receiving target detection method based on convolutional neural network
CN116600132B (en) Coal mine video data self-adaptive compression method
CN116912949B (en) A gait recognition method based on perspective-aware part-wise attention mechanism
Ghasemzadeh et al. Deepsportlab: a unified framework for ball detection, player instance segmentation and pose estimation in team sports scenes
CN111274895A (en) CNN micro-expression identification method based on cavity convolution
CN105303200A (en) Human face identification method for handheld device
CN113033468A (en) Specific person re-identification method based on multi-source image information
CN114666457A (en) Video and audio program broadcasting guide method, device, equipment, system and medium
Ouyang et al. The comparison and analysis of extracting video key frame
CN113052136A (en) Pedestrian detection method based on improved Faster RCNN
CN114937239B (en) Pedestrian multi-target tracking identification method and tracking identification device
CN114372502B (en) Angle-adaptive elliptical template target detector
CN116682048A (en) Method and device for detecting violation of badminton serve height
CN117789253B (en) A video person re-identification method based on dual networks
CN115512263A (en) A dynamic visual monitoring method and device for falling objects
CN114782803A (en) Method for monitoring transmission line galloping based on compression sampling and image recognition
Peng et al. Arena: A Patch-of-Interest ViT Inference Acceleration System for Edge-Assisted Video Analytics

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant