CN104915000A - Multisensory biological recognition interaction method for naked eye 3D advertisement - Google Patents
Multisensory biological recognition interaction method for naked eye 3D advertisement Download PDFInfo
- Publication number
- CN104915000A CN104915000A CN201510282106.XA CN201510282106A CN104915000A CN 104915000 A CN104915000 A CN 104915000A CN 201510282106 A CN201510282106 A CN 201510282106A CN 104915000 A CN104915000 A CN 104915000A
- Authority
- CN
- China
- Prior art keywords
- crowd
- human
- feature
- classification
- advertisement
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Image Analysis (AREA)
Abstract
本发明涉及一种用于裸眼3D广告的多感知生物识别交互方法,其步骤为:⑴在场景中的人体检测:在自动投放针对性的广告内容平台的建设中,通过摄像头对固定场景中的运动人体进行监视与检测;⑵人群的特征检测:提取人群的人脸特征、发型特征和形态特征,用来区分不同人群的年龄和性别;⑶人群智能识别与分类:利用提取的人群的人脸特征和形态特征,对人群进行性别和年龄段的智能识别分类;⑷广告推荐投放决策:根据人群分类的结果,将不同的广告进行分类投放,采用个性化的推荐技术,将不同的信息传递给想要指定的人群;⑸智能人机交互:通过体感技术,语音识别技术将捕获的人体动作和语音信号转变为控制信号,来进行人与广告机的人机智能交互。
The present invention relates to a multi-sensory biometric interaction method for naked-eye 3D advertisements, the steps of which are: (1) human body detection in the scene: in the construction of a platform for automatically delivering targeted advertisement content, the camera monitors the human body in a fixed scene Monitoring and detection of moving human body; (2) crowd feature detection: extract the face features, hairstyle features and morphological features of the crowd to distinguish the age and gender of different crowds; (3) crowd intelligent recognition and classification: use the extracted crowd face Features and morphological characteristics, intelligently identify and classify the gender and age of the crowd; (4) Advertisement recommendation decision: According to the results of the crowd classification, different advertisements are classified and placed, and personalized recommendation technology is used to deliver different information to People who want to be designated; ⑸Intelligent human-computer interaction: Through somatosensory technology and voice recognition technology, the captured human motion and voice signals are converted into control signals to carry out human-computer intelligent interaction between people and advertising machines.
Description
技术领域technical field
本发明属于智能识别领域,特别是一种用于裸眼3D广告的多感知生物识别交互方法。The invention belongs to the field of intelligent identification, in particular to a multi-sensory biological identification interaction method for naked-eye 3D advertisements.
背景技术Background technique
3D显示技术,特别是裸眼3D技术及产品正在推动一场从平板显示转向立体显示的技术革命。国际上的研究已经如火如荼,如索尼、三星、LG公司已经推出裸眼3D产品。目前我国3D产业尚处于发展初期,裸眼3D产业也刚刚起步,尽管国内已经有多家企业积极投入到裸眼3D技术和产品研发中,但我国裸眼3D技术市场发展时间短,且没有专业正规的科研机构,同时由于立体成像技术是一种交叉边缘科技技术,研究的人少之又少,市场上裸眼3D产品售价较高,市场处于教育和培育阶段。目前已上市的裸眼3D设备中,视差障壁式3D技术是比较常见的类型,这种技术多运用在3D手机(夏普SH8168U)、便携式游戏机(任天堂3DS)这类小尺寸的设备上。而东芝的56吋3D电视则使用了柱状透镜技术,因为电视机对亮度、分辨率的要求较高,前者还难以胜任比较苛刻的应用环境。3D display technology, especially glasses-free 3D technology and products are promoting a technological revolution from flat panel display to stereoscopic display. International research has been in full swing, such as Sony, Samsung, and LG companies have launched glasses-free 3D products. At present, my country's 3D industry is still in the early stage of development, and the naked eye 3D industry has just started. Although many domestic companies have actively invested in the research and development of naked eye 3D technology and products, the development time of the naked eye 3D technology market in China is short, and there is no professional and formal scientific research. At the same time, because stereoscopic imaging technology is a cross-border technology, there are very few people who study it. The price of naked-eye 3D products in the market is relatively high, and the market is in the stage of education and cultivation. Among the glasses-free 3D devices currently on the market, the parallax barrier 3D technology is a relatively common type. This technology is mostly used in small-sized devices such as 3D mobile phones (Sharp SH8168U) and portable game consoles (Nintendo 3DS). Toshiba's 56-inch 3D TV uses lenticular lens technology, because the TV has higher requirements on brightness and resolution, and the former is still difficult to meet the harsh application environment.
裸眼3D广告机是利用人两眼具有视差的特性,在不需要任何辅助设备(如3D眼镜、头盔等)的情况下,即可获得具有空间、深度的逼真立体影像。画中事物即可以凸出于画面之外,也可以深藏于画面之中。色彩艳丽、层次分明、活灵活现、栩栩如生,是真正意义上的三维立体影像。裸眼立体影像以其真实生动的表现力,优美高雅的环境感染力,强烈震撼的视觉冲击力深受广大消费者的青睐。这种新、特、奇表现手法的影像产品,广泛应用到广告传媒、展览展示、旅游招商、婚纱摄影、科研教学、游戏娱乐、工业应用、建筑设计、手机等多个行业。Glasses-free 3D advertising machine uses the parallax characteristic of human eyes to obtain realistic three-dimensional images with space and depth without any auxiliary equipment (such as 3D glasses, helmets, etc.). Things in the painting can protrude outside the picture, or can be hidden deep in the picture. Colorful, distinct, vivid and lifelike, it is a true three-dimensional image. Glasses-free three-dimensional images are favored by consumers for their true and vivid expressive power, elegant and elegant environmental appeal, and strong and shocking visual impact. This new, unique, and unique image product is widely used in advertising media, exhibitions, tourism investment, wedding photography, scientific research and teaching, game entertainment, industrial applications, architectural design, mobile phones and other industries.
数字广告机(以下简称广告机)是一种新媒体概念,指在商场、机场及其他人流汇聚的公共场所,通过大屏幕终端显示设备,发布商业、财经或娱乐信息的多媒体专业视听系统。作为传统户外广告的替代者,广告机所包含的信息量非常大,可以随时方便地更新,并且能够联网,这都是传统户外广告所无法比拟的。广告机不仅可以向公众发布多媒体广告信息、带来全新的商业创新价值,而且被认为是显示产业新成长的动力。经过近五年的发展,广告机已经走过了单机、网络、交互等几个阶段,并逐步向智能化迈进。英特尔中国区总监施养维表示:“传统的、基于平面表现形式的内容及广告推送对消费者的吸引力越来越小,互动式的媒体传播和营销方式即将成为主流”。Digital advertising machine (hereinafter referred to as advertising machine) is a new media concept, which refers to a multimedia professional audio-visual system that releases commercial, financial or entertainment information through large-screen terminal display equipment in shopping malls, airports and other public places where other people gather. As a substitute for traditional outdoor advertising, the advertising machine contains a large amount of information, can be easily updated at any time, and can be connected to the Internet, which is unmatched by traditional outdoor advertising. Advertising machines can not only release multimedia advertising information to the public and bring new commercial innovation value, but also are considered to be the driving force for the new growth of the display industry. After nearly five years of development, the advertising machine has gone through several stages such as stand-alone, network, and interactive, and is gradually moving towards intelligence. Shi Yangwei, director of Intel China, said: "Traditional content and advertisements based on graphic representations are becoming less and less attractive to consumers, and interactive media communication and marketing methods will soon become the mainstream."
智能化的裸眼3D广告机产品作为“第五媒体”,将为互动式媒体传播应用的实现提供必不可少的技术平台。”智能化的裸眼3D广告机是指终端显示设备根据所识别的不同消费者特质而自动推出针对性的裸眼3D广告信息。基于对这一趋势的分析,英特尔认为:智能化裸眼3D广告机系统不仅要能支持更清晰、生动的裸眼3D画面显示以吸引消费者,拥有更友善的触屏互动操作界面,还需要能识别消费者特质、即时自动推出定制化的广告信息。这些应用的实现需要强大的处理器技术、硬件平台以及软件解决方案的支持。As the "fifth media", the intelligent naked-eye 3D advertising machine products will provide an indispensable technical platform for the realization of interactive media communication applications. "Intelligent naked-eye 3D advertising machine means that the terminal display device automatically launches targeted naked-eye 3D advertising information according to the identified characteristics of different consumers. Based on the analysis of this trend, Intel believes that: intelligent naked-eye 3D advertising machine system Not only must it be able to support clearer and vivid naked-eye 3D images to attract consumers, it must have a more friendly touch-screen interactive operation interface, but it also needs to be able to identify consumer characteristics and automatically launch customized advertising information in real time. The realization of these applications requires Supported by powerful processor technology, hardware platform and software solutions.
随着裸眼3D广告机互联网技术的发展、新的应用、及其需要处理的海量数据的增长,基于“智能互动平台”的裸眼3D广告机处理方案显得尤为重要。对于智能裸眼3D广告机的硬件资源上能满足不同应用模式的需求,同时,针对裸眼3D广告机行业开发的控制器和显示器等硬件也日趋丰富,而相对薄弱的环节是相应的软件解决方案,即智能识别技术。虽然目前国际上报道了一些智能识别技术,但是它们大多处于实验室研发阶段,而且考虑的人物对象和环境也比较单一。然而,事实上裸眼3D广告机的应用领域是完全开放式的复杂环境,服务对象及环境都极不确定:进入摄像机视野的人群数量、性别年龄分布、站立或行走姿势等都是无法约束的,摄像机前的环境,如车辆穿行、明暗变化、天气影响等也是无法控制的。这一切都给人体检测、追踪、识别带来困难。此外,现有的一些识别技术多基于人物的脸部特征,少数基于头发颜色和听力特征,但是裸眼3D广告机应用环境的上述特性使得单一特征的精确检测几乎无法实现,因此不可仅仅依赖少数几种特征做出决策。With the development of naked-eye 3D advertising machine Internet technology, new applications, and the growth of massive data that needs to be processed, the naked-eye 3D advertising machine processing solution based on the "intelligent interactive platform" is particularly important. The hardware resources of smart naked-eye 3D advertising machines can meet the needs of different application modes. At the same time, hardware such as controllers and displays developed for the naked-eye 3D advertising machine industry is also becoming more and more abundant, and the relatively weak link is the corresponding software solutions. That is, intelligent recognition technology. Although some intelligent recognition technologies have been reported internationally, most of them are in the laboratory research and development stage, and the person objects and environments considered are relatively single. However, in fact, the application field of naked-eye 3D advertising machines is a completely open and complex environment, and the service objects and environments are extremely uncertain: the number of people entering the camera's field of view, gender and age distribution, standing or walking posture, etc. cannot be restricted. The environment in front of the camera, such as vehicles passing through, changes in light and shade, and weather effects, are also uncontrollable. All these bring difficulties to human body detection, tracking and identification. In addition, some existing recognition technologies are mostly based on the facial features of people, and a few are based on hair color and hearing features. However, the above-mentioned characteristics of the naked-eye 3D advertising machine application environment make it almost impossible to achieve accurate detection of a single feature, so it is not possible to rely solely on a few characteristics to make a decision.
此外,目前体感技术,语音识别技术等支持人机交互技术相继出现,通过体感传感器以及语音采集器,捕获人体的动作或者语音,通过信息处理技术,将人体的工作或者语音转换为相应的指令,用来控制智能广告机的播放以及进行相应的人机交互具有重要的意义。In addition, human-computer interaction technologies such as somatosensory technology and speech recognition technology have emerged one after another. Through somatosensory sensors and voice collectors, the movements or voices of the human body are captured, and the work or voice of the human body is converted into corresponding instructions through information processing technology. It is of great significance to control the playback of intelligent advertising machines and perform corresponding human-computer interaction.
发明内容Contents of the invention
本发明的目的在于填补现有技术的空白,提供一种将多种特征的提取与融合且具有智能推荐技术的用于裸眼3D广告的多感知生物识别交互方法。The purpose of the present invention is to fill in the gaps in the prior art, and provide a multi-sensory biometric interaction method for naked-eye 3D advertisements that extracts and fuses multiple features and has intelligent recommendation technology.
本发明解决其技术问题所采用的技术方案是:The technical solution adopted by the present invention to solve its technical problems is:
一种用于裸眼3D广告的多感知生物识别交互方法,其步骤为:A multi-sensory biometric interaction method for naked-eye 3D advertisements, the steps of which are:
⑴在场景中的人体检测:在自动投放针对性的广告内容平台的建设中,通过摄像头对固定场景中的运动人体进行监视与检测,运动人体检测包括人体检测和跟踪,即检测和跟踪出现在视频中的人群;人的个数,即视频里人群的个数统计;⑴Human detection in the scene: In the construction of the platform for automatic delivery of targeted advertising content, the camera monitors and detects the moving human body in the fixed scene. The moving human body detection includes human body detection and tracking, that is, detection and tracking occur in The crowd in the video; the number of people, that is, the statistics of the number of crowd in the video;
⑵人群的特征检测:提取人群的人脸特征、发型特征、声音特征和形态特征,用来区分不同人群的年龄和性别;(2) Crowd feature detection: Extract the face features, hairstyle features, voice features and morphological features of the crowd to distinguish the age and gender of different groups;
⑶人群智能识别与分类:利用提取的人群的特征,对人群进行性别和年龄段的智能识别分类;(3) Crowd intelligent identification and classification: use the extracted characteristics of the crowd to intelligently identify and classify the gender and age of the crowd;
⑷广告推荐投放决策:根据人群分类的结果,将不同的广告进行分类投放,采用个性化的推荐技术,将不同的信息传递给想要指定的人群;⑷ Advertisement recommendation placement decision: According to the results of crowd classification, different advertisements are classified and placed, and personalized recommendation technology is used to deliver different information to the desired crowd;
⑸智能人机交互:通过体感技术,语音识别技术将捕获的人体动作和语音信号转变为控制信号,来进行人与广告机的人机智能交互。⑸Intelligent human-computer interaction: Through somatosensory technology and voice recognition technology, the captured human body movements and voice signals are converted into control signals to carry out human-computer intelligent interaction between people and advertising machines.
而且,步骤⑴中所述的人体检测和跟踪采用静态图像人体检测,其在特征选择上,采用Haar-like特征,EOH特征、纹理特征、颜色特征,在分类器的设计上,采用的是统计学习分类的方法。Moreover, the human body detection and tracking described in step (1) adopt static image human body detection, which adopts Haar-like features, EOH features, texture features, and color features in feature selection. In the design of classifiers, it uses statistical Learn how to classify.
而且,步骤⑴所述的人的个数采用动态更新模型,利用高斯分布来描述背景点颜色的概率分布,新运动人体的检测是根据当前帧图像与通过混合高斯模型建立起来的背景模型进行对比,获得当前帧的前景图像中的所有运动人体,然后,再与已记录的运动人体数据进行比较。这样,就得到许多当前帧中带运动性质的目标,将较小的运动目标与已经被跟踪的人体有重叠的目标丢弃,并对剩余的部分按照大小排序,之后就新加10个最大的人体来继续追踪。Moreover, the number of people described in step (1) adopts a dynamic update model, and uses a Gaussian distribution to describe the probability distribution of the background point color, and the detection of a new moving human body is based on the comparison between the current frame image and the background model established by the mixed Gaussian model , to obtain all moving human bodies in the foreground image of the current frame, and then compare with the recorded moving human body data. In this way, many moving objects in the current frame are obtained, and the smaller moving objects that overlap with the tracked human body are discarded, and the remaining parts are sorted by size, and then 10 largest human bodies are newly added. Come and keep track.
而且,步骤⑵所述的人脸特征和形态特征提取分为整体特征提取和局部特征提取两种方式,整体特征提取算法对人脸进行粗略的识别,局部特征提取从细节上进行补充。Moreover, the face feature and morphological feature extraction described in step (2) is divided into two ways: overall feature extraction and local feature extraction. The overall feature extraction algorithm roughly recognizes the face, and the local feature extraction supplements the details.
而且,步骤⑵所述的发型特征提取主要步骤如下:And, the main steps of the hairstyle feature extraction described in step (2) are as follows:
a、从视频中获取一帧图像进行检测,定时获取图像或是已经检测到场景中有人物时定时获取单帖图像进行处理。a. Obtain a frame of image from the video for detection, acquire images at regular intervals or acquire a single image at regular intervals for processing when a person in the scene has been detected.
b、图像预处理。b. Image preprocessing.
c、发型分割。c, hairstyle segmentation.
d、发型模型形成。d. Formation of hairstyle model.
而且,步骤⑶所述的人群智能识别与分类采用多分类器来实现决策融合。Moreover, the crowd intelligent identification and classification described in step (3) uses multiple classifiers to realize decision fusion.
而且,步骤⑶所述的性别识别的算法的步骤如下:And, the steps of the algorithm of the gender identification described in step (3) are as follows:
a、提取到人体的人脸、肤色、发型、装饰等各种特征,人脸特征包括整体和局部特征;a. Various features such as human face, skin color, hairstyle, decoration, etc. are extracted, and facial features include overall and local features;
b、利用线性判别方法提取训练检测到的人体的特征,同时利用Fisher方法提取训练样本和目标图像的特征;b. Use the linear discriminant method to extract the characteristics of the human body detected by training, and use the Fisher method to extract the characteristics of the training samples and target images;
c、利用动态聚类方法对目标图像进行分类,并计算单分类器的平均识别率;c, utilize the dynamic clustering method to classify the target image, and calculate the average recognition rate of a single classifier;
d、利用男女训练样本对SVR进行训练,得到一组参数值,利用训练好的SVR对测试样本进行分类;d. Use the male and female training samples to train the SVR to obtain a set of parameter values, and use the trained SVR to classify the test samples;
e、利用各种融合规则进行组合分类,比较不同融合规则的分类精度。e. Use various fusion rules for combined classification, and compare the classification accuracy of different fusion rules.
而且,步骤⑶所述的年龄段识别的算法的步骤如下:And, the steps of the algorithm of the age group identification described in step (3) are as follows:
a、提取到人体的人脸、肤色、发型、形态各种特征,对人脸,选取鉴别能力较强的左右眼睛、鼻子和嘴巴作为局部区域,在一定程度上降低计算复杂度,标定特征点分割出人脸的左右眼睛、鼻子和嘴巴区域;a. Extract the human face, skin color, hairstyle, and various features of the shape. For the face, select the left and right eyes, nose and mouth with strong discrimination ability as the local area, reduce the computational complexity to a certain extent, and calibrate the feature points Segment the left and right eyes, nose and mouth areas of the face;
b、利用线性判别方法提取训练检测到的人体的特征,同时利用Fisher方法提取训练样本和目标图像的特征;b. Use the linear discriminant method to extract the characteristics of the human body detected by training, and use the Fisher method to extract the characteristics of the training samples and target images;
c、利用动态聚类方法对目标图像进行分类,并计算单分类器的平均识别率;c, utilize the dynamic clustering method to classify the target image, and calculate the average recognition rate of a single classifier;
d、利用不同年龄段训练样本对SVR进行训练,得到一组参数值,利用训练好的SVR对测试样本进行分类;d. Use the training samples of different age groups to train the SVR to obtain a set of parameter values, and use the trained SVR to classify the test samples;
e、利用各种融合规则进行组合分类,比较不同融合规则的分类精度。根据对目标图像的分类迭代增加分类类别,缩小年龄段分类区间,以提高年龄估计精度。e. Use various fusion rules for combined classification, and compare the classification accuracy of different fusion rules. According to the classification of the target image, the classification category is iteratively increased, and the age classification interval is narrowed, so as to improve the age estimation accuracy.
而且,步骤⑷所述的广告推荐投放决策采用人口统计学的推荐技术,广告投放采用人工方式分类。Moreover, the advertisement recommendation placement decision described in step (4) adopts demographic recommendation technology, and the advertisement placement adopts manual classification.
本发明的优点和积极效果是:Advantage and positive effect of the present invention are:
本发明针对智能裸眼3D广告机,通过对复杂环境下的目标人群的多种特征的提取与融合,最终实现裸眼3D广告机广告的智能分类与投放决策以及基于人脸识别、体感技术和语音识别技术的人机多感知智能交互。由此通过多特征的提取、计算以及分类,对裸眼3D广告机的投放决策提供了重要的数据基础,填补了智能识别技术在裸眼3D广告机应用领域的空白。The present invention aims at the intelligent naked-eye 3D advertising machine, through the extraction and fusion of various characteristics of the target population in a complex environment, and finally realizes the intelligent classification and placement decision of the naked-eye 3D advertising machine advertisement and based on face recognition, somatosensory technology and voice recognition. Human-computer multi-sensory intelligent interaction of technology. Therefore, through the extraction, calculation and classification of multiple features, it provides an important data basis for the decision-making of the naked-eye 3D advertising machine, and fills the gap of intelligent recognition technology in the application field of the naked-eye 3D advertising machine.
本发明通过体感传感器以及语音采集器,捕获人体的动作或者语音信号,通过信息处理技术,将人体的动作或者语音信号转换为相应的指令,用来控制智能广告机的播放以及进行相应的人机交互。由此将人脸识别、体感技术和语音识别技术相融合,实现多感知生物识别技术在智能广告机上的应用,可以为不同的群体服务,吸引顾客的注意力。The present invention captures human body motions or voice signals through somatosensory sensors and voice collectors, and converts human body motions or voice signals into corresponding instructions through information processing technology, which is used to control the playback of intelligent advertising machines and perform corresponding man-machine interact. In this way, face recognition, somatosensory technology and voice recognition technology are integrated to realize the application of multi-sensory biometric technology on intelligent advertising machines, which can serve different groups and attract customers' attention.
附图说明Description of drawings
图1是本发明的多感知生物识别交互方法流程图;Fig. 1 is a flow chart of the multi-sensory biometric identification interaction method of the present invention;
图2是基于静态图像的人体检测框图;Figure 2 is a block diagram of human body detection based on static images;
图3是基于视频的人体检测框图;Fig. 3 is a block diagram of human body detection based on video;
图4是语音信号的处理系统框图;Fig. 4 is the processing system block diagram of voice signal;
图5是MFCC特征参数的计算方框图;Fig. 5 is the calculation block diagram of MFCC feature parameter;
图6是性别、年龄段的智能识别框图。Fig. 6 is a block diagram of intelligent identification of gender and age group.
具体实施方式Detailed ways
下面通过具体实施例对本发明作进一步详述,以下实施例只是描述性的,不是限定性的,不能以此限定本发明的保护范围。The present invention will be further described in detail below through the specific examples, the following examples are only descriptive, not restrictive, and cannot limit the protection scope of the present invention with this.
一种用于裸眼3D广告的多感知生物识别交互方法,其步骤为:A multi-sensory biometric interaction method for naked-eye 3D advertisements, the steps of which are:
⑴复杂场景中的人体检测:在自动投放针对性的广告内容平台的建设中,通过摄像头对固定场景中的运动人体进行监视与检测,是进一步自动识别性别、年龄段等的基础,主要有以下两个研究子内容:⑴Human detection in complex scenes: In the construction of a platform for automatic delivery of targeted advertising content, monitoring and detection of moving human bodies in fixed scenes through cameras is the basis for further automatic identification of gender, age group, etc., mainly as follows Two study sub-contents:
人体检测和跟踪:检测和跟踪出现在视频中的人群,可以同时进行多个目标的检测和跟踪,Human detection and tracking: detection and tracking of crowds appearing in the video, multiple targets can be detected and tracked at the same time,
人的个数:视频里人群的个数统计。Number of people: Statistics of the number of people in the video.
静态图像人体检测,如图1所示问题实际上是一个标准的模式识别问题,其中最为关键的两大方面内容是:特征的选择以及分类器的设计,在特征选择上,采用Haar-like特征,EOH特征、纹理特征、颜色特征等,在分类器的设计上,采用的是统计学习分类的方法,考虑使用已经成功应用于人脸检测领域的Adaboost学习算法,该算法在特征挑选和分类器设计上取得了令人满意的效果,用于人体检测和跟踪,本发明基于视频的运动人体检测与跟踪,摄像头的场景背景一般变换不大,这给运动人体检测和跟踪带来了很大的便利,同时大大降低了了检测与跟踪的复杂度。Static image human detection, as shown in Figure 1, is actually a standard pattern recognition problem. The two most critical aspects are: feature selection and classifier design. In feature selection, Haar-like features are used. , EOH feature, texture feature, color feature, etc. In the design of the classifier, the method of statistical learning classification is adopted, and the Adaboost learning algorithm that has been successfully applied in the field of face detection is considered. Satisfactory results have been obtained in the design, and it is used for human body detection and tracking. The present invention is based on video-based moving human body detection and tracking. The scene background of the camera generally changes little, which brings great benefits to moving human body detection and tracking. Convenience, while greatly reducing the complexity of detection and tracking.
基于视频的人体检测框图如图2所示,在处理视频检测运动人体之前,利用视频图像中的运动信息对其运动区域进行标记,去除不满足要求的区域。然后使用训练好的分类器对筛选出来的区域进行检测以得到最终的运动人体实例。前景检测主要是将视频图像分成运动的人体前景区域,与摄像头中不动部分的背景区域。自适应背景模型为静止背景建立背景模型,通过将当前图像帧和背景模型进行比较,确定出变化较大的区域即认为是前景区域。这种方法的计算速度很快,可以获得关于运动目标区域的完整精确的描述,但对场景中光照条件、大面积运动和噪声比较敏感,所以在实际应用中需采用一定的算法进行背景模型的动态更新以适应环境的变化。The block diagram of human body detection based on video is shown in Figure 2. Before processing the video to detect moving human body, the motion information in the video image is used to mark its moving area and remove the area that does not meet the requirements. Then use the trained classifier to detect the selected regions to get the final moving human instance. Foreground detection is mainly to divide the video image into the moving foreground area of the human body and the background area of the motionless part of the camera. The adaptive background model establishes a background model for the static background. By comparing the current image frame with the background model, the area with a large change is determined to be the foreground area. The calculation speed of this method is very fast, and a complete and accurate description of the moving target area can be obtained, but it is sensitive to the lighting conditions, large-area motion and noise in the scene, so it is necessary to use a certain algorithm for the background model in practical applications. Dynamically updated to adapt to changes in the environment.
本发明采用动态更新模型是用高斯分布(正态分布)来描述背景点颜色的概率分布。新运动人体的检测是根据当前帧图像与通过混合高斯模型建立起来的背景模型进行对比,获得当前帧的前景图像中的所有运动人体。其次,再与已记录的运动人体数据进行比较。这样,就得到许多当前帧中带运动性质的目标,将较小的运动目标(可能是由噪声引起的)与已经被跟踪的人体有重叠的目标丢弃,并对剩余的部分按照大小排序,之后就新加10个最大的人体来继续追踪,用于视频里人群的个数统计。The present invention uses a dynamic update model to describe the probability distribution of background point colors with Gaussian distribution (normal distribution). The detection of the new moving human body is based on comparing the current frame image with the background model established by the Gaussian mixture model to obtain all moving human bodies in the foreground image of the current frame. Secondly, it is compared with the recorded sports human body data. In this way, many moving targets in the current frame are obtained, and the smaller moving targets (probably caused by noise) that overlap with the tracked human body are discarded, and the remaining parts are sorted by size, and then Add 10 largest human bodies to continue tracking, and use them to count the number of people in the video.
⑵人群的特征检测:提取人群的一些特征,用来区分不同人群的年龄和性别,这些特征可综合使用,以提高系统的识别精度和可靠性,常用的特征有以下几种:(2) Feature detection of the crowd: Extract some features of the crowd to distinguish the age and gender of different groups. These features can be used comprehensively to improve the recognition accuracy and reliability of the system. The commonly used features are as follows:
a、人脸特征:人脸的整体或者局部特征。a. Facial features: the overall or partial features of the human face.
b、发型特征:主要指长短头发、以及光头等发型。b. Hairstyle features: mainly refers to long or short hair, and bald hair.
c、声音特征:语音、咳嗽、喘息、脚步声等。c. Sound characteristics: speech, coughing, wheezing, footsteps, etc.
b、形态特征:人体的身高,关键关节点的特征。b. Morphological characteristics: the height of the human body, the characteristics of key joint points.
人脸特征和形态特征提取可以分为整体特征提取和局部特征提取两种方式,整体特征提取算法对人脸进行粗略的识别,局部特征提取从细节上进行补充,但当识别出某些独特的局部特征(如光头、络腮胡等)时,则这些局部特征可以直接用来确定性别,整体特征和局部特征提取互相补充使人脸识别获得更佳效果。The extraction of facial features and morphological features can be divided into two methods: overall feature extraction and local feature extraction. The overall feature extraction algorithm roughly recognizes the face, and the local feature extraction supplements the details. When local features (such as bald head, beard, etc.) are used, these local features can be directly used to determine gender, and the extraction of overall features and local features complement each other to achieve better results in face recognition.
此外,特征融合也可以成功引用于局部特征提取,通过对一幅图像进行多次处理,如用局部二值模式进行处理,可以得到多幅特征图像,然后进行融合与特征提取,最后来进行人脸匹配。In addition, feature fusion can also be successfully applied to local feature extraction. By processing an image multiple times, such as using local binary mode, multiple feature images can be obtained, and then fusion and feature extraction are performed, and finally artificial The face matches.
四维局部二值模式特征提取是一个基于局部二值模式改进的新的方法,局部二值模式仅仅考虑了中心像素点与领域像素点的大小的局部结构关系,很好的描述了图像的局部纹理特征。但是三维局部二值模式却忽略了图像本身中心点的像素值对于局部细节特征的影响,四维局部二值模式特征就是在三维局部二值模式特征的基础上加上中心点的像素值,并将考虑梯度信息考虑进去,形成一个新的融合局部二值模式特征。The four-dimensional local binary mode feature extraction is a new method based on the improvement of the local binary mode. The local binary mode only considers the local structure relationship between the center pixel and the area pixel, and describes the local texture of the image well. feature. However, the three-dimensional local binary mode ignores the influence of the pixel value of the center point of the image itself on the local detail features. Taking the gradient information into account, a new fused local binary pattern feature is formed.
其中,对于发型特征的提取,Among them, for the extraction of hairstyle features,
发型特征提取的主要步骤如下:The main steps of hair style feature extraction are as follows:
第一步:从视频中获取一帧图像进行检测。可以定时获取图像或是已经检测到场景中有人物时定时获取单帖图像进行处理。Step 1: Obtain a frame of image from the video for detection. Images can be acquired regularly or a single image can be acquired regularly for processing when a person in the scene has been detected.
第二步:图像预处理。这一步处理的效果直接影响后面发型模型提取的效果。发型提取与人脸提取有些区别,人脸模型的刚性比较强,并且人脸的反光度较头发低,即人脸对光照的敏感度比头发要低。因此,这一步的处理至关重要。The second step: image preprocessing. The effect of this step directly affects the effect of subsequent hairstyle model extraction. Hairstyle extraction is somewhat different from face extraction. The rigidity of the face model is relatively strong, and the reflectivity of the face is lower than that of the hair, that is, the sensitivity of the face to light is lower than that of the hair. Therefore, the treatment of this step is very important.
第三步:发型分割。这一步是整个过程的难点,主要是头发的柔性比较大,其形状不固定,并且随光照变化影响较大,无论是光照强度还是光照方向。The third step: hairstyle segmentation. This step is the difficulty of the whole process, mainly because the hair is relatively flexible, its shape is not fixed, and it is greatly affected by changes in light, whether it is light intensity or light direction.
第四步:发型模型形成。在进行分割后,可以从分割图中提取出发型模型,取出的发型模型用于性别或年龄判断。Step 4: The hairstyle model is formed. After segmentation, the hairstyle model can be extracted from the segmentation map, and the extracted hairstyle model is used for gender or age judgment.
对于声音特征的提取,For the extraction of sound features,
声音的特征提取占有举足轻重的作用,在识别系统之前有数个必须经历的环节包括数字化,预滤波,采样,A/D变换,声音信号的预处理(包括分帧与加窗),具体流程如图4所示。Sound feature extraction plays a decisive role. There are several links that must be experienced before the recognition system, including digitization, pre-filtering, sampling, A/D conversion, and pre-processing of sound signals (including framing and windowing). The specific process is shown in the figure 4.
对于声音的特征提取方法采用的是梅尔频标倒谱参数(MFCC),这种方法把人耳的听觉特性考虑在内,即把声音的频谱转化到Mel频标的非线性频谱区域当中,然后再通过一定的同态处理过程转换到频谱域上,其计算流程图如图5所示.For the sound feature extraction method, the Mel frequency scale cepstrum parameter (MFCC) is used. This method takes the auditory characteristics of the human ear into consideration, that is, the frequency spectrum of the sound is transformed into the nonlinear spectrum region of the Mel frequency scale, and then Then it is converted to the spectral domain through a certain homomorphic processing process, and its calculation flow chart is shown in Figure 5.
与其他传统的声音特征提取方法相比,MFCC参数具有良好的识别性能和抗噪声能力,而且这种参数比基于声道模型的LPCC相比具有更好的鲁邦性,更符合人耳的听觉特性,而且当信噪比降低时仍然具有较好的识别性能。Compared with other traditional sound feature extraction methods, MFCC parameters have good recognition performance and anti-noise ability, and this parameter has better Lubang performance than LPCC based on channel model, which is more in line with the human ear characteristics, and still have good recognition performance when the signal-to-noise ratio decreases.
至于对语音特征进行分类的方法我们通过SVM(支持向量机)进行分类识别,通过将训练好所得到的MFCC参数传输给SVM分类器,然后投影到高维空间,使它们成为线性可分,再利用线性划分的原理来判断分类边界,最终达成语音识别的目的。As for the method of classifying speech features, we use SVM (Support Vector Machine) for classification and recognition, and transfer the trained MFCC parameters to the SVM classifier, and then project them into a high-dimensional space to make them linearly separable. The principle of linear division is used to judge the classification boundary, and finally achieve the purpose of speech recognition.
⑶人群智能识别与分类:利用提取的人群特征,对人群进行性别和年龄段的智能识别分类,利用统计学或者人工智能之类的方法对提取的人群特征进行性别和年龄段识别分类。(3) Crowd intelligent identification and classification: use the extracted crowd characteristics to intelligently identify and classify the gender and age groups of the crowd, and use statistics or artificial intelligence methods to identify and classify the extracted crowd features by gender and age group.
性别识别:对男女进行分类。Gender Recognition: Classify males and females.
年龄段识别:对老年、中年、少年进行区分和分类。Age Group Identification: Distinguish and classify old, middle-aged, and young people.
采用分类器来对检测的人体进行年龄分类是人脸识别领域中关于年龄变化研究的主要方法,可以说这样的年龄估计方法是行之有效的。但是由于单个分类器都有其各自的优缺点,因而在同样情况下不同分类器的识别结果可能相差非常大。多分类器决策融合是模式发展一种趋势,其目的旨在提高分类精度,并度量模式的密集程度。Using a classifier to classify the age of the detected human body is the main method of age change research in the field of face recognition. It can be said that such an age estimation method is effective. But because a single classifier has its own advantages and disadvantages, the recognition results of different classifiers may be very different in the same situation. Multi-classifier decision fusion is a trend in pattern development, and its purpose is to improve classification accuracy and measure the density of patterns.
·Fisher线性判别法。作为单分类器来说,Fisher线性判别法识别性能在不同的数据库上均不错。但是,利用Fisher线性判别法在进行降维处理时,还是会丢失一些有用信息,而这些信息对于后面步骤可能很重要,这也是该方法的不足之处。• Fisher's linear discriminant method. As a single classifier, Fisher's linear discriminant method has good recognition performance on different databases. However, when using Fisher's linear discriminant method for dimensionality reduction, some useful information will still be lost, which may be important for subsequent steps, which is also the shortcoming of this method.
·动态聚类算法。动态聚类算法采用C-均值算法,它是一种常用的基于近邻法则的无监督学习方法。其基本思想很简单,首先确定需要的群数c,选好c个代表点,用这些代表点作为初始类型,再对样本集H中每个样本X找出相距最近的代表点,将X归到这个最近的代表点所在的群中去。这样,第一次迭代就用近邻法则将H初步分为c群。下一次迭代就在这个基础上以上次迭代所得的各群的均值向量作为新的代表点,再次按近邻法则将H分为c群,直到分群稳定为止。· Dynamic clustering algorithm. The dynamic clustering algorithm uses the C-means algorithm, which is a commonly used unsupervised learning method based on the nearest neighbor rule. The basic idea is very simple. First, determine the required number of groups c, select c representative points, use these representative points as the initial type, and then find the nearest representative point for each sample X in the sample set H, and return X to Go to the group where the nearest representative point is located. In this way, H is initially divided into c groups by the nearest neighbor rule in the first iteration. In the next iteration, on this basis, the mean vector of each group obtained in the previous iteration is used as a new representative point, and H is divided into c groups according to the neighbor rule again until the grouping is stable.
·支持向量回归机(Support Vector Regression,简称SVR)。支持向量回归机算法是支持向量机方法在回归问题上的推广。通过引入不敏感损失函数和核函数,可以很好地应用于非线性回归分析,并且对小样本集问题具有良好的预测性能。Support Vector Regression (SVR for short). The support vector regression machine algorithm is the generalization of the support vector machine method on the regression problem. By introducing insensitive loss function and kernel function, it can be well applied to nonlinear regression analysis, and has good predictive performance for small sample set problems.
在前面所述提取特征的基础上,我们可以运用分类器来进行人的性别和年龄识别,如图6所示。下面介绍详细的算法:On the basis of the extracted features mentioned above, we can use the classifier to identify the gender and age of people, as shown in Figure 6. The detailed algorithm is described below:
①性别识别的算法的主要步骤如下:① The main steps of the gender recognition algorithm are as follows:
a、根据前面的介绍的方法提取到人体的人脸、肤色、发型、装饰等各种特征。人脸特征包括整体和局部特征。a. According to the method introduced above, various features such as human face, skin color, hairstyle, decoration, etc. are extracted. Facial features include overall and local features.
b、利用线性判别方法提取训练检测到的人体的特征,同时利用Fisher方法提取训练样本和目标图像的特征。b. Use the linear discriminant method to extract the features of the human body detected by training, and use the Fisher method to extract the features of the training samples and target images.
c、利用动态聚类方法对目标图像进行分类,并计算单分类器的平均识别率。c. Use the dynamic clustering method to classify the target image, and calculate the average recognition rate of a single classifier.
d、利用男女训练样本对SVR进行训练,得到一组参数值。利用训练好的SVR对测试样本进行分类。d. Use the male and female training samples to train the SVR to obtain a set of parameter values. Use the trained SVR to classify the test samples.
e、利用各种融合规则进行组合分类,比较不同融合规则的分类精度。e. Use various fusion rules for combined classification, and compare the classification accuracy of different fusion rules.
②年龄段识别的算法的主要步骤如下:② The main steps of the age group identification algorithm are as follows:
a、根据前面的介绍的方法提取到人体的人脸、肤色、发型、装饰等各种特征。对人脸,选取鉴别能力较强的左右眼睛、鼻子和嘴巴作为局部区域,在一定程度上降低计算复杂度。标定特征点分割出人脸的左右眼睛、鼻子和嘴巴区域。a. According to the method introduced above, various features such as human face, skin color, hairstyle, decoration, etc. are extracted. For human faces, the left and right eyes, nose and mouth with strong discrimination ability are selected as local regions to reduce the computational complexity to a certain extent. The calibration feature points are used to segment the left and right eye, nose and mouth areas of the face.
b、利用线性判别方法提取训练检测到的人体的特征,同时利用Fisher方法提取训练样本和目标图像的特征。b. Use the linear discriminant method to extract the features of the human body detected by training, and use the Fisher method to extract the features of the training samples and target images.
c、利用动态聚类方法对目标图像进行分类,并计算单分类器的平均识别率。c. Use the dynamic clustering method to classify the target image, and calculate the average recognition rate of a single classifier.
d、利用不同年龄段(老、中、少等)训练样本对SVR进行训练,得到一组参数值。利用训练好的SVR对测试样本进行分类。d. Using training samples of different age groups (old, middle, young, etc.) to train the SVR to obtain a set of parameter values. Use the trained SVR to classify the test samples.
e、利用各种融合规则进行组合分类,比较不同融合规则的分类精度。根据对目标图像的分类迭代增加分类类别,缩小年龄段分类区间,以提高年龄估计精度。⑷广告推荐投放决策:根据人群分类的结果,将不同的广告进行分类投放。采用个性化的推荐技术,在适当的时间,将恰当信息传递给想要传递的人群,使广告期望达到更好的效果。e. Use various fusion rules for combined classification, and compare the classification accuracy of different fusion rules. According to the classification of the target image, the classification category is iteratively increased, and the age classification interval is narrowed, so as to improve the age estimation accuracy. ⑷ Advertisement recommendation placement decision: According to the results of crowd classification, different advertisements are classified and placed. Using personalized recommendation technology, at the right time, the right information is delivered to the people who want to deliver it, so that the advertisement can expect to achieve better results.
采用人口统计学的推荐技术(Demographic-based),该技术最常用的人口学特征变量是年龄、性别和地理位置,有时收入变量也会被使用(如高档宾馆中的顾客)。年龄和性别是由系统识别得到,地理位置由系统终端所在地决定。Demographic-based recommendation technology is adopted. The most commonly used demographic characteristic variables of this technology are age, gender and geographical location, and sometimes income variables are also used (such as customers in high-end hotels). Age and gender are identified by the system, and geographic location is determined by the location of the system terminal.
广告初始阶段通过人工方式进行分类,并标注适宜人群属性(年龄、性别、地理位置和收入等)。若有可能,系统增加观察量:用户对某一广告的观看时间。利用该时间,得到某类人群对各个广告的喜爱程度,以便调整各个广告的播放策略。In the initial stage of the advertisement, it is manually classified and marked with suitable crowd attributes (age, gender, geographical location and income, etc.). If possible, the system adds observations: the time a user watches an ad. The time is used to obtain the liking degree of a certain group of people for each advertisement, so as to adjust the playing strategy of each advertisement.
⑸智能人机交互:通过体感技术,语音识别技术等将捕获的人体动作和语音信号转变为控制信号,来进行广告机的人机智能交互。通过人脸识别技术,可以识别人的身份、年龄以及性别,然后为不同的人群播放不同的广告,例如,为小孩子播放动画广告,为女士播放化妆品广告,实现人机交互,在一定程度上可以吸引人的注意力。⑸Intelligent human-computer interaction: Through somatosensory technology, voice recognition technology, etc., the captured human motion and voice signals are converted into control signals to carry out the human-computer intelligent interaction of the advertising machine. Through face recognition technology, it is possible to identify a person's identity, age, and gender, and then play different advertisements for different groups of people. For example, play animation advertisements for children and cosmetic advertisements for women to achieve human-computer interaction. To a certain extent Can attract people's attention.
通过体感传感器以及语音采集器,捕获人体的动作或者语音信号,通过信息处理技术,将人体的动作或者语音信号转换为相应的指令,用来控制智能广告机的播放以及进行相应的人机交互。Through somatosensory sensors and voice collectors, human body movements or voice signals are captured, and through information processing technology, human body movements or voice signals are converted into corresponding instructions, which are used to control the playback of intelligent advertising machines and perform corresponding human-computer interaction.
其中,可以通过体感传感器,捕获人体的动作信号,如向左,向右,向上,向下,以及手掌合闭等信号,通过信号处理,将人体的动作信号转换为智能广告机的控制信号,来控制广告视频的播放,暂停,快进,快退,片源的更换,屏幕放大,缩小等功能,该技术可以吸引小朋友以及年轻的广告群体。Among them, the motion signals of the human body can be captured through the somatosensory sensor, such as left, right, up, down, and palm closing and other signals, and the human body’s motion signals can be converted into control signals of the intelligent advertising machine through signal processing. To control advertising video playback, pause, fast forward, fast rewind, source replacement, screen zoom in, zoom out and other functions, this technology can attract children and young advertising groups.
此外,通过语音识别技术,将人的语音如开始,暂停,播放,快进,快退,片源的更换,屏幕放大,缩小等语音信号,转换为相应的控制信号,来控制广告视频的播放。该技术将大大方便残疾人等弱势群体。In addition, through voice recognition technology, human voices such as start, pause, play, fast forward, rewind, change of film source, screen zoom in, zoom out and other voice signals are converted into corresponding control signals to control the playback of advertising videos . The technology will greatly facilitate vulnerable groups such as the disabled.
由此将人脸识别、体感技术和语音识别技术相融合,实现多感知生物识别技术在智能广告机上的应用,可以为不同的群体服务,吸引顾客的注意力。In this way, face recognition, somatosensory technology and voice recognition technology are integrated to realize the application of multi-sensory biometric technology on intelligent advertising machines, which can serve different groups and attract customers' attention.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510282106.XA CN104915000A (en) | 2015-05-27 | 2015-05-27 | Multisensory biological recognition interaction method for naked eye 3D advertisement |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510282106.XA CN104915000A (en) | 2015-05-27 | 2015-05-27 | Multisensory biological recognition interaction method for naked eye 3D advertisement |
Publications (1)
Publication Number | Publication Date |
---|---|
CN104915000A true CN104915000A (en) | 2015-09-16 |
Family
ID=54084136
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510282106.XA Pending CN104915000A (en) | 2015-05-27 | 2015-05-27 | Multisensory biological recognition interaction method for naked eye 3D advertisement |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104915000A (en) |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105550679A (en) * | 2016-02-29 | 2016-05-04 | 深圳前海勇艺达机器人有限公司 | Judgment method for robot to cyclically monitor recording |
WO2017049918A1 (en) * | 2015-09-25 | 2017-03-30 | 杭州海康威视数字技术股份有限公司 | Information publishing terminal and method |
CN106599656A (en) * | 2016-11-28 | 2017-04-26 | 深圳超多维科技有限公司 | Display method, device and electronic equipment |
CN106650601A (en) * | 2016-10-18 | 2017-05-10 | 武汉慧能机器人科技有限公司 | Human-machine interaction method and intelligent robot |
CN106625711A (en) * | 2016-12-30 | 2017-05-10 | 华南智能机器人创新研究院 | Method for positioning intelligent interaction of robot |
CN107146096A (en) * | 2017-03-07 | 2017-09-08 | 浙江工业大学 | Intelligent video advertisement display method and device |
CN107221252A (en) * | 2017-06-22 | 2017-09-29 | 蚌埠市麦沃特信息科技有限公司 | A kind of energy-saving signboard |
CN107369398A (en) * | 2017-06-22 | 2017-11-21 | 蚌埠市麦沃特信息科技有限公司 | A kind of voice advertisement method |
CN107678545A (en) * | 2017-09-26 | 2018-02-09 | 深圳市维冠视界科技股份有限公司 | A kind of information interactive terminal and method |
WO2018035831A1 (en) * | 2016-08-25 | 2018-03-01 | 武克易 | Advertisement video playback system |
CN107766782A (en) * | 2016-08-23 | 2018-03-06 | 中兴通讯股份有限公司 | A kind of method and device of age-colony classification |
CN108022543A (en) * | 2017-11-27 | 2018-05-11 | 深圳中科呼图电子商务有限公司 | A kind of advertisement autonomous demenstration method, system and advertisement machine and application |
CN108076128A (en) * | 2016-12-28 | 2018-05-25 | 北京市商汤科技开发有限公司 | User property extracting method, device and electronic equipment |
CN108269133A (en) * | 2018-03-23 | 2018-07-10 | 深圳悠易阅科技有限公司 | A kind of combination human bioequivalence and the intelligent advertisement push method and terminal of speech recognition |
CN109117833A (en) * | 2018-10-29 | 2019-01-01 | 天津市汇诚智慧体育科技有限公司 | A kind of full crowd movement's intelligent management system based on big data |
CN109886730A (en) * | 2019-01-15 | 2019-06-14 | 北京奇艺世纪科技有限公司 | Put-on method, equipment and system, service equipment and method, processing equipment, medium |
CN109934149A (en) * | 2019-03-06 | 2019-06-25 | 百度在线网络技术(北京)有限公司 | Method and apparatus for output information |
CN110002302A (en) * | 2018-08-09 | 2019-07-12 | 浙江新再灵科技股份有限公司 | A kind of elevator switch door detection system and method based on deep learning |
CN110070865A (en) * | 2019-04-03 | 2019-07-30 | 北京容联易通信息技术有限公司 | A kind of guidance robot with voice and image identification function |
CN110674760A (en) * | 2019-09-27 | 2020-01-10 | 广州云从人工智能技术有限公司 | Interactive guiding method, device, equipment and medium based on mobile robot |
CN113158834A (en) * | 2021-03-31 | 2021-07-23 | 中北大学南通智能光机电研究院 | Multi-perception face recognition magic cube |
CN113435518A (en) * | 2021-06-29 | 2021-09-24 | 青岛海尔科技有限公司 | Feature fusion interaction method and device based on multiple modes |
CN113961680A (en) * | 2021-09-28 | 2022-01-21 | 京东方智慧物联科技有限公司 | Human-computer interaction based session processing method and device, medium and electronic equipment |
CN115099866A (en) * | 2022-07-07 | 2022-09-23 | 悦诚智慧(厦门)科技有限公司 | Advertisement delivery system based on AI glasses |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070092245A1 (en) * | 2005-10-20 | 2007-04-26 | Honeywell International Inc. | Face detection and tracking in a wide field of view |
CN101236599A (en) * | 2007-12-29 | 2008-08-06 | 浙江工业大学 | Face recognition detection device based on multi-camera information fusion |
CN102289672A (en) * | 2011-06-03 | 2011-12-21 | 天津大学 | Infrared gait identification method adopting double-channel feature fusion |
CN102298709A (en) * | 2011-09-07 | 2011-12-28 | 江西财经大学 | Energy-saving intelligent identification digital signage fused with multiple characteristics in complicated environment |
CN102368295A (en) * | 2011-09-07 | 2012-03-07 | 常州蓝城信息科技有限公司 | Method for extracting age characteristics under complex environment |
CN102368294A (en) * | 2011-09-07 | 2012-03-07 | 常州蓝城信息科技有限公司 | Method for extracting sex characteristics under complex environment |
CN102446327A (en) * | 2010-10-15 | 2012-05-09 | 汉王科技股份有限公司 | Advertisement putting device and method based on face recognition |
CN102542252A (en) * | 2011-11-18 | 2012-07-04 | 江西财经大学 | Intelligent advertisement delivery system |
CN102930454A (en) * | 2012-10-07 | 2013-02-13 | 乐配(天津)科技有限公司 | Intelligent 3D (Three Dimensional) advertisement recommendation method based on multiple perception technologies |
-
2015
- 2015-05-27 CN CN201510282106.XA patent/CN104915000A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070092245A1 (en) * | 2005-10-20 | 2007-04-26 | Honeywell International Inc. | Face detection and tracking in a wide field of view |
CN101236599A (en) * | 2007-12-29 | 2008-08-06 | 浙江工业大学 | Face recognition detection device based on multi-camera information fusion |
CN102446327A (en) * | 2010-10-15 | 2012-05-09 | 汉王科技股份有限公司 | Advertisement putting device and method based on face recognition |
CN102289672A (en) * | 2011-06-03 | 2011-12-21 | 天津大学 | Infrared gait identification method adopting double-channel feature fusion |
CN102298709A (en) * | 2011-09-07 | 2011-12-28 | 江西财经大学 | Energy-saving intelligent identification digital signage fused with multiple characteristics in complicated environment |
CN102368295A (en) * | 2011-09-07 | 2012-03-07 | 常州蓝城信息科技有限公司 | Method for extracting age characteristics under complex environment |
CN102368294A (en) * | 2011-09-07 | 2012-03-07 | 常州蓝城信息科技有限公司 | Method for extracting sex characteristics under complex environment |
CN102542252A (en) * | 2011-11-18 | 2012-07-04 | 江西财经大学 | Intelligent advertisement delivery system |
CN102930454A (en) * | 2012-10-07 | 2013-02-13 | 乐配(天津)科技有限公司 | Intelligent 3D (Three Dimensional) advertisement recommendation method based on multiple perception technologies |
Non-Patent Citations (2)
Title |
---|
中央文化企业国有资产监督管理领导小组办公室等: "《文化产业专题研究报告 (下)》", 31 December 2014 * |
李学庆等编: "《和谐人机环境 2007》", 31 May 2008 * |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106557940B (en) * | 2015-09-25 | 2019-09-17 | 杭州海康威视数字技术股份有限公司 | Information release terminal and method |
WO2017049918A1 (en) * | 2015-09-25 | 2017-03-30 | 杭州海康威视数字技术股份有限公司 | Information publishing terminal and method |
CN106557940A (en) * | 2015-09-25 | 2017-04-05 | 杭州海康威视数字技术股份有限公司 | Information release terminal and method |
CN105550679B (en) * | 2016-02-29 | 2019-02-15 | 深圳前海勇艺达机器人有限公司 | A kind of judgment method of robot cycle monitoring recording |
CN105550679A (en) * | 2016-02-29 | 2016-05-04 | 深圳前海勇艺达机器人有限公司 | Judgment method for robot to cyclically monitor recording |
CN107766782A (en) * | 2016-08-23 | 2018-03-06 | 中兴通讯股份有限公司 | A kind of method and device of age-colony classification |
WO2018035831A1 (en) * | 2016-08-25 | 2018-03-01 | 武克易 | Advertisement video playback system |
CN106650601A (en) * | 2016-10-18 | 2017-05-10 | 武汉慧能机器人科技有限公司 | Human-machine interaction method and intelligent robot |
CN106599656A (en) * | 2016-11-28 | 2017-04-26 | 深圳超多维科技有限公司 | Display method, device and electronic equipment |
WO2018121541A1 (en) * | 2016-12-28 | 2018-07-05 | 北京市商汤科技开发有限公司 | User attribute extraction method and apparatus, and electronic device |
CN108076128A (en) * | 2016-12-28 | 2018-05-25 | 北京市商汤科技开发有限公司 | User property extracting method, device and electronic equipment |
CN106625711A (en) * | 2016-12-30 | 2017-05-10 | 华南智能机器人创新研究院 | Method for positioning intelligent interaction of robot |
CN107146096B (en) * | 2017-03-07 | 2020-08-18 | 浙江工业大学 | Intelligent video advertisement display method and device |
CN107146096A (en) * | 2017-03-07 | 2017-09-08 | 浙江工业大学 | Intelligent video advertisement display method and device |
CN107221252A (en) * | 2017-06-22 | 2017-09-29 | 蚌埠市麦沃特信息科技有限公司 | A kind of energy-saving signboard |
CN107369398A (en) * | 2017-06-22 | 2017-11-21 | 蚌埠市麦沃特信息科技有限公司 | A kind of voice advertisement method |
CN107678545A (en) * | 2017-09-26 | 2018-02-09 | 深圳市维冠视界科技股份有限公司 | A kind of information interactive terminal and method |
CN108022543A (en) * | 2017-11-27 | 2018-05-11 | 深圳中科呼图电子商务有限公司 | A kind of advertisement autonomous demenstration method, system and advertisement machine and application |
CN108269133A (en) * | 2018-03-23 | 2018-07-10 | 深圳悠易阅科技有限公司 | A kind of combination human bioequivalence and the intelligent advertisement push method and terminal of speech recognition |
CN110002302A (en) * | 2018-08-09 | 2019-07-12 | 浙江新再灵科技股份有限公司 | A kind of elevator switch door detection system and method based on deep learning |
CN109117833A (en) * | 2018-10-29 | 2019-01-01 | 天津市汇诚智慧体育科技有限公司 | A kind of full crowd movement's intelligent management system based on big data |
CN109886730A (en) * | 2019-01-15 | 2019-06-14 | 北京奇艺世纪科技有限公司 | Put-on method, equipment and system, service equipment and method, processing equipment, medium |
CN109934149B (en) * | 2019-03-06 | 2022-08-09 | 百度在线网络技术(北京)有限公司 | Method and apparatus for outputting information |
CN109934149A (en) * | 2019-03-06 | 2019-06-25 | 百度在线网络技术(北京)有限公司 | Method and apparatus for output information |
CN110070865A (en) * | 2019-04-03 | 2019-07-30 | 北京容联易通信息技术有限公司 | A kind of guidance robot with voice and image identification function |
CN110674760A (en) * | 2019-09-27 | 2020-01-10 | 广州云从人工智能技术有限公司 | Interactive guiding method, device, equipment and medium based on mobile robot |
CN113158834A (en) * | 2021-03-31 | 2021-07-23 | 中北大学南通智能光机电研究院 | Multi-perception face recognition magic cube |
CN113435518A (en) * | 2021-06-29 | 2021-09-24 | 青岛海尔科技有限公司 | Feature fusion interaction method and device based on multiple modes |
CN113435518B (en) * | 2021-06-29 | 2024-03-22 | 青岛海尔科技有限公司 | Multi-mode-based interaction method and device for feature fusion |
CN113961680A (en) * | 2021-09-28 | 2022-01-21 | 京东方智慧物联科技有限公司 | Human-computer interaction based session processing method and device, medium and electronic equipment |
CN115099866A (en) * | 2022-07-07 | 2022-09-23 | 悦诚智慧(厦门)科技有限公司 | Advertisement delivery system based on AI glasses |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104915000A (en) | Multisensory biological recognition interaction method for naked eye 3D advertisement | |
CN107197384B (en) | The multi-modal exchange method of virtual robot and system applied to net cast platform | |
CN106951867B (en) | Face identification method, device, system and equipment based on convolutional neural networks | |
CN109635680B (en) | Multitask attribute identification method and device, electronic equipment and storage medium | |
CN107146096B (en) | Intelligent video advertisement display method and device | |
CN102930454A (en) | Intelligent 3D (Three Dimensional) advertisement recommendation method based on multiple perception technologies | |
CN109948447B (en) | Discovery and evolution presentation method of character network relationship based on video image recognition | |
CN102298709A (en) | Energy-saving intelligent identification digital signage fused with multiple characteristics in complicated environment | |
KR20160012902A (en) | Method and device for playing advertisements based on associated information between audiences | |
CN109922373A (en) | Method for processing video frequency, device and storage medium | |
CN106778506A (en) | A kind of expression recognition method for merging depth image and multi-channel feature | |
Xu et al. | Saliency prediction on omnidirectional image with generative adversarial imitation learning | |
CN112686928B (en) | Moving target visual tracking method based on multi-source information fusion | |
CN108269133A (en) | A kind of combination human bioequivalence and the intelligent advertisement push method and terminal of speech recognition | |
CN105426850A (en) | Human face identification based related information pushing device and method | |
Yargıç et al. | A lip reading application on MS Kinect camera | |
CN107341434A (en) | Processing method, device and the terminal device of video image | |
CN103854016B (en) | Jointly there is human body behavior classifying identification method and the system of feature based on directivity | |
CN105913275A (en) | Clothes advertisement putting method and system based on video leading role identification | |
CN102129644A (en) | Intelligent advertising system having functions of audience characteristic perception and counting | |
CN116484318A (en) | A speech training feedback method, device and storage medium | |
CN103605986A (en) | Human motion recognition method based on local features | |
CN102088597A (en) | Method for estimating video visual salience through dynamic and static combination | |
CN108256513A (en) | A kind of intelligent video analysis method and intelligent video record system | |
CN105303571A (en) | Time-space saliency detection method for video processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20150916 |
|
RJ01 | Rejection of invention patent application after publication |