[go: up one dir, main page]

CN111652155A - A method and system for recognizing human motion intention - Google Patents

A method and system for recognizing human motion intention Download PDF

Info

Publication number
CN111652155A
CN111652155A CN202010500206.6A CN202010500206A CN111652155A CN 111652155 A CN111652155 A CN 111652155A CN 202010500206 A CN202010500206 A CN 202010500206A CN 111652155 A CN111652155 A CN 111652155A
Authority
CN
China
Prior art keywords
motion
coordinates
human body
human
video information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010500206.6A
Other languages
Chinese (zh)
Inventor
王兴坚
张卿
王少萍
苗忆南
安麦灵
张超
张敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN202010500206.6A priority Critical patent/CN111652155A/en
Publication of CN111652155A publication Critical patent/CN111652155A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J11/00Manipulators not otherwise provided for
    • B25J11/008Manipulators for service tasks
    • B25J11/009Nursing, e.g. carrying sick persons, pushing wheelchairs, distributing drugs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H20/00ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
    • G16H20/30ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to physical therapies or activities, e.g. physiotherapy, acupressure or exercising

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Medical Informatics (AREA)
  • Psychiatry (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Physical Education & Sports Medicine (AREA)
  • Epidemiology (AREA)
  • Social Psychology (AREA)
  • Primary Health Care (AREA)
  • Public Health (AREA)
  • Nursing (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method and a system for identifying human motion intention. The method comprises the following steps: acquiring a motion coordinate in the motion of a human body; determining coordinates of a sight line attention point in human motion through a wearable glasses type eye movement instrument; recognizing the coordinates of obstacles in a visual field scene in the motion of a human body through a trained convolutional neural network model; and predicting the human body movement yaw angle information through a trained recurrent neural network model according to the movement coordinate, the attention point coordinate and the obstacle coordinate. The exoskeleton robot can quickly and accurately identify the walking movement direction of the human body so as to provide accurate wearer movement information for the exoskeleton robot.

Description

一种人体运动意图的识别方法及系统A method and system for recognizing human motion intention

技术领域technical field

本发明涉及计算机模式识别技术领域,特别是涉及一种人体运动意图的识别方法及系统。The present invention relates to the technical field of computer pattern recognition, in particular to a method and system for recognizing human motion intention.

背景技术Background technique

随着科学技术的不断进步,机器参与人类的生活的比例日益增多。机械外骨骼可以用于医疗、康复训练、军事辅助等情景。动力外骨骼可以提高单兵作战的能力,助力外骨骼可以代偿病人的行走能力以及辅助病人进行康复训练。With the continuous advancement of science and technology, the proportion of machines participating in human life is increasing. Mechanical exoskeletons can be used in medical treatment, rehabilitation training, military assistance and other scenarios. Powered exoskeletons can improve the ability of individual soldiers to fight, and power exoskeletons can compensate for the patient's walking ability and assist patients in rehabilitation training.

要实现高效的人机协同工作,就需要极强的人机交互能力和智能控制。而传统的人机交互方法中,机器人冗余运动、人机交互系统运动误差累积以及用户交互体验差等问题仍然没有得到有效的解决。而作为人机交互系统的参与者,人类控制系统本身是目前地球上最高级、复杂、智能的系统。为了提升人机交互能力,最有效的途径就是仿效人类,将生物学、控制技术、传感器数据融合、机器学习等多个学科领域有机融合,以获得机器人智能、友好的最佳平台。建立智能的人机交互策略,主要可以分为以下三个阶段:人体运动意图估计、机器人助力控制和稳定性控制。如何对人体运动意图定性、定量地捕获分析并快速、准确地映射到柔顺控制系统的输入,进而完成基于以外骨骼穿戴操作者为主控制模型的半自动控制策略和基于人和机器人协调控制模型的主动智能控制策略是给穿戴者提供自然舒适、合理健康和友好临场感觉的前提和基础。在下肢外骨骼机器人与穿戴者交互助力运动过程中,感知系统对于实时、快速、精准地捕获人体运动意图一直是一项研究难点。To achieve efficient human-machine collaborative work, strong human-machine interaction capabilities and intelligent control are required. In traditional human-computer interaction methods, the problems of redundant robot motion, accumulation of motion errors in human-computer interaction systems, and poor user interaction experience have not been effectively solved. As a participant in the human-computer interaction system, the human control system itself is the most advanced, complex and intelligent system on the earth. In order to improve the ability of human-computer interaction, the most effective way is to imitate human beings and organically integrate multiple disciplines such as biology, control technology, sensor data fusion, and machine learning to obtain the best platform for robots to be intelligent and friendly. The establishment of an intelligent human-computer interaction strategy can be divided into the following three stages: human motion intention estimation, robot assist control and stability control. How to qualitatively and quantitatively capture and analyze human motion intention and map it to the input of the compliance control system quickly and accurately, and then complete the semi-automatic control strategy based on the exoskeleton wearing operator-based control model and the active control model based on the human-robot coordinated control model The intelligent control strategy is the premise and foundation to provide the wearer with natural comfort, reasonable health and friendly presence. In the process of the lower limb exoskeleton robot interacting with the wearer to assist the movement, the perception system has always been a research difficulty to capture the human movement intention in real time, quickly and accurately.

发明内容SUMMARY OF THE INVENTION

本发明的目的是提供一种人体运动意图的识别方法及系统,用以精准快速地获得人体运动意图。The purpose of the present invention is to provide a method and system for recognizing human body motion intention, so as to obtain the human body motion intention accurately and quickly.

为实现上述目的,本发明提供了如下方案:For achieving the above object, the present invention provides the following scheme:

一种人体运动意图的识别方法,所述方法包括:A method for identifying human motion intention, the method comprising:

获取人体运动中的运动坐标;Get the motion coordinates in human motion;

通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标;Determine the coordinates of the attention point of sight in human movement through a wearable eye-tracking device;

通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标;Identify the coordinates of obstacles in the visual field of human motion through the trained convolutional neural network model;

根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。According to the motion coordinates, the attention point coordinates and the obstacle coordinates, the human body motion yaw angle information is predicted through the trained recurrent neural network model.

可选的,所述获取人体运动中的运动坐标,具体包括:Optionally, the acquiring motion coordinates in the motion of the human body specifically includes:

获取穿戴在人体上的单目摄像头中的视频信息;Obtain video information from a monocular camera worn on the human body;

获取穿戴在人体上的IMU传感器的运动信息;Obtain the motion information of the IMU sensor worn on the human body;

采用视觉惯性里程计方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。Using the visual inertial odometry scheme, the video information and the motion information are processed to obtain motion coordinates in the motion of the human body.

可选的,所述通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标,具体包括:Optionally, the determining the coordinates of the attention point of sight in the human body movement by using a wearable eye-tracking device specifically includes:

通过所述可穿戴的眼镜式眼动仪获取人眼瞳孔的视频信息;Obtain the video information of the pupil of the human eye through the wearable glasses eye tracker;

根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。According to the video information of the pupil of the human eye, the coordinates of the attention point of the line of sight in the movement of the human body are determined.

可选的,所述通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标,具体包括:Optionally, identifying the coordinates of obstacles in the visual field scene in human motion through the trained convolutional neural network model specifically includes:

通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;Obtain the video information of the visual field scene in the human body movement through the wearable glasses-type eye tracker;

通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;Identify each frame of image in the video information of the visual field by using the trained convolutional neural network model, and determine the obstacles in the image;

通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。The obstacles in each frame of image are selected by a rectangular frame, and the coordinates of the obstacles in the image are obtained.

可选的,还包括:采用样条插值方法将所述运动坐标、关注点坐标以及所述障碍物坐标同步成等间隔的信号。Optionally, the method further includes: using a spline interpolation method to synchronize the motion coordinates, the focus point coordinates and the obstacle coordinates into signals at equal intervals.

本发明还提供了一种人体运动意图的识别系统,所述系统包括:The present invention also provides a system for identifying human motion intentions, the system comprising:

运动坐标获取模块,用于获取人体运动中的运动坐标;The motion coordinate acquisition module is used to acquire motion coordinates in human motion;

关注点坐标确定模块,用于通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标;The attention point coordinate determination module is used to determine the coordinates of the line of sight attention point in the human body movement through the wearable glasses eye tracker;

障碍物坐标识别模块,用于通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标;The obstacle coordinate recognition module is used to identify the obstacle coordinates in the visual field of human motion through the trained convolutional neural network model;

预测模块,用于根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。The prediction module is used for predicting the yaw angle information of human body motion through the trained recurrent neural network model according to the motion coordinates, the attention point coordinates and the obstacle coordinates.

可选的,所述运动坐标获取模块具体包括:Optionally, the motion coordinate acquisition module specifically includes:

第一视频信息获取单元,用于获取穿戴在人体上的单目摄像头中的视频信息;a first video information acquisition unit, used for acquiring video information in a monocular camera worn on the human body;

运动信息获取单元,用于获取穿戴在人体上的IMU传感器的运动信息;A motion information acquisition unit, used to acquire motion information of the IMU sensor worn on the human body;

处理单元,用于采用视觉惯性里程计方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。The processing unit is configured to use the visual inertial odometry scheme to process the video information and the motion information to obtain motion coordinates in the motion of the human body.

可选的,所述关注点坐标确定模块具体包括:Optionally, the coordinate determination module of the point of interest specifically includes:

第二视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人眼瞳孔的视频信息;a second video information acquisition unit, configured to acquire video information of the pupil of the human eye through the wearable eye-tracker;

关注点坐标确定单元,用于根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。The attention point coordinate determination unit is used for determining the coordinates of the line of sight attention point in the motion of the human body according to the video information of the pupil of the human eye.

可选的,所述障碍物坐标识别模块具体包括:Optionally, the obstacle coordinate identification module specifically includes:

第三视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;a third video information acquisition unit, configured to acquire video information of visual field scenes in human motion through the wearable glasses eye tracker;

障碍物识别单元,用于通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;The obstacle identification unit is used to identify each frame of image in the video information of the visual field through the trained convolutional neural network model, and determine the obstacle in the image;

坐标确定单元,用于通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。The coordinate determination unit is used to frame the obstacles in each frame of images through a rectangular frame to obtain the coordinates of the obstacles in the image.

可选的,还包括:Optionally, also include:

插值模块,用于采用样条插值方法将所述运动坐标、关注点坐标以及所述障碍物坐标同步成等间隔的信号。The interpolation module is used for synchronizing the motion coordinates, the attention point coordinates and the obstacle coordinates into signals at equal intervals by using a spline interpolation method.

根据本发明提供的具体实施例,本发明公开了以下技术效果:According to the specific embodiments provided by the present invention, the present invention discloses the following technical effects:

(1)本发明中利用穿戴单目摄像头和IMU集成的模块,采用视觉惯性里程计方案,可以实现无需外部定位系统,实现对人体自身在空间环境中的定位,该定位信息可用于解算人体运动过程中的偏航角信息;(1) In the present invention, a module integrated with a wearable monocular camera and an IMU is used, and the visual inertial odometry solution is adopted, which can realize the positioning of the human body itself in the space environment without the need for an external positioning system, and the positioning information can be used to solve the human body. Yaw angle information during motion;

(2)本发明中利用眼动仪实时获取人眼关注的兴趣点,通过该数据可以分析穿戴者对空间环境内物体的关注程度,结合发明中所述的识别出的物体的坐标点,其中可分析物体对于人体行走方向的影响;(2) In the present invention, the eye tracker is used to obtain the interest points that the human eye pays attention to in real time. Through this data, the wearer's degree of attention to objects in the space environment can be analyzed. Combined with the coordinate points of the identified objects described in the invention, wherein It can analyze the influence of objects on the walking direction of the human body;

(3)本发明中采用循环神经网络作为融合多种信息的网络模型,其好处在于循环神经网络对于历史信息有记忆能力,结合历史信息和当前的运动信息,以及一段时间内人在环境中的关注点,可以预测出人下一时刻的偏航角信息,精准快速地获得人体运动意图。(3) In the present invention, the cyclic neural network is used as a network model that integrates various information. The focus point can predict the yaw angle information of the person at the next moment, and obtain the human motion intention accurately and quickly.

附图说明Description of drawings

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the accompanying drawings required in the embodiments will be briefly introduced below. Obviously, the drawings in the following description are only some of the present invention. In the embodiments, for those of ordinary skill in the art, other drawings can also be obtained according to these drawings without creative labor.

图1为本发明实施例人体运动意图的识别方法的流程图;1 is a flowchart of a method for recognizing human motion intention according to an embodiment of the present invention;

图2为本发明实施例LSTM循环神经网络的结构示意图;2 is a schematic structural diagram of an LSTM cyclic neural network according to an embodiment of the present invention;

图3为本发明实施例人体运动意图的识别系统的结构框图。FIG. 3 is a structural block diagram of a system for recognizing human motion intention according to an embodiment of the present invention.

具体实施方式Detailed ways

下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only a part of the embodiments of the present invention, but not all of the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative efforts shall fall within the protection scope of the present invention.

本发明的目的是提供一种人体运动意图的识别方法及系统,用以精准快速地获得人体运动意图。The purpose of the present invention is to provide a method and system for recognizing human body motion intention, so as to obtain the human body motion intention accurately and quickly.

为使本发明的上述目的、特征和优点能够更加明显易懂,下面结合附图和具体实施方式对本发明作进一步详细的说明。In order to make the above objects, features and advantages of the present invention more clearly understood, the present invention will be described in further detail below with reference to the accompanying drawings and specific embodiments.

如图1所示,一种人体运动意图的识别方法包括以下步骤:As shown in Figure 1, a method for identifying human motion intention includes the following steps:

步骤101:获取人体运动中的运动坐标。具体的,获取穿戴在人体上的单目摄像头中的视频信息;获取穿戴在人体上的IMU传感器的运动信息;采用视觉惯性里程方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。Step 101: Obtain motion coordinates in human body motion. Specifically, the video information in the monocular camera worn on the human body is obtained; the motion information of the IMU sensor worn on the human body is obtained; the visual inertial mileage scheme is used to process the video information and the motion information to obtain the human body Motion coordinates in motion.

本发明实例中,选用VINS-Mono作为视觉惯性里程计的方案,首先利用张正友标记法,求得相机的内参和外参,IMU出厂时提供IMU的漂移和随机误差,利用ORB方法识别图像角点,由K最近邻法实现两帧特征点之间的匹配,利用N点透视法求解相机的旋转和平移,估计相机的位姿,IMU运动学计算IMU漂移误差,利用IMU预积分方法,定义状态向量为滑动窗口内所有相机状态、相机到IMU的变化参数、所有3D点的逆深度,观测点是观测到的匹配的特征点,利用图优化方法求解观测点对于状态向量的最大后验估计,进而更精确测量人体在空间环境中的坐标点。In the example of the present invention, VINS-Mono is selected as the solution of the visual inertial odometer. First, the Zhang Zhengyou marking method is used to obtain the internal and external parameters of the camera. The IMU provides the drift and random error of the IMU when it leaves the factory, and uses the ORB method to identify the image corners. , the K nearest neighbor method is used to achieve the matching between the feature points of the two frames, the N-point perspective method is used to solve the rotation and translation of the camera, the pose of the camera is estimated, the IMU kinematics is used to calculate the IMU drift error, and the IMU pre-integration method is used to define the state The vector is all camera states in the sliding window, the change parameters from the camera to the IMU, the inverse depth of all 3D points, and the observation points are the observed matching feature points. The graph optimization method is used to solve the maximum a posteriori estimation of the observation points for the state vector, Then, the coordinate points of the human body in the space environment can be measured more accurately.

步骤102:通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标。具体的,通过所述可穿戴的眼镜式眼动仪获取人眼通孔的视频信息;根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。Step 102 : Determine the coordinates of the attention point of sight in the movement of the human body through the wearable eye-tracking device. Specifically, the video information of the through-hole of the human eye is obtained through the wearable eye-tracking device; the coordinates of the attention point of the line of sight in the movement of the human body are determined according to the video information of the pupil of the human eye.

采用可穿戴的眼镜式眼动仪,人体穿戴了眼动仪后,通过该眼动仪可录制人眼角度下的周围环境视频;穿戴式眼动仪利用其设备上的红外摄像头和红外线获得人眼瞳孔的视频信息,根据眼动仪内置的算法,求解出穿戴者的在该周围环境视频中的眼睛所注视的平面二维像素坐标点(x,y)和以眼动仪设备为原点的三维坐标系下的三维坐标点(X,Y,Z)。A wearable eye-tracking device with glasses is used. After the human body wears the eye-tracking device, the video of the surrounding environment from the angle of the human eye can be recorded through the eye-tracking device. The video information of the eye pupil, according to the built-in algorithm of the eye tracker, to solve the plane two-dimensional pixel coordinate point (x, y) that the wearer's eyes in the surrounding video are looking at and the eye tracker device as the origin. The three-dimensional coordinate point (X, Y, Z) in the three-dimensional coordinate system.

步骤103:通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标。具体的,通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。Step 103: Recognize the coordinates of obstacles in the visual field scene during human motion through the trained convolutional neural network model. Specifically, the video information of the visual field scene in human motion is obtained through the wearable eye-tracking device; the trained convolutional neural network model is used to identify each frame of the image in the video information of the visual field scene, and determine the visual field scene video information in the image. The obstacles in each frame are selected by a rectangular frame, and the coordinates of the obstacles in the image are obtained.

所述眼镜式眼动仪可以录制人眼角度下的周围环境视频,通过该视频流输入至目标检测的卷积神经网络。该卷积神经网络采用现有的开源方案,事先经过开源的目标检测数据集进行训练,可以识别环境中预先设定好的标签物体,比如人体、椅子、显示器等。将视频分割成单帧图片,对视频中每一帧输入至卷积神经网络进行识别,用矩形框标记出每一帧图像中的标签物体的平面位置,可得到所识别的矩形框中心坐标(x,y)和矩形框的高度h和宽度w,作为人眼所观测到的物体位置坐标(x,y,w,h)。The glasses-type eye tracker can record the video of the surrounding environment from the angle of the human eye, and input the video stream to the convolutional neural network for target detection. The convolutional neural network adopts the existing open source solution and is trained on the open source target detection data set in advance, and can identify pre-set label objects in the environment, such as human body, chair, monitor, etc. The video is divided into single-frame pictures, and each frame in the video is input to the convolutional neural network for identification, and the plane position of the label object in each frame of image is marked with a rectangular frame, and the center coordinates of the identified rectangular frame can be obtained ( x, y) and the height h and width w of the rectangular frame, as the position coordinates (x, y, w, h) of the object observed by the human eye.

步骤104:根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。Step 104: Predict the yaw angle information of human body movement through the trained recurrent neural network model according to the motion coordinates, the attention point coordinates and the obstacle coordinates.

循环神经网络为长短时记忆网络模型,该网络输入人眼注视点坐标向量、障碍物体坐标向量(x,y,w,h),运动坐标向量,将上述向量坐标整合成一个向量作为输入向量,用姿态坐标信息求解出下一时刻和当前时刻的偏航角的变化量作为当前输入向量对应的标签数据,输入至长短时记忆网络模型。训练时长短时记忆网络设定步长为3,即训练当前时刻T时,结合了历史的T-1、T-2时刻的位置和姿态坐标在网络中曾经输出过的全连接层向量,经过当前时刻的网络共同计算出新的输出,输出偏航角角度值,用于预测人体运动的方向。The recurrent neural network is a long and short-term memory network model. The network inputs the coordinate vector of the human eye gaze point, the coordinate vector of the obstacle object (x, y, w, h), and the motion coordinate vector, and integrates the above vector coordinates into a vector as the input vector, Using the attitude coordinate information, the variation of the yaw angle at the next moment and the current moment is obtained as the label data corresponding to the current input vector, and input to the long and short-term memory network model. Training time The short-term memory network sets the step size to 3, that is, when training the current time T, the fully connected layer vector that has been output in the network by combining the historical positions and attitude coordinates of T-1 and T-2 times is passed. The network at the current moment jointly calculates a new output, and outputs the yaw angle value, which is used to predict the direction of human movement.

本发明实例中,循环神经网络选择为长短时记忆LSTM网络,图2为其网络结构。该网络中设置有记忆门、遗忘门,可以消除循环神经网络的梯度消失问题,并且具有记忆长时间的信息的能力。经过同步的信号,输入至双向长短时记忆网络中,经过网络处理,输出当前时刻按照轨迹运动所需要的转角信息。In the example of the present invention, the cyclic neural network is selected as a long-short-term memory LSTM network, and FIG. 2 is its network structure. The network is equipped with a memory gate and a forgetting gate, which can eliminate the gradient disappearance problem of the recurrent neural network, and has the ability to memorize long-term information. The synchronized signal is input into the bidirectional long-term and short-term memory network, and after network processing, the information of the rotation angle required by the trajectory movement at the current moment is output.

LSTM的关键就是memory block(记忆块),主要包含了三个门(forget gate、inputgate、output gate)与一个记忆单元(cell)。方框内上方的那条水平线,被称为cell state(单元状态)。LSTM的网络更新关系如下:The key to LSTM is the memory block, which mainly includes three gates (forget gate, inputgate, output gate) and a memory unit (cell). The horizontal line above the box is called the cell state. The network update relationship of LSTM is as follows:

it=α(Wxixt+Whiht-1+Wcict-1+bi)i t =α(W xi x t +W hi h t-1 +W ci c t-1 +b i )

ft=α(Wxfxt+Whfht-1+Wcfct-1+bf)f t =α(W xf x t +W hf h t-1 +W cf c t-1 +b f )

ot=α(Wxoxt+Whoht-1+Wcoct-1+bo)o t =α(W xo x t +W ho h t-1 +W co c t-1 +b o )

ct=ft*ct-1+it*tanh(Wxcxt+Whcht-1+bc)c t =f t *c t-1 +i t *tanh(W xc x t +W hc h t-1 +b c )

ht=ot*tanh(ct)h t =o t *tanh(c t )

其中,it是输入门,ft是遗忘门,ot是输出门,ct是单元状态,ht是t状态的输出。where i t is the input gate, f t is the forget gate, o t is the output gate, c t is the cell state, and h t is the output of the t state.

输入至循环神经网络模型中的信号,不同信号的频率不同,并且相位也有差距,需要进行处理,利用样条插值方法,将上述信息同步成等间隔的信号。The signals input to the cyclic neural network model have different frequencies and different phases, and need to be processed. The spline interpolation method is used to synchronize the above information into equally spaced signals.

如图3所示,本发明还提供了一种人体运动意图的识别系统,所述系统包括:As shown in FIG. 3 , the present invention also provides a system for recognizing human motion intention, the system comprising:

运动坐标获取模块301,用于获取人体运动中的运动坐标。The motion coordinate acquisition module 301 is used for acquiring motion coordinates in human body motion.

所述运动坐标获取模块301具体包括:The motion coordinate acquisition module 301 specifically includes:

第一视频信息获取单元,用于获取穿戴在人体上的单目摄像头中的视频信息;a first video information acquisition unit, used for acquiring video information in a monocular camera worn on the human body;

运动信息获取单元,用于获取穿戴在人体上的IMU传感器的运动信息;A motion information acquisition unit, used to acquire motion information of the IMU sensor worn on the human body;

处理单元,用于采用视觉惯性里程计方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。The processing unit is configured to use the visual inertial odometry scheme to process the video information and the motion information to obtain motion coordinates in the motion of the human body.

关注点坐标确定模块302,用于通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标。The attention point coordinate determination module 302 is used for determining the coordinates of the line of sight attention point in the human body movement through the wearable eye-tracker with glasses.

所述关注点坐标确定模块302具体包括:The attention point coordinate determination module 302 specifically includes:

第二视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人眼瞳孔的视频信息;a second video information acquisition unit, configured to acquire video information of the pupil of the human eye through the wearable eye-tracker;

关注点坐标确定单元,用于根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。The attention point coordinate determination unit is used for determining the coordinates of the line of sight attention point in the motion of the human body according to the video information of the pupil of the human eye.

障碍物坐标识别模块303,用于通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标。The obstacle coordinate identification module 303 is configured to identify the obstacle coordinates in the visual field scene during human motion through the trained convolutional neural network model.

所述障碍物坐标识别模块303具体包括:The obstacle coordinate identification module 303 specifically includes:

第三视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;a third video information acquisition unit, configured to acquire video information of visual field scenes in human motion through the wearable glasses eye tracker;

障碍物识别单元,用于通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;The obstacle identification unit is used to identify each frame of image in the video information of the visual field through the trained convolutional neural network model, and determine the obstacle in the image;

坐标确定单元,用于通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。The coordinate determination unit is used to frame the obstacles in each frame of images through a rectangular frame to obtain the coordinates of the obstacles in the image.

预测模块304,用于根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。The prediction module 304 is configured to predict the yaw angle information of the human body movement through the trained recurrent neural network model according to the movement coordinates, the attention point coordinates and the obstacle coordinates.

该系统还包括:The system also includes:

插值模块,用于采用样条插值方法将所述运动坐标、关注点坐标以及所述障碍物坐标同步成等间隔的信号。The interpolation module is used for synchronizing the motion coordinates, the attention point coordinates and the obstacle coordinates into signals at equal intervals by using a spline interpolation method.

本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例公开的系统而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。The various embodiments in this specification are described in a progressive manner, and each embodiment focuses on the differences from other embodiments, and the same and similar parts between the various embodiments can be referred to each other. For the system disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant part can be referred to the description of the method.

本文中应用了具体个例对本发明的原理及实施方式进行了阐述,以上实施例的说明只是用于帮助理解本发明的方法及其核心思想;同时,对于本领域的一般技术人员,依据本发明的思想,在具体实施方式及应用范围上均会有改变之处。综上所述,本说明书内容不应理解为对本发明的限制。In this paper, specific examples are used to illustrate the principles and implementations of the present invention. The descriptions of the above embodiments are only used to help understand the methods and core ideas of the present invention; meanwhile, for those skilled in the art, according to the present invention There will be changes in the specific implementation and application scope. In conclusion, the contents of this specification should not be construed as limiting the present invention.

Claims (10)

1.一种人体运动意图的识别方法,其特征在于,所述方法包括:1. a recognition method of human body motion intention, is characterized in that, described method comprises: 获取人体运动中的运动坐标;Get the motion coordinates in human motion; 通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标;Determine the coordinates of the attention point of sight in human movement through a wearable eye-tracking device; 通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标;Identify the coordinates of obstacles in the visual field of human motion through the trained convolutional neural network model; 根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。According to the motion coordinates, the attention point coordinates and the obstacle coordinates, the human body motion yaw angle information is predicted through the trained recurrent neural network model. 2.根据权利要求1所述的人体运动意图的识别方法,其特征在于,所述获取人体运动中的运动坐标,具体包括:2. the identification method of human body motion intention according to claim 1, is characterized in that, described obtaining the motion coordinates in human body motion, specifically comprises: 获取穿戴在人体上的单目摄像头中的视频信息;Obtain video information from a monocular camera worn on the human body; 获取穿戴在人体上的IMU传感器的运动信息;Obtain the motion information of the IMU sensor worn on the human body; 采用视觉惯性里程计方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。Using the visual inertial odometry scheme, the video information and the motion information are processed to obtain motion coordinates in the motion of the human body. 3.根据权利要求1所述的人体运动意图的识别方法,其特征在于,所述通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标,具体包括:3. The method for recognizing human motion intention according to claim 1, characterized in that, determining the coordinates of the attention point of sight in human motion by a wearable eye-tracking device, specifically comprising: 通过所述可穿戴的眼镜式眼动仪获取人眼瞳孔的视频信息;Obtain the video information of the pupil of the human eye through the wearable glasses eye tracker; 根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。According to the video information of the pupil of the human eye, the coordinates of the attention point of the line of sight in the movement of the human body are determined. 4.根据权利要求1所述的人体运动意图的识别方法,其特征在于,所述通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标,具体包括:4. the identification method of human body motion intention according to claim 1, is characterized in that, described by the trained convolutional neural network model identifying the obstacle coordinates in the visual field scene in the human body motion, specifically comprises: 通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;Obtain the video information of the visual field scene in the human body movement through the wearable glasses-type eye tracker; 通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;Identify each frame of image in the video information of the visual field by using the trained convolutional neural network model, and determine the obstacles in the image; 通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。The obstacles in each frame of image are selected by a rectangular frame, and the coordinates of the obstacles in the image are obtained. 5.根据权利要求1所述的人体运动意图的识别方法,其特征在于,还包括:采用样条插值方法将所述运动坐标、关注点坐标以及所述障碍物坐标同步成等间隔的信号。5 . The method for recognizing human motion intention according to claim 1 , further comprising: using a spline interpolation method to synchronize the motion coordinates, the coordinates of the point of interest, and the coordinates of the obstacle into equally spaced signals. 6 . 6.一种人体运动意图的识别系统,其特征在于,所述系统包括:6. A recognition system for human body motion intention, wherein the system comprises: 运动坐标获取模块,用于获取人体运动中的运动坐标;A motion coordinate acquisition module, used to acquire motion coordinates in human motion; 关注点坐标确定模块,用于通过可穿戴的眼镜式眼动仪确定人体运动中视线关注点的坐标;The attention point coordinate determination module is used to determine the coordinates of the line of sight attention point in the human body movement through the wearable glasses eye tracker; 障碍物坐标识别模块,用于通过训练好的卷积神经网络模型识别人体运动中视野场景中的障碍物坐标;The obstacle coordinate recognition module is used to identify the obstacle coordinates in the visual field of human motion through the trained convolutional neural network model; 预测模块,用于根据所述运动坐标、关注点坐标以及所述障碍物坐标,通过训练好的循环神经网络模型预测人体运动偏航角信息。The prediction module is used for predicting the yaw angle information of human body motion through the trained recurrent neural network model according to the motion coordinates, the attention point coordinates and the obstacle coordinates. 7.根据权利要求6所述的人体运动意图的识别系统,其特征在于,所述运动坐标获取模块具体包括:7. The recognition system of human body motion intention according to claim 6, wherein the motion coordinate acquisition module specifically comprises: 第一视频信息获取单元,用于获取穿戴在人体上的单目摄像头中的视频信息;a first video information acquisition unit, used for acquiring video information in a monocular camera worn on the human body; 运动信息获取单元,用于获取穿戴在人体上的IMU传感器的运动信息;A motion information acquisition unit, used to acquire motion information of the IMU sensor worn on the human body; 处理单元,用于采用视觉惯性里程计方案,对所述视频信息以及所述运动信息进行处理,得到人体运动中的运动坐标。The processing unit is configured to use the visual inertial odometry scheme to process the video information and the motion information to obtain motion coordinates in the motion of the human body. 8.根据权利要求6所述的人体运动意图的识别系统,其特征在于,所述关注点坐标确定模块具体包括:8. The recognition system of human body motion intention according to claim 6, wherein the coordinate determination module of the point of interest specifically comprises: 第二视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人眼瞳孔的视频信息;a second video information acquisition unit, configured to acquire video information of the pupil of the human eye through the wearable eye-tracker; 关注点坐标确定单元,用于根据所述人眼瞳孔的视频信息,确定人体运动中视线关注点的坐标。The attention point coordinate determination unit is used for determining the coordinates of the line of sight attention point in the motion of the human body according to the video information of the pupil of the human eye. 9.根据权利要求6所述的人体运动意图的识别系统,其特征在于,所述障碍物坐标识别模块具体包括:9. The recognition system of human motion intention according to claim 6, wherein the obstacle coordinate recognition module specifically comprises: 第三视频信息获取单元,用于通过所述可穿戴的眼镜式眼动仪获取人体运动中视野场景视频信息;a third video information acquisition unit, configured to acquire video information of visual field scenes in human motion through the wearable glasses eye tracker; 障碍物识别单元,用于通过训练好的卷积神经网络模型对所述视野场景视频信息中每一帧图像进行识别,确定图像中的障碍物;The obstacle identification unit is used to identify each frame of image in the video information of the visual field through the trained convolutional neural network model, and determine the obstacle in the image; 坐标确定单元,用于通过矩形框对每一帧图像中的障碍物进行框选,得到图像中的障碍物的坐标。The coordinate determination unit is used to frame the obstacles in each frame of images through a rectangular frame to obtain the coordinates of the obstacles in the image. 10.根据权利要求6所述的人体运动意图的识别系统,其特征在于,还包括:10. The recognition system of human motion intention according to claim 6, characterized in that, further comprising: 插值模块,用于采用样条插值方法将所述运动坐标、关注点坐标以及所述障碍物坐标同步成等间隔的信号。The interpolation module is used for synchronizing the motion coordinates, the attention point coordinates and the obstacle coordinates into signals with equal intervals by using a spline interpolation method.
CN202010500206.6A 2020-06-04 2020-06-04 A method and system for recognizing human motion intention Pending CN111652155A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010500206.6A CN111652155A (en) 2020-06-04 2020-06-04 A method and system for recognizing human motion intention

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010500206.6A CN111652155A (en) 2020-06-04 2020-06-04 A method and system for recognizing human motion intention

Publications (1)

Publication Number Publication Date
CN111652155A true CN111652155A (en) 2020-09-11

Family

ID=72347230

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010500206.6A Pending CN111652155A (en) 2020-06-04 2020-06-04 A method and system for recognizing human motion intention

Country Status (1)

Country Link
CN (1) CN111652155A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111986781A (en) * 2020-08-24 2020-11-24 龙马智芯(珠海横琴)科技有限公司 Psychotherapy method, device and user terminal based on human-computer interaction
CN114037738A (en) * 2021-11-18 2022-02-11 湖北三闾智能科技有限公司 Human vision-driven upper limb auxiliary robot control method
US20220405954A1 (en) * 2021-06-15 2022-12-22 Acronis International Gmbh Systems and methods for determining environment dimensions based on environment pose

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150309563A1 (en) * 2013-09-17 2015-10-29 Medibotics Llc Motion Recognition Clothing [TM] with Flexible Electromagnetic, Light, or Sonic Energy Pathways
US9727790B1 (en) * 2010-06-04 2017-08-08 Masoud Vaziri Method and apparatus for a wearable computer with natural user interface
CN109345542A (en) * 2018-09-18 2019-02-15 重庆大学 A wearable visual gaze target positioning device and method
CN109782902A (en) * 2018-12-17 2019-05-21 中国科学院深圳先进技术研究院 A kind of operation indicating method and glasses
CN110095116A (en) * 2019-04-29 2019-08-06 桂林电子科技大学 A kind of localization method of vision positioning and inertial navigation combination based on LIFT
CN110125909A (en) * 2019-05-22 2019-08-16 南京师范大学镇江创新发展研究院 A kind of multi-information fusion human body exoskeleton robot Control protection system
US20200167957A1 (en) * 2018-11-28 2020-05-28 National Taiwan University Correcting method and device for eye-tracking

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9727790B1 (en) * 2010-06-04 2017-08-08 Masoud Vaziri Method and apparatus for a wearable computer with natural user interface
US20150309563A1 (en) * 2013-09-17 2015-10-29 Medibotics Llc Motion Recognition Clothing [TM] with Flexible Electromagnetic, Light, or Sonic Energy Pathways
CN109345542A (en) * 2018-09-18 2019-02-15 重庆大学 A wearable visual gaze target positioning device and method
US20200167957A1 (en) * 2018-11-28 2020-05-28 National Taiwan University Correcting method and device for eye-tracking
CN109782902A (en) * 2018-12-17 2019-05-21 中国科学院深圳先进技术研究院 A kind of operation indicating method and glasses
CN110095116A (en) * 2019-04-29 2019-08-06 桂林电子科技大学 A kind of localization method of vision positioning and inertial navigation combination based on LIFT
CN110125909A (en) * 2019-05-22 2019-08-16 南京师范大学镇江创新发展研究院 A kind of multi-information fusion human body exoskeleton robot Control protection system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
柳锴: "基于人体上肢协同运动特征的外骨骼机器人设计方法研究", 《中国优秀博硕士学位论文全文数据库(博士) 信息科技辑》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111986781A (en) * 2020-08-24 2020-11-24 龙马智芯(珠海横琴)科技有限公司 Psychotherapy method, device and user terminal based on human-computer interaction
CN111986781B (en) * 2020-08-24 2021-08-06 龙马智芯(珠海横琴)科技有限公司 Psychotherapy device and user terminal based on human-computer interaction
US20220405954A1 (en) * 2021-06-15 2022-12-22 Acronis International Gmbh Systems and methods for determining environment dimensions based on environment pose
US12165352B2 (en) * 2021-06-15 2024-12-10 Acronis International Gmbh Systems and methods for determining environment dimensions based on environment pose
CN114037738A (en) * 2021-11-18 2022-02-11 湖北三闾智能科技有限公司 Human vision-driven upper limb auxiliary robot control method
CN114037738B (en) * 2021-11-18 2025-01-17 湖北三闾智能科技有限公司 Control method of human vision driven upper limb auxiliary robot

Similar Documents

Publication Publication Date Title
Rabhi et al. Intelligent control wheelchair using a new visual joystick
Al-Rahayfeh et al. Eye tracking and head movement detection: A state-of-art survey
CN113362452B (en) Hand posture three-dimensional reconstruction method and device and storage medium
CN106529409B (en) A Method for Measuring Eye Gaze Angle Based on Head Posture
CN111652155A (en) A method and system for recognizing human motion intention
CN109559332B (en) A Gaze Tracking Method Combining Bidirectional LSTM and Itracker
Sun et al. Real-time gaze estimation with online calibration
Tao et al. Trajectory planning of upper limb rehabilitation robot based on human pose estimation
Liu et al. Real-time robust stereo visual SLAM system based on bionic eyes
Lemley et al. Eye tracking in augmented spaces: A deep learning approach
Liu et al. Ego+ x: An egocentric vision system for global 3d human pose estimation and social interaction characterization
Li et al. Learning laparoscope actions via video features for proactive robotic field-of-view control
Amer et al. Wheelchair control system based eye gaze
Zhao et al. Reducing the device complexity for 3D human pose estimation: A deep learning approach using monocular camera and IMUs
Jin et al. Human-robot interaction for assisted object grasping by a wearable robotic object manipulation aid for the blind
Perez et al. Robotic wheelchair controlled through a vision-based interface
Wu et al. The visual footsteps planning system for exoskeleton robots under complex terrain
Orlov et al. A gaze-contingent intention decoding engine for human augmentation
Chhimpa et al. Revolutionizing Gaze-based Human-Computer Interaction using Iris Tracking: A Webcam-Based Low-Cost Approach with Calibration, Regression and Real-Time Re-calibration
CN115120250A (en) Intelligent brain-controlled wheelchair system based on electroencephalogram signals and SLAM control
Chee et al. Eye tracking electronic wheelchair for physically challenged person
Hassani et al. Implementation of wheelchair controller using mouth and tongue gesture
Kim et al. A human-robot interface using eye-gaze tracking system for people with motor disabilities
Sowmya et al. Eye gaze controlled wheelchair
CN108008811A (en) A kind of method and terminal using non-touch screen mode operating terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200911