CN108992072B - A gait event-based human lower limb movement intention recognition method - Google Patents
A gait event-based human lower limb movement intention recognition method Download PDFInfo
- Publication number
- CN108992072B CN108992072B CN201810915082.0A CN201810915082A CN108992072B CN 108992072 B CN108992072 B CN 108992072B CN 201810915082 A CN201810915082 A CN 201810915082A CN 108992072 B CN108992072 B CN 108992072B
- Authority
- CN
- China
- Prior art keywords
- gait
- stride
- event
- time
- events
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Measuring devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor or mobility of a limb
- A61B5/112—Gait analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6813—Specially adapted to be attached to a specific body part
- A61B5/6828—Leg
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Physics & Mathematics (AREA)
- Public Health (AREA)
- Pathology (AREA)
- Physiology (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Signal Processing (AREA)
- Dentistry (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Rehabilitation Tools (AREA)
Abstract
Description
技术领域technical field
本发明涉及一种基于步态事件的人体下肢运动意图识别方法。The invention relates to a method for recognizing human lower limb movement intention based on gait events.
背景技术Background technique
随着机电技术的不断发展,机器人系统的研究越来越成为现阶段的热门课题,其中一种可穿戴式的下肢外骨骼机器人已经取得了阶段性的进展,并在军事和医疗领域有着广泛的应用。With the continuous development of electromechanical technology, the research of robotic systems has become a hot topic at this stage. One of the wearable lower limb exoskeleton robots has made staged progress and has a wide range of applications in the military and medical fields. application.
外骨骼机器人识别人体下肢的运动意图则是实现外骨骼机器人智能化的关键技术之一。而现阶段主要实现下肢运动意图识别的方法主要有两种,一种是利用人体电信号来识别人体运动意图,例如肌电(EMG)或脑电波(EOG);一种是检测人机交互力来识别人体运动意图。Exoskeleton robot recognition of the motion intention of human lower limbs is one of the key technologies to realize the intelligentization of exoskeleton robots. At this stage, there are mainly two methods to realize lower limb motion intention recognition. One is to use human body electrical signals to identify human body motion intentions, such as electromyography (EMG) or brain waves (EOG); the other is to detect human-computer interaction force. to identify human motion intentions.
发明内容SUMMARY OF THE INVENTION
本发明针对现有技术的不足,提供了一种基于步态事件的人体下肢运动意图识别方法。Aiming at the deficiencies of the prior art, the present invention provides a method for recognizing human lower limb movement intention based on gait events.
本发明首先检测人体下肢实时步态参量(步频与步幅),然后根据实时的步频与步幅来调整标准转角轨迹,进而得到同步人体步频与步幅的关节转角控制目标轨迹,最后该目标轨迹作为控制系统的跟踪目标轨迹进行运动控制,实现下肢外骨骼同步人体步态进行运动,即实现了人体下肢运动意图的识别。The invention firstly detects the real-time gait parameters (step frequency and stride length) of the lower limbs of the human body, then adjusts the standard rotation angle trajectory according to the real-time cadence frequency and stride length, and then obtains the joint rotation angle control target trajectory synchronizing the human body's cadence frequency and stride length, and finally The target trajectory is used as the tracking target trajectory of the control system for motion control, so that the lower limb exoskeleton can move synchronously with the human body's gait, that is, the recognition of the motion intention of the lower limbs of the human body is realized.
本发明技术方案的具体内容如下:The specific content of the technical solution of the present invention is as follows:
(1)人体下肢步态事件定义。(1) Definition of human lower limb gait events.
(2)步态事件实时检测算法。(2) Real-time detection algorithm of gait events.
(3)基于步态事件得到实时步频与步幅。(3) Real-time stride frequency and stride length are obtained based on gait events.
(4)根据实时步频与步幅调整标准关节转角轨迹,并生成同步步频与步幅的参考关节转角轨迹。(4) Adjust the standard joint angle trajectory according to the real-time cadence and stride, and generate the reference joint angle trajectory of the synchronized cadence and stride.
本发明的有益效果是:提出一种全新的下肢运动意图识别的方法,能够通过安装在下肢的惯性检测单元(IMU)实时获取关节转角数据,并根据所测得的数据得到同步步频与步幅的关节转角参考轨迹作为运动控制的跟踪轨迹,最终实现了下肢外骨骼同步人体步态。The beneficial effects of the present invention are as follows: a brand-new lower limb motion intention recognition method is proposed, which can acquire joint rotation angle data in real time through an inertial detection unit (IMU) installed on the lower limb, and obtain synchronous cadence and cadence according to the measured data. The joint rotation angle reference trajectory of the width is used as the tracking trajectory of the motion control, and finally the human gait synchronization of the lower limb exoskeleton is realized.
附图说明Description of drawings
图1是整体方法框图;Figure 1 is a block diagram of the overall method;
图2是步态事件定义图。Figure 2 is a gait event definition diagram.
具体实施方式Detailed ways
本发明的整体方案主要分为3步,其中第一步为实时步态事件检测,第二步为实时步频与步幅的检测,最后一步为调整标准关节转角轨迹曲线并生成参考转角轨迹。具体如图1所示。The overall scheme of the present invention is mainly divided into three steps, wherein the first step is real-time gait event detection, the second step is real-time gait frequency and stride detection, and the last step is to adjust the standard joint corner trajectory curve and generate a reference corner trajectory. Specifically as shown in Figure 1.
(1)步态事件检测(1) Gait event detection
本发明将一个完整的步态周期分为4个步态相位,其中包括支撑初期(ISt)、支撑中期(MSt)、支撑末期(TSt)和摆动期(Sw)。定义每个步态相位的开始为步态事件,其中包括接触开始事件(IC)、大腿直立事件(TE)、脚趾预离地事件(pTO)和脚趾离地事件(TO),具体如图2所示。The present invention divides a complete gait cycle into four gait phases, including the initial stance (ISt), mid-stance (MSt), end-stance (TSt) and swing phase (Sw). The start of each gait phase is defined as a gait event, including the contact initiation event (IC), the thigh erection event (TE), the toe pre-off event (pTO) and the toe off the ground event (TO), as shown in Figure 2 shown.
根据步态事件的定义和实验结果可以发现步态事件与下肢髋关节的转角存在着一定的映射关系,具体规律如表1所示。According to the definition of gait events and the experimental results, it can be found that there is a certain mapping relationship between gait events and the rotation angle of the lower extremity hip joint. The specific rules are shown in Table 1.
表1是步态事件与髋关节转角规律表Table 1 is the regularity of gait events and hip joint rotation
数据采集卡采集人体下肢的IMU实时髋关节转角数据,并根据上述规律编写程序便可实时检测出每个步态周期中每个步态事件。The data acquisition card collects the IMU real-time hip joint angle data of the lower limbs of the human body, and writes a program according to the above rules to detect each gait event in each gait cycle in real time.
(2)步频和步幅算法(2) Step frequency and stride length algorithm
2.1步频算法2.1 Step frequency algorithm
根据步态事件检测的结果,可以得到相邻的同一步态事件的时间间隔,然后可以推出步态周期Ts计算公式如下:According to the result of gait event detection, the time interval of adjacent same gait events can be obtained, and then the calculation formula of gait cycle T s can be derived as follows:
式中tn和tn-1分别表示第n次和第n-1次该步态事件发生的时间点。为了提升步态周期计算的精准度,将步态周期中定义的4个步态事件都用来计算步态周期并取平均值,具体公式如下:where t n and t n-1 represent the time points of the nth and n-1th gait events, respectively. In order to improve the accuracy of gait cycle calculation, the four gait events defined in the gait cycle are used to calculate the gait cycle and take the average value. The specific formula is as follows:
式中和分别表示第n次和第n-1次第i个步态事件发生的时间点,ωs表示步态频率,k等于4。得到了步态周期后,再定义步态周期的相位角为:in the formula and represent the time points of the nth and n-1th ith gait events, respectively, ωs represents the gait frequency, and k is equal to 4. After obtaining the gait cycle, define the phase angle of the gait cycle as:
其中φs表示步态相位角函数,取值范围为0到2π,然后根据相位角定义步态进程百分比s为:where φ s represents the gait phase angle function, which ranges from 0 to 2π, and then defines the gait progress percentage s according to the phase angle as:
2.2步幅算法2.2 Stride algorithm
在本设计中定义步幅为接触开始事件或脚趾预离地事件发生时两脚质心之间的距离。根据上述定义和下肢简化模型,可以得到步幅的计算公式如下:In this design, the stride is defined as the distance between the center of mass of the feet when the contact start event or the toe pre-ground event occurs. According to the above definition and the simplified model of lower limbs, the calculation formula of stride length can be obtained as follows:
L=l3sinθ3+l4sinθ4-(l1sinθ1+l2sinθ2)L=l 3 sinθ 3 +l 4 sinθ 4 -(l 1 sinθ 1 +l 2 sinθ 2 )
式中θ1和θ2表示膝关节转角,θ3和θ4表示髋关节转角,在一个完整的步态周期中,存在着两个步态事件可以用来计算步幅,为了提高计算结果的精度,可将步幅计算公式表示如下:where θ 1 and θ 2 represent the knee joint rotation angle, θ 3 and θ 4 represent the hip joint rotation angle, In a complete gait cycle, there are two gait events that can be used to calculate the stride length. In order to improve the accuracy of the calculation results, the stride length calculation formula can be expressed as follows:
式中和表示当IC(i=1)和pTO(i=2)步态事件发生时髋关节转角值,和表示当IC(i=1)和pTO(i=2)步态事件发生时膝关节转角值。in the formula and represents the value of hip joint rotation when IC(i=1) and pTO(i=2) gait events occur, and Indicates the knee joint rotation value when IC (i=1) and pTO (i=2) gait events occur.
(3)关节转角轨迹调整算法(3) Joint rotation angle trajectory adjustment algorithm
外骨骼参考关节跟踪轨迹是基于标准关节轨迹生成的,具体包括髋关节和膝关节。而标准关节转角轨迹则是通过参考临床步态分析数据库(CGA)或者实验结果得到的。为了实现外骨骼机器人同步下肢运动意图,标准关节转角轨迹需要根据实时检测到的步频和步幅来调整周期和幅值。Exoskeleton reference joint tracking trajectories are generated based on standard joint trajectories, including hip and knee joints. The standard joint rotation trajectory is obtained by referring to the clinical gait analysis database (CGA) or experimental results. In order to realize the synchronized lower limb motion intention of the exoskeleton robot, the standard joint angle trajectory needs to adjust the period and amplitude according to the stride frequency and stride detected in real time.
3.1步频调整算法3.1 Step frequency adjustment algorithm
假设标准髋关节和膝关节转角曲线函数分别为fh(s)和fk(s),s变量表示步态进程百分比,s是时变参数,具体定义如下:Assuming that the standard hip and knee rotation curve functions are f h (s) and f k (s), respectively, the s variable represents the percentage of gait progress, and s is a time-varying parameter, which is specifically defined as follows:
式中φs(t)表示步态相位角函数,可在步频检测算法中得到。进而可以推出同步步频的关节转角函数如下:where φ s (t) represents the gait phase angle function, which can be obtained in the gait detection algorithm. Then, the joint angle function of the synchronous cadence can be derived as follows:
fh(t)=fh(s(t))f h (t)=f h (s(t))
fk(t)=fk(s(t))f k (t)=f k (s(t))
3.2步幅调整算法3.2 Stride adjustment algorithm
假设标准步态的步幅为Ls,步幅检测的结果为Lr,进而可以定义步幅调整参数K为:Assuming that the stride of the standard gait is L s , and the result of stride detection is L r , the stride adjustment parameter K can be defined as:
根据步态事件检测结果,可以得到TO事件发生时步态进程百分比值,具体计算公式为:According to the detection result of gait event, the percentage value of gait process when TO event occurs can be obtained. The specific calculation formula is:
式中tTO表示TO事件发生的时间点,tTE表示TE事件发生的时间点,Ts表示步态周期。然后调整标准轨迹函数如下式:In the formula, t TO represents the time point of the TO event, t TE represents the time point of the TE event, and T s represents the gait cycle. Then adjust the standard trajectory function as follows:
式中f(s)表示标准关节转角函数,Θ1为步幅拉伸区间,Θ1∈(0,sTO),Θ2为步幅收缩区间,Θ2∈(sTO,100%)。最后应用到步频调整算法中,可以推出公式为:where f(s) represents the standard joint angle function, Θ 1 is the stride stretch interval, Θ 1 ∈(0,s TO ), Θ 2 is the stride contraction interval, Θ 2 ∈(s TO ,100%). Finally, when applied to the step frequency adjustment algorithm, the formula can be derived as:
fn(t)=fn(s(t)) fn (t)= fn (s(t))
式中fn(t)表示参考转角轨迹函数,即同步人体步频和步幅的转角轨迹曲线。where f n (t) represents the reference corner trajectory function, that is, the corner trajectory curve that synchronizes the human cadence and stride length.
综上,本发明设计了一种基于步态事件检测的人体意图识别方法。该方法用到了4个惯性检测单元(IMU)来实时检测人体下肢关节转角(髋关节和膝关节转角),该传感器在价格上相对较低且安装方便,适合机器人产业化发展的目标。在算法的复杂度上,本发明设计的算法也相对较为简单,降低了计算量。To sum up, the present invention designs a human intention recognition method based on gait event detection. This method uses four inertial detection units (IMUs) to detect the joint rotation angles of the lower limbs (hip and knee joints) in real time. In terms of the complexity of the algorithm, the algorithm designed by the present invention is also relatively simple, which reduces the amount of calculation.
Claims (3)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810915082.0A CN108992072B (en) | 2018-08-13 | 2018-08-13 | A gait event-based human lower limb movement intention recognition method |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810915082.0A CN108992072B (en) | 2018-08-13 | 2018-08-13 | A gait event-based human lower limb movement intention recognition method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN108992072A CN108992072A (en) | 2018-12-14 |
| CN108992072B true CN108992072B (en) | 2020-09-11 |
Family
ID=64595801
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201810915082.0A Active CN108992072B (en) | 2018-08-13 | 2018-08-13 | A gait event-based human lower limb movement intention recognition method |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN108992072B (en) |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111469117B (en) * | 2020-04-14 | 2022-06-03 | 武汉理工大学 | Human motion mode detection method of rigid-flexible coupling active exoskeleton |
| CN113425290A (en) * | 2021-06-15 | 2021-09-24 | 燕山大学 | Joint coupling time sequence calculation method for human body rhythm movement |
| CN114569411B (en) * | 2022-02-21 | 2024-10-08 | 长沙优龙机器人有限公司 | Gait self-adaptive control method and system for hemiplegic patient |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104302251A (en) * | 2012-03-22 | 2015-01-21 | 埃克苏仿生公司 | Human machine interface for lower extremity orthotics |
| CN105992554A (en) * | 2013-12-09 | 2016-10-05 | 哈佛大学校长及研究员协会 | Assistive flexible suits, flexible suit systems, and methods of manufacture and control thereof for assisting human mobility |
| CN107361992A (en) * | 2016-05-13 | 2017-11-21 | 深圳市肯綮科技有限公司 | A kind of human body lower limbs move power assisting device |
-
2018
- 2018-08-13 CN CN201810915082.0A patent/CN108992072B/en active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104302251A (en) * | 2012-03-22 | 2015-01-21 | 埃克苏仿生公司 | Human machine interface for lower extremity orthotics |
| CN105992554A (en) * | 2013-12-09 | 2016-10-05 | 哈佛大学校长及研究员协会 | Assistive flexible suits, flexible suit systems, and methods of manufacture and control thereof for assisting human mobility |
| CN107361992A (en) * | 2016-05-13 | 2017-11-21 | 深圳市肯綮科技有限公司 | A kind of human body lower limbs move power assisting device |
Non-Patent Citations (1)
| Title |
|---|
| Gait-Event-Based Synchronization Method for Gait Rehabilitation Robots via a Bioinspired Adaptive Oscillator;Gong Chen等;《IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING》;20170630;第64卷(第6期);第1345-1356页 * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN108992072A (en) | 2018-12-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Schmitz et al. | The measurement of in vivo joint angles during a squat using a single camera markerless motion capture system as compared to a marker based system | |
| Joshi et al. | Classification of gait phases from lower limb EMG: Application to exoskeleton orthosis | |
| CN106067178B (en) | A kind of continuous estimation method of hand joint movement based on muscle synergistic activation model | |
| Chen et al. | Gait-event-based synchronization method for gait rehabilitation robots via a bioinspired adaptive oscillator | |
| CN107273611B (en) | A gait planning method for a lower limb rehabilitation robot based on the walking characteristics of the lower limbs | |
| CN108992072B (en) | A gait event-based human lower limb movement intention recognition method | |
| CN108960155B (en) | Adult gait extraction and abnormal analysis method based on Kinect | |
| Li et al. | A human joint torque estimation method for elbow exoskeleton control | |
| Chen et al. | Adaptive control strategy for gait rehabilitation robot to assist-when-needed | |
| Ling et al. | AE-CNN-based multisource data fusion for gait motion step length estimation | |
| Al-Quraishi et al. | Impact of feature extraction techniques on classification accuracy for EMG based ankle joint movements | |
| Bae et al. | Real-time estimation of lower extremity joint torques in normal gait | |
| Wen et al. | Gait measurement and quantitative analysis in patients with Parkinson's disease for rehabilitation assessment | |
| Lu et al. | Gait-Event-Based human intention recognition approach for lower limb | |
| Tong et al. | BP-AR-based human joint angle estimation using multi-channel sEMG | |
| Kapti et al. | Wearable acceleration sensor application in unilateral trans-tibial amputation prostheses | |
| Meng et al. | Effect of walking variations on complementary filter based inertial data fusion for ankle angle measurement | |
| Lyu et al. | Comparison of particle filter to established filtering methods in electromyography biofeedback | |
| Su et al. | Process mining over sensor data: Goal recognition for powered transhumeral prostheses | |
| Mamikoglu et al. | Electromyography based joint angle estimation and control of a robotic leg | |
| Chen et al. | An adaptive gait learning strategy for lower limb exoskeleton robot | |
| Hao et al. | A gait patterns recognition approach based on surface electromyography and three-axis acceleration signals | |
| Shen et al. | Simultaneous gait event intention detection using single semg sensor for lower limb exoskeleton | |
| Shao et al. | A method for self-service rehabilitation training of human lower limbs | |
| Bao et al. | Motion identification based on sEMG for flexible pneumatic hand rehabilitator |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |






























