[go: up one dir, main page]

CN108992072B - A gait event-based human lower limb movement intention recognition method - Google Patents

A gait event-based human lower limb movement intention recognition method Download PDF

Info

Publication number
CN108992072B
CN108992072B CN201810915082.0A CN201810915082A CN108992072B CN 108992072 B CN108992072 B CN 108992072B CN 201810915082 A CN201810915082 A CN 201810915082A CN 108992072 B CN108992072 B CN 108992072B
Authority
CN
China
Prior art keywords
gait
stride
event
time
events
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810915082.0A
Other languages
Chinese (zh)
Other versions
CN108992072A (en
Inventor
陈正
卢形
黄方昊
周时钊
朱世强
梅珑
金来
郭凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Sanlian Robot Technology Co ltd
Zhejiang University ZJU
Original Assignee
Anhui Sanlian Robot Technology Co ltd
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Sanlian Robot Technology Co ltd, Zhejiang University ZJU filed Critical Anhui Sanlian Robot Technology Co ltd
Priority to CN201810915082.0A priority Critical patent/CN108992072B/en
Publication of CN108992072A publication Critical patent/CN108992072A/en
Application granted granted Critical
Publication of CN108992072B publication Critical patent/CN108992072B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Measuring devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor or mobility of a limb
    • A61B5/112Gait analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6802Sensor mounted on worn items
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6813Specially adapted to be attached to a specific body part
    • A61B5/6828Leg
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Physics & Mathematics (AREA)
  • Public Health (AREA)
  • Pathology (AREA)
  • Physiology (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Signal Processing (AREA)
  • Dentistry (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
  • Rehabilitation Tools (AREA)

Abstract

The invention discloses a human body intention identification method based on gait event detection. The method comprises the steps of firstly defining the gait event definition of the lower limbs of the human body, then detecting the gait event in real time, then obtaining the real-time step frequency and the stride based on the gait event, finally adjusting the standard joint corner track according to the real-time step frequency and the stride, and generating the reference joint corner track for synchronizing the step frequency and the stride. According to the invention, joint corner data can be acquired in real time through an inertia detection unit (IMU) installed on the lower limb, and a joint corner reference track for synchronizing the step frequency and the stride is obtained according to the acquired data and is used as a tracking track for motion control, so that the synchronous human gait of the lower limb exoskeleton is finally realized.

Description

一种基于步态事件的人体下肢运动意图识别方法A gait event-based human lower limb movement intention recognition method

技术领域technical field

本发明涉及一种基于步态事件的人体下肢运动意图识别方法。The invention relates to a method for recognizing human lower limb movement intention based on gait events.

背景技术Background technique

随着机电技术的不断发展,机器人系统的研究越来越成为现阶段的热门课题,其中一种可穿戴式的下肢外骨骼机器人已经取得了阶段性的进展,并在军事和医疗领域有着广泛的应用。With the continuous development of electromechanical technology, the research of robotic systems has become a hot topic at this stage. One of the wearable lower limb exoskeleton robots has made staged progress and has a wide range of applications in the military and medical fields. application.

外骨骼机器人识别人体下肢的运动意图则是实现外骨骼机器人智能化的关键技术之一。而现阶段主要实现下肢运动意图识别的方法主要有两种,一种是利用人体电信号来识别人体运动意图,例如肌电(EMG)或脑电波(EOG);一种是检测人机交互力来识别人体运动意图。Exoskeleton robot recognition of the motion intention of human lower limbs is one of the key technologies to realize the intelligentization of exoskeleton robots. At this stage, there are mainly two methods to realize lower limb motion intention recognition. One is to use human body electrical signals to identify human body motion intentions, such as electromyography (EMG) or brain waves (EOG); the other is to detect human-computer interaction force. to identify human motion intentions.

发明内容SUMMARY OF THE INVENTION

本发明针对现有技术的不足,提供了一种基于步态事件的人体下肢运动意图识别方法。Aiming at the deficiencies of the prior art, the present invention provides a method for recognizing human lower limb movement intention based on gait events.

本发明首先检测人体下肢实时步态参量(步频与步幅),然后根据实时的步频与步幅来调整标准转角轨迹,进而得到同步人体步频与步幅的关节转角控制目标轨迹,最后该目标轨迹作为控制系统的跟踪目标轨迹进行运动控制,实现下肢外骨骼同步人体步态进行运动,即实现了人体下肢运动意图的识别。The invention firstly detects the real-time gait parameters (step frequency and stride length) of the lower limbs of the human body, then adjusts the standard rotation angle trajectory according to the real-time cadence frequency and stride length, and then obtains the joint rotation angle control target trajectory synchronizing the human body's cadence frequency and stride length, and finally The target trajectory is used as the tracking target trajectory of the control system for motion control, so that the lower limb exoskeleton can move synchronously with the human body's gait, that is, the recognition of the motion intention of the lower limbs of the human body is realized.

本发明技术方案的具体内容如下:The specific content of the technical solution of the present invention is as follows:

(1)人体下肢步态事件定义。(1) Definition of human lower limb gait events.

(2)步态事件实时检测算法。(2) Real-time detection algorithm of gait events.

(3)基于步态事件得到实时步频与步幅。(3) Real-time stride frequency and stride length are obtained based on gait events.

(4)根据实时步频与步幅调整标准关节转角轨迹,并生成同步步频与步幅的参考关节转角轨迹。(4) Adjust the standard joint angle trajectory according to the real-time cadence and stride, and generate the reference joint angle trajectory of the synchronized cadence and stride.

本发明的有益效果是:提出一种全新的下肢运动意图识别的方法,能够通过安装在下肢的惯性检测单元(IMU)实时获取关节转角数据,并根据所测得的数据得到同步步频与步幅的关节转角参考轨迹作为运动控制的跟踪轨迹,最终实现了下肢外骨骼同步人体步态。The beneficial effects of the present invention are as follows: a brand-new lower limb motion intention recognition method is proposed, which can acquire joint rotation angle data in real time through an inertial detection unit (IMU) installed on the lower limb, and obtain synchronous cadence and cadence according to the measured data. The joint rotation angle reference trajectory of the width is used as the tracking trajectory of the motion control, and finally the human gait synchronization of the lower limb exoskeleton is realized.

附图说明Description of drawings

图1是整体方法框图;Figure 1 is a block diagram of the overall method;

图2是步态事件定义图。Figure 2 is a gait event definition diagram.

具体实施方式Detailed ways

本发明的整体方案主要分为3步,其中第一步为实时步态事件检测,第二步为实时步频与步幅的检测,最后一步为调整标准关节转角轨迹曲线并生成参考转角轨迹。具体如图1所示。The overall scheme of the present invention is mainly divided into three steps, wherein the first step is real-time gait event detection, the second step is real-time gait frequency and stride detection, and the last step is to adjust the standard joint corner trajectory curve and generate a reference corner trajectory. Specifically as shown in Figure 1.

(1)步态事件检测(1) Gait event detection

本发明将一个完整的步态周期分为4个步态相位,其中包括支撑初期(ISt)、支撑中期(MSt)、支撑末期(TSt)和摆动期(Sw)。定义每个步态相位的开始为步态事件,其中包括接触开始事件(IC)、大腿直立事件(TE)、脚趾预离地事件(pTO)和脚趾离地事件(TO),具体如图2所示。The present invention divides a complete gait cycle into four gait phases, including the initial stance (ISt), mid-stance (MSt), end-stance (TSt) and swing phase (Sw). The start of each gait phase is defined as a gait event, including the contact initiation event (IC), the thigh erection event (TE), the toe pre-off event (pTO) and the toe off the ground event (TO), as shown in Figure 2 shown.

根据步态事件的定义和实验结果可以发现步态事件与下肢髋关节的转角存在着一定的映射关系,具体规律如表1所示。According to the definition of gait events and the experimental results, it can be found that there is a certain mapping relationship between gait events and the rotation angle of the lower extremity hip joint. The specific rules are shown in Table 1.

表1是步态事件与髋关节转角规律表Table 1 is the regularity of gait events and hip joint rotation

步态事件Gait events ICIC TETE pTOpTO TOTO 髋关节转角Hip Rotation 最大值maximum value 大约为零about zero 最小值minimum 大约为零about zero

数据采集卡采集人体下肢的IMU实时髋关节转角数据,并根据上述规律编写程序便可实时检测出每个步态周期中每个步态事件。The data acquisition card collects the IMU real-time hip joint angle data of the lower limbs of the human body, and writes a program according to the above rules to detect each gait event in each gait cycle in real time.

(2)步频和步幅算法(2) Step frequency and stride length algorithm

2.1步频算法2.1 Step frequency algorithm

根据步态事件检测的结果,可以得到相邻的同一步态事件的时间间隔,然后可以推出步态周期Ts计算公式如下:According to the result of gait event detection, the time interval of adjacent same gait events can be obtained, and then the calculation formula of gait cycle T s can be derived as follows:

Figure BDA0001762770260000031
Figure BDA0001762770260000031

式中tn和tn-1分别表示第n次和第n-1次该步态事件发生的时间点。为了提升步态周期计算的精准度,将步态周期中定义的4个步态事件都用来计算步态周期并取平均值,具体公式如下:where t n and t n-1 represent the time points of the nth and n-1th gait events, respectively. In order to improve the accuracy of gait cycle calculation, the four gait events defined in the gait cycle are used to calculate the gait cycle and take the average value. The specific formula is as follows:

Figure BDA0001762770260000032
Figure BDA0001762770260000032

式中

Figure BDA0001762770260000033
Figure BDA0001762770260000034
分别表示第n次和第n-1次第i个步态事件发生的时间点,ωs表示步态频率,k等于4。得到了步态周期后,再定义步态周期的相位角为:in the formula
Figure BDA0001762770260000033
and
Figure BDA0001762770260000034
represent the time points of the nth and n-1th ith gait events, respectively, ωs represents the gait frequency, and k is equal to 4. After obtaining the gait cycle, define the phase angle of the gait cycle as:

Figure BDA0001762770260000035
Figure BDA0001762770260000035

其中φs表示步态相位角函数,取值范围为0到2π,然后根据相位角定义步态进程百分比s为:where φ s represents the gait phase angle function, which ranges from 0 to 2π, and then defines the gait progress percentage s according to the phase angle as:

Figure BDA0001762770260000041
Figure BDA0001762770260000041

2.2步幅算法2.2 Stride algorithm

在本设计中定义步幅为接触开始事件或脚趾预离地事件发生时两脚质心之间的距离。根据上述定义和下肢简化模型,可以得到步幅的计算公式如下:In this design, the stride is defined as the distance between the center of mass of the feet when the contact start event or the toe pre-ground event occurs. According to the above definition and the simplified model of lower limbs, the calculation formula of stride length can be obtained as follows:

L=l3sinθ3+l4sinθ4-(l1sinθ1+l2sinθ2)L=l 3 sinθ 3 +l 4 sinθ 4 -(l 1 sinθ 1 +l 2 sinθ 2 )

式中θ1和θ2表示膝关节转角,θ3和θ4表示髋关节转角,

Figure BDA0001762770260000042
在一个完整的步态周期中,存在着两个步态事件可以用来计算步幅,为了提高计算结果的精度,可将步幅计算公式表示如下:where θ 1 and θ 2 represent the knee joint rotation angle, θ 3 and θ 4 represent the hip joint rotation angle,
Figure BDA0001762770260000042
In a complete gait cycle, there are two gait events that can be used to calculate the stride length. In order to improve the accuracy of the calculation results, the stride length calculation formula can be expressed as follows:

Figure BDA0001762770260000043
Figure BDA0001762770260000043

式中

Figure BDA0001762770260000044
Figure BDA0001762770260000045
表示当IC(i=1)和pTO(i=2)步态事件发生时髋关节转角值,
Figure BDA0001762770260000046
Figure BDA0001762770260000047
表示当IC(i=1)和pTO(i=2)步态事件发生时膝关节转角值。in the formula
Figure BDA0001762770260000044
and
Figure BDA0001762770260000045
represents the value of hip joint rotation when IC(i=1) and pTO(i=2) gait events occur,
Figure BDA0001762770260000046
and
Figure BDA0001762770260000047
Indicates the knee joint rotation value when IC (i=1) and pTO (i=2) gait events occur.

(3)关节转角轨迹调整算法(3) Joint rotation angle trajectory adjustment algorithm

外骨骼参考关节跟踪轨迹是基于标准关节轨迹生成的,具体包括髋关节和膝关节。而标准关节转角轨迹则是通过参考临床步态分析数据库(CGA)或者实验结果得到的。为了实现外骨骼机器人同步下肢运动意图,标准关节转角轨迹需要根据实时检测到的步频和步幅来调整周期和幅值。Exoskeleton reference joint tracking trajectories are generated based on standard joint trajectories, including hip and knee joints. The standard joint rotation trajectory is obtained by referring to the clinical gait analysis database (CGA) or experimental results. In order to realize the synchronized lower limb motion intention of the exoskeleton robot, the standard joint angle trajectory needs to adjust the period and amplitude according to the stride frequency and stride detected in real time.

3.1步频调整算法3.1 Step frequency adjustment algorithm

假设标准髋关节和膝关节转角曲线函数分别为fh(s)和fk(s),s变量表示步态进程百分比,s是时变参数,具体定义如下:Assuming that the standard hip and knee rotation curve functions are f h (s) and f k (s), respectively, the s variable represents the percentage of gait progress, and s is a time-varying parameter, which is specifically defined as follows:

Figure BDA0001762770260000051
Figure BDA0001762770260000051

式中φs(t)表示步态相位角函数,可在步频检测算法中得到。进而可以推出同步步频的关节转角函数如下:where φ s (t) represents the gait phase angle function, which can be obtained in the gait detection algorithm. Then, the joint angle function of the synchronous cadence can be derived as follows:

fh(t)=fh(s(t))f h (t)=f h (s(t))

fk(t)=fk(s(t))f k (t)=f k (s(t))

3.2步幅调整算法3.2 Stride adjustment algorithm

假设标准步态的步幅为Ls,步幅检测的结果为Lr,进而可以定义步幅调整参数K为:Assuming that the stride of the standard gait is L s , and the result of stride detection is L r , the stride adjustment parameter K can be defined as:

Figure BDA0001762770260000052
Figure BDA0001762770260000052

根据步态事件检测结果,可以得到TO事件发生时步态进程百分比值,具体计算公式为:According to the detection result of gait event, the percentage value of gait process when TO event occurs can be obtained. The specific calculation formula is:

Figure BDA0001762770260000053
Figure BDA0001762770260000053

式中tTO表示TO事件发生的时间点,tTE表示TE事件发生的时间点,Ts表示步态周期。然后调整标准轨迹函数如下式:In the formula, t TO represents the time point of the TO event, t TE represents the time point of the TE event, and T s represents the gait cycle. Then adjust the standard trajectory function as follows:

Figure BDA0001762770260000054
Figure BDA0001762770260000054

式中f(s)表示标准关节转角函数,Θ1为步幅拉伸区间,Θ1∈(0,sTO),Θ2为步幅收缩区间,Θ2∈(sTO,100%)。最后应用到步频调整算法中,可以推出公式为:where f(s) represents the standard joint angle function, Θ 1 is the stride stretch interval, Θ 1 ∈(0,s TO ), Θ 2 is the stride contraction interval, Θ 2 ∈(s TO ,100%). Finally, when applied to the step frequency adjustment algorithm, the formula can be derived as:

fn(t)=fn(s(t)) fn (t)= fn (s(t))

式中fn(t)表示参考转角轨迹函数,即同步人体步频和步幅的转角轨迹曲线。where f n (t) represents the reference corner trajectory function, that is, the corner trajectory curve that synchronizes the human cadence and stride length.

综上,本发明设计了一种基于步态事件检测的人体意图识别方法。该方法用到了4个惯性检测单元(IMU)来实时检测人体下肢关节转角(髋关节和膝关节转角),该传感器在价格上相对较低且安装方便,适合机器人产业化发展的目标。在算法的复杂度上,本发明设计的算法也相对较为简单,降低了计算量。To sum up, the present invention designs a human intention recognition method based on gait event detection. This method uses four inertial detection units (IMUs) to detect the joint rotation angles of the lower limbs (hip and knee joints) in real time. In terms of the complexity of the algorithm, the algorithm designed by the present invention is also relatively simple, which reduces the amount of calculation.

Claims (3)

1.一种基于步态事件的人体下肢运动意图识别方法,其特征在于,包括以下步骤:1. a human lower limb movement intention recognition method based on gait event, is characterized in that, comprises the following steps: (1)定义人体下肢步态事件;(1) Define human lower limb gait events; 每个步态相位的开始为步态事件,包括接触开始事件、大腿直立事件、脚趾预离地事件和脚趾离地事件,这些事件与髋关节转角存在对应的映射关系;The start of each gait phase is a gait event, including contact start event, thigh upright event, toe pre-lifting event, and toe-lifting event, these events have a corresponding mapping relationship with hip joint rotation; (2)实时检测步态事件;(2) Real-time detection of gait events; 采集人体下肢的实时髋关节转角数据,依据上述映射关系检测出每个步态周期中的每个步态事件,其中每个步态周期分为四个步态相位,包括支撑初期、支撑中期、支撑末期和摆动期;Collect real-time hip joint rotation angle data of human lower limbs, and detect each gait event in each gait cycle according to the above mapping relationship. Each gait cycle is divided into four gait phases, including initial support, middle support, end of support and swing; (3)基于步态事件得到实时步频与步幅;(3) Obtain real-time stride frequency and stride length based on gait events; 得到实时步频的过程为:The process of obtaining the real-time cadence is as follows: 根据步态事件检测的结果,得到相邻的同一步态事件的时间间隔,从而推出步态周期TsAccording to the result of gait event detection, the time interval between adjacent same gait events is obtained, and the gait cycle T s is deduced:
Figure FDA0002550988240000011
Figure FDA0002550988240000011
式中tn和tn-1分别表示第n次和第n-1次该步态事件发生的时间点,ωs表示步态频率;where t n and t n-1 represent the time points of the nth and n-1th gait events, respectively, and ωs represents the gait frequency; 得到了步态周期后,再定义步态周期的相位角为:After obtaining the gait cycle, define the phase angle of the gait cycle as:
Figure FDA0002550988240000012
Figure FDA0002550988240000012
其中φS步态相位角函数,取值范围为0到2π,然后根据相位角定义步态进程百分比s为:where φ S is the gait phase angle function, which ranges from 0 to 2π, and then defines the gait process percentage s according to the phase angle as:
Figure FDA0002550988240000021
Figure FDA0002550988240000021
得到实时步幅的过程为:The process of getting the real-time stride is: 定义步幅为接触开始事件或脚趾预离地事件发生时两脚质心之间的距离,则步幅的计算公式如下:The stride is defined as the distance between the centers of mass of the two feet when the contact start event or the toe pre-lifting event occurs, and the calculation formula of the stride is as follows: L=l3sinθ3+l4sinθ4-(l1sinθ1+l2sinθ2)L=l 3 sinθ 3 +l 4 sinθ 4 -(l 1 sinθ 1 +l 2 sinθ 2 ) 式中θ1和θ2表示膝关节转角,θ3和θ4表示髋关节转角,
Figure FDA0002550988240000022
where θ 1 and θ 2 represent the knee joint rotation angle, θ 3 and θ 4 represent the hip joint rotation angle,
Figure FDA0002550988240000022
(4)根据实时步频与步幅调整标准关节转角轨迹,并生成同步步频与步幅的参考关节转角轨迹;(4) Adjust the standard joint angle trajectory according to the real-time cadence and stride, and generate the reference joint angle trajectory of synchronous cadence and stride; 步频调整:Cadence adjustment: 假设标准髋关节和膝关节转角曲线函数分别为fh(s)和fk(s),则同步步频的关节转角函数如下:Assuming that the standard hip and knee joint rotation curve functions are f h (s) and f k (s), respectively, the joint rotation function of the synchronous cadence is as follows: fh(t)=fh(s(t))f h (t)=f h (s(t)) fk(t)=fk(s(t))f k (t)=f k (s(t)) 步幅调整:Stride Adjustment: 假设标准步态的步幅为Ls,步幅检测的结果为Lr,则定义步幅调整参数K为:Assuming that the stride of the standard gait is L s and the result of stride detection is L r , the stride adjustment parameter K is defined as:
Figure FDA0002550988240000023
Figure FDA0002550988240000023
根据步态事件检测结果,得到脚趾离地事件发生时步态进程百分比值:According to the gait event detection results, the gait process percentage value when the toe-off event occurs is obtained:
Figure FDA0002550988240000031
Figure FDA0002550988240000031
式中tTO表示脚趾离地事件发生的时间点,tTE表示大腿直立事件发生的时间点;然后调整标准轨迹函数如下式:In the formula, t TO represents the time point of the toe off the ground event, and t TE represents the time point of the thigh upright event; then adjust the standard trajectory function as follows:
Figure FDA0002550988240000032
Figure FDA0002550988240000032
式中f(s)表示标准关节转角函数,Θ1为步幅拉伸区间,Θ1∈(0,sTO),Θ2为步幅收缩区间,Θ2∈(sTO,100%),最后应用到步频调整中,得:where f(s) represents the standard joint angle function, Θ 1 is the stride stretch interval, Θ 1 ∈(0, s TO ), Θ 2 is the stride contraction interval, Θ 2 ∈(s TO , 100%), Finally applied to the cadence adjustment, we get: fn(t)=fn(s(t)) fn (t)= fn (s(t)) 式中fn(t)表示参考转角轨迹函数,即同步人体步频和步幅的转角轨迹曲线。where f n (t) represents the reference corner trajectory function, that is, the corner trajectory curve that synchronizes the human cadence and stride length.
2.根据权利要求1所述的一种基于步态事件的人体下肢运动意图识别方法,其特征在于:为了提升步态周期计算的精准度,将步态周期中定义的四个步态事件都用来计算步态周期并取平均值,具体公式如下:2. a kind of human lower limb motion intention recognition method based on gait event according to claim 1 is characterized in that: in order to improve the accuracy of gait cycle calculation, the four gait events defined in the gait cycle are all Used to calculate the gait cycle and take the average, the specific formula is as follows:
Figure FDA0002550988240000033
Figure FDA0002550988240000033
式中
Figure FDA0002550988240000034
Figure FDA0002550988240000035
分别表示第n次和第n-1次第i个步态事件发生的时间点。
in the formula
Figure FDA0002550988240000034
and
Figure FDA0002550988240000035
represent the time points of the nth and n-1th ith gait events, respectively.
3.根据权利要求1所述的一种基于步态事件的人体下肢运动意图识别方法,其特征在于:在一个完整的步态周期中,存在着两个步态事件可以用来计算步幅,为了提高计算结果的精度,将步幅计算公式表示如下:3. a kind of human lower limb motion intention recognition method based on gait event according to claim 1 is characterized in that: in a complete gait cycle, there are two gait events that can be used to calculate stride length, In order to improve the accuracy of the calculation results, the stride calculation formula is expressed as follows:
Figure FDA0002550988240000041
Figure FDA0002550988240000041
式中
Figure FDA0002550988240000042
Figure FDA0002550988240000043
表示当接触开始事件和脚趾预离地事件发生时髋关节转角值,
Figure FDA0002550988240000044
Figure FDA0002550988240000045
表示当接触开始事件和脚趾预离地事件发生时膝关节转角值。
in the formula
Figure FDA0002550988240000042
and
Figure FDA0002550988240000043
Represents the value of the hip joint rotation when the contact start event and the toe pre-ground event occur,
Figure FDA0002550988240000044
and
Figure FDA0002550988240000045
Indicates the knee joint rotation value when the contact start event and the toe pre-ground event occur.
CN201810915082.0A 2018-08-13 2018-08-13 A gait event-based human lower limb movement intention recognition method Active CN108992072B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810915082.0A CN108992072B (en) 2018-08-13 2018-08-13 A gait event-based human lower limb movement intention recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810915082.0A CN108992072B (en) 2018-08-13 2018-08-13 A gait event-based human lower limb movement intention recognition method

Publications (2)

Publication Number Publication Date
CN108992072A CN108992072A (en) 2018-12-14
CN108992072B true CN108992072B (en) 2020-09-11

Family

ID=64595801

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810915082.0A Active CN108992072B (en) 2018-08-13 2018-08-13 A gait event-based human lower limb movement intention recognition method

Country Status (1)

Country Link
CN (1) CN108992072B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111469117B (en) * 2020-04-14 2022-06-03 武汉理工大学 Human motion mode detection method of rigid-flexible coupling active exoskeleton
CN113425290A (en) * 2021-06-15 2021-09-24 燕山大学 Joint coupling time sequence calculation method for human body rhythm movement
CN114569411B (en) * 2022-02-21 2024-10-08 长沙优龙机器人有限公司 Gait self-adaptive control method and system for hemiplegic patient

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104302251A (en) * 2012-03-22 2015-01-21 埃克苏仿生公司 Human machine interface for lower extremity orthotics
CN105992554A (en) * 2013-12-09 2016-10-05 哈佛大学校长及研究员协会 Assistive flexible suits, flexible suit systems, and methods of manufacture and control thereof for assisting human mobility
CN107361992A (en) * 2016-05-13 2017-11-21 深圳市肯綮科技有限公司 A kind of human body lower limbs move power assisting device

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104302251A (en) * 2012-03-22 2015-01-21 埃克苏仿生公司 Human machine interface for lower extremity orthotics
CN105992554A (en) * 2013-12-09 2016-10-05 哈佛大学校长及研究员协会 Assistive flexible suits, flexible suit systems, and methods of manufacture and control thereof for assisting human mobility
CN107361992A (en) * 2016-05-13 2017-11-21 深圳市肯綮科技有限公司 A kind of human body lower limbs move power assisting device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Gait-Event-Based Synchronization Method for Gait Rehabilitation Robots via a Bioinspired Adaptive Oscillator;Gong Chen等;《IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING》;20170630;第64卷(第6期);第1345-1356页 *

Also Published As

Publication number Publication date
CN108992072A (en) 2018-12-14

Similar Documents

Publication Publication Date Title
Schmitz et al. The measurement of in vivo joint angles during a squat using a single camera markerless motion capture system as compared to a marker based system
Joshi et al. Classification of gait phases from lower limb EMG: Application to exoskeleton orthosis
CN106067178B (en) A kind of continuous estimation method of hand joint movement based on muscle synergistic activation model
Chen et al. Gait-event-based synchronization method for gait rehabilitation robots via a bioinspired adaptive oscillator
CN107273611B (en) A gait planning method for a lower limb rehabilitation robot based on the walking characteristics of the lower limbs
CN108992072B (en) A gait event-based human lower limb movement intention recognition method
CN108960155B (en) Adult gait extraction and abnormal analysis method based on Kinect
Li et al. A human joint torque estimation method for elbow exoskeleton control
Chen et al. Adaptive control strategy for gait rehabilitation robot to assist-when-needed
Ling et al. AE-CNN-based multisource data fusion for gait motion step length estimation
Al-Quraishi et al. Impact of feature extraction techniques on classification accuracy for EMG based ankle joint movements
Bae et al. Real-time estimation of lower extremity joint torques in normal gait
Wen et al. Gait measurement and quantitative analysis in patients with Parkinson's disease for rehabilitation assessment
Lu et al. Gait-Event-Based human intention recognition approach for lower limb
Tong et al. BP-AR-based human joint angle estimation using multi-channel sEMG
Kapti et al. Wearable acceleration sensor application in unilateral trans-tibial amputation prostheses
Meng et al. Effect of walking variations on complementary filter based inertial data fusion for ankle angle measurement
Lyu et al. Comparison of particle filter to established filtering methods in electromyography biofeedback
Su et al. Process mining over sensor data: Goal recognition for powered transhumeral prostheses
Mamikoglu et al. Electromyography based joint angle estimation and control of a robotic leg
Chen et al. An adaptive gait learning strategy for lower limb exoskeleton robot
Hao et al. A gait patterns recognition approach based on surface electromyography and three-axis acceleration signals
Shen et al. Simultaneous gait event intention detection using single semg sensor for lower limb exoskeleton
Shao et al. A method for self-service rehabilitation training of human lower limbs
Bao et al. Motion identification based on sEMG for flexible pneumatic hand rehabilitator

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant