TWI775128B - Gesture control device and control method thereof - Google Patents
Gesture control device and control method thereof Download PDFInfo
- Publication number
- TWI775128B TWI775128B TW109127565A TW109127565A TWI775128B TW I775128 B TWI775128 B TW I775128B TW 109127565 A TW109127565 A TW 109127565A TW 109127565 A TW109127565 A TW 109127565A TW I775128 B TWI775128 B TW I775128B
- Authority
- TW
- Taiwan
- Prior art keywords
- image
- gesture
- unit
- dimensional
- feature
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 15
- 238000012545 processing Methods 0.000 claims abstract description 42
- 238000010191 image analysis Methods 0.000 claims abstract description 23
- 238000004458 analytical method Methods 0.000 claims abstract description 6
- 238000004364 calculation method Methods 0.000 claims description 22
- 238000012549 training Methods 0.000 claims description 19
- 238000012805 post-processing Methods 0.000 claims description 3
- 230000003993 interaction Effects 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000013461 design Methods 0.000 description 2
- 230000003238 somatosensory effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 210000004905 finger nail Anatomy 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 210000000707 wrist Anatomy 0.000 description 1
Images
Landscapes
- Image Analysis (AREA)
- User Interface Of Digital Computer (AREA)
- Electrotherapy Devices (AREA)
- Selective Calling Equipment (AREA)
Abstract
一種手勢控制裝置及其控制方法,係包括一影像處理模組,該影像處理模組係電性連接一攝影單元,該影像處理模組係用以接收來自於該攝影單元所擷取之二維影像進行後續處理,包括抓取手部二維影像及對該手部二維影像進行灰階化處理及計算該二維影像之特徵點,該影像處理模組另電性連接一影像分析模組,該影像分析模組係用以接收計算完成之二維影像資訊進行後續分析,包括對於計算後之二維影像進行標示,並對所標示之特徵位置進行節點繪製或是計算其他節點,最後再由該手勢判斷單元進行手勢判讀,最後再輸出該手勢判斷之結果。A gesture control device and a control method thereof, comprising an image processing module, the image processing module is electrically connected to a photographing unit, and the image processing module is used for receiving a two-dimensional image captured by the photographing unit The image is subjected to subsequent processing, including capturing the two-dimensional image of the hand, performing grayscale processing on the two-dimensional image of the hand, and calculating the feature points of the two-dimensional image. The image processing module is also electrically connected to an image analysis module. , the image analysis module is used to receive the calculated two-dimensional image information for subsequent analysis, including marking the calculated two-dimensional image, drawing nodes for the marked feature positions or calculating other nodes, and finally The gesture judgment unit performs gesture interpretation, and finally outputs the result of the gesture judgment.
Description
本發明係有關一種控制裝置,尤指一種具有人機互動功能之手勢控制裝置及其控制方法。The present invention relates to a control device, in particular to a gesture control device with human-computer interaction function and a control method thereof.
傳統的人機互動介面為滑鼠、鍵盤和搖桿,隨著科技發展日新月異,為了使人機互動可以更加便利,體感控制提供了一種全新的輸入方式,最常見的體感控制變是透過手勢辨識來進行全新的人機互動。The traditional human-computer interaction interface is mouse, keyboard and joystick. With the rapid development of technology, in order to make human-computer interaction more convenient, somatosensory control provides a new input method. The most common somatosensory control is through Gesture recognition for a new human-computer interaction.
由於手勢是人與人在日常生活中常用的溝通方式,是一種相當直覺且方便的方法,因此,透過技術的引領下,手勢辨逐漸應用於眾多領域之一,包括人機介面設計、醫療復健、虛擬實境與遊戲設計等。Since gestures are a common way of communication between people in daily life, it is a very intuitive and convenient method. Therefore, under the guidance of technology, gesture recognition is gradually applied in one of many fields, including human-machine interface design, medical rehabilitation Health, virtual reality and game design.
而辨識手勢的資訊主要有兩種方式,一種是動態手勢,另一種則是靜態手勢;動態手勢資訊包括手部移動軌跡、位置資訊與時序關係,而靜態手勢資訊則主要為手形變化,藉由分析手勢資訊並根據不同的手勢,來達到人機互動的功能There are two main ways to identify gesture information, one is dynamic gestures and the other is static gestures; dynamic gesture information includes hand movement trajectory, position information and timing relationship, while static gesture information is mainly hand shape changes. Analyze gesture information and achieve the function of human-computer interaction according to different gestures
目前常見手勢辨識的技術,係為利用深度攝影機來取得具有三維影像,每張三維影像影像必須做前置處理,如影像二值化和清除影像背景、消除雜訊等等,再從一連串影像中擷取並分析出使用者手部位置及手勢等相關訊息。再利用手部位置的影像座標數值來控制顯示器之游標的移動;然後,透過三維影像需要在前置處理花費較長時間,致使造成移動游標的速度與精確度難以與滑鼠相較,且深度攝影機此類裝置費用太高,造成目前手勢辨識裝置及技術難以降低成本門檻。At present, the common gesture recognition technology is to use a depth camera to obtain a 3D image. Each 3D image must be pre-processed, such as image binarization, clearing the image background, eliminating noise, etc., and then from a series of images. Capture and analyze the user's hand position and gestures and other related information. Then use the image coordinate value of the hand position to control the movement of the cursor on the display; then, through the 3D image, it takes a long time in the pre-processing, which makes the speed and accuracy of moving the cursor difficult to compare with the mouse, and the depth The cost of such devices as cameras is too high, making it difficult to reduce the cost threshold of current gesture recognition devices and technologies.
針對上述之缺失,本發明之主要目的在於提供一種手勢控制裝置及其控制方法,藉由具有抓取二維影像之攝影機與演算法之搭配,以判斷手勢之變化,達成人機互動之目的,並藉此降低該手勢控制裝置之成本。In view of the above deficiencies, the main purpose of the present invention is to provide a gesture control device and a control method thereof, through the combination of a camera for capturing a two-dimensional image and an algorithm, to determine the change of gestures and achieve the purpose of human-computer interaction, And thereby reduce the cost of the gesture control device.
為達成上述之目的,本發明係主要提供一種手勢控制裝置及其控制方法,係包括一影像處理模組,該影像處理模組係電性連接一攝影單元,該影像處理模組係用以接收來自於該攝影單元所擷取之二維影像進行後續處理,於該影像處理模組內更包括一影像處理單元及一特徵計算單元,其中該影像處理單元係於接收所擷取之二維影像後,進行後製處理,包括抓取手部二維影像及對該手部二維影像進行灰階化處理,待處理過之二維影像輸入進到該特徵計算單元,即開始計算該二維影像之特徵點;該影像處理模組另電性連接一影像分析模組,該影像分析模組係用以接收計算完成之二維影像資訊進行後續分析,其中該影像分析模組內更包括一影像分析單元及一手勢判斷單元,其中該影像分析單元係用以接收二維影像資訊進行,對於計算後之二維影像進行標示,並對所標示之特徵位置進行節點繪製或是計算其他節點,最後再由該手勢判斷單元進行手勢判讀,最後再輸出該手勢判斷之結果。In order to achieve the above object, the present invention mainly provides a gesture control device and a control method thereof, comprising an image processing module, the image processing module is electrically connected to a photographing unit, and the image processing module is used for receiving The two-dimensional image captured by the photographing unit is subjected to subsequent processing. The image processing module further includes an image processing unit and a feature calculation unit, wherein the image processing unit receives the captured two-dimensional image. After that, post-processing is performed, including capturing the 2D image of the hand and performing grayscale processing on the 2D image of the hand. The 2D image to be processed is input into the feature calculation unit, and the calculation of the 2D image is started. feature points of the image; the image processing module is also electrically connected to an image analysis module, and the image analysis module is used for receiving the two-dimensional image information completed by the calculation for subsequent analysis, wherein the image analysis module further includes a An image analysis unit and a gesture determination unit, wherein the image analysis unit is used for receiving two-dimensional image information, marking the calculated two-dimensional image, and performing node drawing for the marked feature position or calculating other nodes, Finally, the gesture judgment unit performs gesture interpretation, and finally outputs the result of the gesture judgment.
為讓本發明之上述和其他目的、特徵和優點能更明顯易懂,下文特舉較佳實施例,並配合所附圖式,作詳細說明如下。In order to make the above-mentioned and other objects, features and advantages of the present invention more clearly understood, preferred embodiments are hereinafter described in detail in conjunction with the accompanying drawings.
先敘明本實施例中所述之「特徵點」,係指透過影像偵測及分析後取得的數個點,其中可以是以關節位置、關節處的皮膚紋路、指甲、掌紋或是手腕周圍之外型等。將所取得的特徵點對比之前訓練時所設定之關節點,進而完成標示關節處的影像關鍵點。利用關鍵點間的連線所形成的圖樣,對比之前所預設之手勢的形狀,,即可用以啟動對應之功能。First, the “feature points” mentioned in this embodiment refer to several points obtained through image detection and analysis, which may be joint positions, skin lines at the joints, fingernails, palm lines, or around the wrist. appearance, etc. The obtained feature points are compared with the joint points set in the previous training, and then the image key points at the joints are marked. Using the pattern formed by the connection between the key points and comparing the shape of the gestures preset before, it can be used to activate the corresponding function.
請參閱第一圖,係為本發明之系統方塊圖。本發明之手勢控制裝置係主要包括一影像處理模組1,該影像處理模組1係電性連接一攝影單元2,該影像處理模組1係用以接收來自於該攝影單元2所擷取之二維影像進行後續處理,而該攝影單元2係為一種具有擷取二維影像功能之攝影機;於該影像處理模組1內更包括一影像處理單元11及一特徵計算單元12,該影像處理單元11與特徵計算單元12電性連接,其中該影像處理單元11係於接收所擷取之二維影像後,進行後製處理,包括抓取手部二維影像及對該手部二維影像進行灰階化處理,待處理過之二維影像輸入進到該特徵計算單元12,即開始計算該二維影像之特徵點(scale-invariant feature transform descriptor) ,而該特徵點之計算方式係針對每個關鍵點(keypoint)擷取16×16之像素大小,再平均切為4×4之格狀大小(cells),每個格狀取梯度及角度值統計為8個箱狀(bins)之直方圖(Histogram),總共16個格狀會得到16個直方圖(8 bins),可合併成16×8=128維度資料,最後針對這些資料進行歸一化處理(L2-Normalizing),即可得到代表該關鍵點之特徵。;亦即特徵點的位置可為二維座標(X,Y),關鍵點經過計算轉換後,其位置座標為三維座標(X,Y,Z)。其計算轉換方式可為經由比較影像中的參考點(例如:臉或背後物件),根據參考點與每一關鍵點的大小比例,進而推算出每一關鍵點的Z值。在取得參考點時,亦可經由手的動作來進行每一關鍵點的Z值的取得或是校正。Please refer to the first figure, which is a system block diagram of the present invention. The gesture control device of the present invention mainly includes an
續參閱第一圖。該影像處理模組1另電性連接一影像分析模組3,該影像分析模組3係用以接收計算完成之二維影像資訊進行後續分析,其中該影像分析模組3內更包括一影像分析單元31及一手勢判斷單元32,該影像分析單元31係訊息電性連接該手勢判斷單元32,其中該影像分析單元31係用以接收二維影像資訊進行,對於計算後之二維影像進行標示,並對所標示之特徵位置進行節點繪製或是計算其他節點,最後再由該手勢判斷單元32進行手勢判讀,最後再輸出該手勢判斷之結果。Continue to refer to the first figure. The
續參閱第一圖。此外,該影像分析模組3係電性連接一手勢訓練模組4,該手勢訓練模組4更包括一手勢紀錄單元41及一手勢訓練單元42,該手勢紀錄單元41係電性連接該手勢訓練單元42,其中該手勢紀錄單元41係用以紀錄手勢影像,並將該些手勢影像載入該手勢訓練單元42進行手勢訓練並產生手勢訓練檔案,該些檔案再傳至該手勢判斷單元32,以作為手勢判斷之依據。Continue to refer to the first figure. In addition, the
請參閱第二圖,係為本發明之手勢控制方法之流程圖。該控制方法之步驟係包括擷取二維手部影像(S1),該二維手部影像係透過一種具有擷取二維影像功能之攝影機進行操作;之後進行計算該二維手部影像之特徵點(S2),包括進行對該手部二維影像進行灰階化處理及特徵點計算,其中該計算之步驟更包括針對每個關鍵點(keypoint)擷取16×16之像素大小(S21),如第三圖之特徵計算流程圖所示,再將每個像素平均切為4×4之格狀大小(cells) (S22),每個格狀取梯度及角度值統計為8個箱狀(bins)之直方圖(Histogram) (S23),總共16個格狀會得到16個直方圖(8 bins),最後合併成16×8=128維度資料(S24),最後針對維度資料進行歸一化處理(L2-Normalizing) (S25),即可得到代表該關鍵點之特徵。Please refer to the second figure, which is a flowchart of the gesture control method of the present invention. The steps of the control method include capturing a two-dimensional hand image ( S1 ), the two-dimensional hand image is operated by a camera with a function of capturing a two-dimensional image; and then calculating the characteristics of the two-dimensional hand image point (S2), including performing gray-scale processing and feature point calculation on the two-dimensional image of the hand, wherein the calculation step further includes extracting a pixel size of 16×16 for each keypoint (S21) , as shown in the feature calculation flow chart in Figure 3, and then cut each pixel into 4 × 4 cells on average (S22), and each cell takes the gradient and angle values and counts them into 8 boxes. The histogram of (bins) (S23), a total of 16 grids will get 16 histograms (8 bins), and finally merge into 16×8=128 dimension data (S24), and finally normalize the dimension data L2-Normalizing (S25), the feature representing the key point can be obtained.
續參閱第三圖。該控制方法之步驟更包括標示該二維手部影像之特徵點(S3),該步驟中係用以計算後之二維影像進行標示,並對所標示之特徵位置進行節點繪製或是計算其他節點;最後,針對該些標示影像進行判斷該二維手部影像之手勢(S4),透過系統內部已存之手勢檔案進行判讀,以輸出最後之判讀結果,即如第四圖至第七圖所示之實施例圖,根據不同的手勢,於系統解讀時,各關鍵點會產生對應移動或是對應手勢之建立。Continue to refer to the third figure. The step of the control method further includes marking the feature points of the two-dimensional hand image (S3), which is used to mark the calculated two-dimensional image in this step, and perform node drawing for the marked feature position or calculate other node; finally, judge the gesture of the two-dimensional hand image based on the marked images (S4), and interpret the gesture file stored in the system to output the final interpretation result, as shown in the fourth to seventh figures. As shown in the embodiment diagram, according to different gestures, when the system interprets, each key point will generate corresponding movement or create corresponding gestures.
惟以上所述之實施方式,是為較佳之實施實例,當不能以此限定本發明實施範圍,若依本發明申請專利範圍及說明書內容所作之等效變化或修飾,皆應屬本發明下述之專利涵蓋範圍。However, the above-mentioned embodiments are preferred implementation examples, which should not limit the scope of the present invention. Any equivalent changes or modifications made according to the scope of the patent application of the present invention and the contents of the description shall belong to the following aspects of the present invention. the scope of patent coverage.
1:影像處理模組1: Image processing module
11:影像處理單元11: Image processing unit
12:特徵計算單元12: Feature calculation unit
2:攝影單元2: Photography unit
3:影像分析模組3: Image analysis module
31:影像分析單元31: Image Analysis Unit
32:手勢判斷單元32: Gesture judgment unit
4:手勢訓練模組4: Gesture training module
41:手勢紀錄單元41: Gesture Recording Unit
42:手勢訓練單元42: Gesture training unit
第一圖、係為本發明之系統方塊圖。The first figure is a system block diagram of the present invention.
第二圖、係為本發明之流程圖。The second figure is a flow chart of the present invention.
第三圖、係為本發明之特徵計算流程圖。The third figure is a flow chart of the characteristic calculation of the present invention.
第四圖、係為本發明之實施例圖(一)。The fourth figure is a figure (1) of the embodiment of the present invention.
第五圖、係為本發明之實施例圖(二)。The fifth figure is the figure (2) of the embodiment of the present invention.
第六圖、係為本發明之實施例圖(三)。The sixth figure is the figure (3) of the embodiment of the present invention.
第七圖、係為本發明之實施例圖(四)。The seventh figure is the figure (4) of the embodiment of the present invention.
1:影像處理模組1: Image processing module
2:攝影單元2: Photography unit
3:影像分析模組3: Image analysis module
4:手勢訓練模組4: Gesture training module
11:影像處理單元11: Image processing unit
12:特徵計算單元12: Feature calculation unit
31:影像分析單元31: Image Analysis Unit
32:手勢判斷單元32: Gesture judgment unit
41:手勢紀錄單元41: Gesture Recording Unit
42:手勢訓練單元42: Gesture training unit
Claims (6)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW109127565A TWI775128B (en) | 2020-08-13 | 2020-08-13 | Gesture control device and control method thereof |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
TW109127565A TWI775128B (en) | 2020-08-13 | 2020-08-13 | Gesture control device and control method thereof |
Publications (2)
Publication Number | Publication Date |
---|---|
TW202207165A TW202207165A (en) | 2022-02-16 |
TWI775128B true TWI775128B (en) | 2022-08-21 |
Family
ID=81323562
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
TW109127565A TWI775128B (en) | 2020-08-13 | 2020-08-13 | Gesture control device and control method thereof |
Country Status (1)
Country | Link |
---|---|
TW (1) | TWI775128B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW200638287A (en) * | 2005-04-20 | 2006-11-01 | Univ Nat Chiao Tung | Image tracking method of an object |
US20180024643A1 (en) * | 2011-08-11 | 2018-01-25 | Eyesight Mobile Technologies Ltd. | Gesture Based Interface System and Method |
US9990050B2 (en) * | 2012-06-18 | 2018-06-05 | Microsoft Technology Licensing, Llc | Dynamic hand gesture recognition using depth data |
TW202027033A (en) * | 2018-12-14 | 2020-07-16 | 大陸商深圳市商湯科技有限公司 | Image processing method and apparatus, electronic device and storage medium |
TWM617136U (en) * | 2020-08-13 | 2021-09-21 | 蔡明勳 | Gesture control device |
-
2020
- 2020-08-13 TW TW109127565A patent/TWI775128B/en not_active IP Right Cessation
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW200638287A (en) * | 2005-04-20 | 2006-11-01 | Univ Nat Chiao Tung | Image tracking method of an object |
US20180024643A1 (en) * | 2011-08-11 | 2018-01-25 | Eyesight Mobile Technologies Ltd. | Gesture Based Interface System and Method |
US9990050B2 (en) * | 2012-06-18 | 2018-06-05 | Microsoft Technology Licensing, Llc | Dynamic hand gesture recognition using depth data |
TW202027033A (en) * | 2018-12-14 | 2020-07-16 | 大陸商深圳市商湯科技有限公司 | Image processing method and apparatus, electronic device and storage medium |
TWM617136U (en) * | 2020-08-13 | 2021-09-21 | 蔡明勳 | Gesture control device |
Also Published As
Publication number | Publication date |
---|---|
TW202207165A (en) | 2022-02-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110991319B (en) | Hand key point detection method, gesture recognition method and related device | |
CN110532984B (en) | Key point detection method, gesture recognition method, device and system | |
Qiang et al. | SqueezeNet and fusion network-based accurate fast fully convolutional network for hand detection and gesture recognition | |
CN102982557B (en) | Method for processing space hand signal gesture command based on depth camera | |
Nair et al. | Hand gesture recognition system for physically challenged people using IOT | |
CN103984928B (en) | Finger gesture recognition methods based on depth image | |
D’Orazio et al. | Recent trends in gesture recognition: how depth data has improved classical approaches | |
CN107885327B (en) | A Fingertip Detection Method Based on Kinect Depth Information | |
US20130335318A1 (en) | Method and apparatus for doing hand and face gesture recognition using 3d sensors and hardware non-linear classifiers | |
CN110837792B (en) | Three-dimensional gesture recognition method and device | |
CN102467657A (en) | Gesture recognition system and method | |
CN103105924B (en) | Man-machine interaction method and device | |
CN103870071B (en) | One kind touches source discrimination and system | |
CN114792443A (en) | Intelligent device gesture recognition control method based on image recognition | |
TW201317843A (en) | Virtual mouse driving apparatus and virtual mouse simulation method | |
CN104067295B (en) | Posture identification method, device for the method and computer program | |
CN104793738A (en) | Non-contact type computer operating method based on Leap Motion | |
CN114445853A (en) | A visual gesture recognition system recognition method | |
CN116909393A (en) | Virtual reality input system based on gesture recognition | |
CN113961067A (en) | Non-contact graffiti drawing method and recognition interaction system based on deep learning | |
CN107292295B (en) | Gesture segmentation method and device | |
CN106484108A (en) | Chinese characters recognition method based on double vision point gesture identification | |
TWI775128B (en) | Gesture control device and control method thereof | |
TWM617136U (en) | Gesture control device | |
Singh et al. | Digitized Interaction: A Gesture-Controlled Whiteboard System with OpenCV, MediaPipe and NumPy |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
GD4A | Issue of patent certificate for granted invention patent | ||
MM4A | Annulment or lapse of patent due to non-payment of fees |