CN206451130U - A kind of body-sensing it is interactive immerse type virtual reality system - Google Patents
A kind of body-sensing it is interactive immerse type virtual reality system Download PDFInfo
- Publication number
- CN206451130U CN206451130U CN201720067170.0U CN201720067170U CN206451130U CN 206451130 U CN206451130 U CN 206451130U CN 201720067170 U CN201720067170 U CN 201720067170U CN 206451130 U CN206451130 U CN 206451130U
- Authority
- CN
- China
- Prior art keywords
- computer
- gloves
- helmet
- control box
- frequency divider
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 230000002452 interceptive effect Effects 0.000 title claims description 10
- 238000003325 tomography Methods 0.000 claims abstract description 12
- 238000012806 monitoring device Methods 0.000 claims description 3
- 210000004556 brain Anatomy 0.000 abstract description 5
- 238000005259 measurement Methods 0.000 abstract description 4
- 230000010365 information processing Effects 0.000 abstract description 2
- 230000003993 interaction Effects 0.000 abstract 1
- 238000005516 engineering process Methods 0.000 description 6
- 238000011161 development Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000005672 electromagnetic field Effects 0.000 description 2
- 238000000034 method Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 244000144992 flock Species 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Landscapes
- Position Input By Displaying (AREA)
Abstract
Type virtual reality system is immersed the utility model discloses a kind of interaction of body-sensing, including center-controlling computer, left eye tomography computer, right eye imagery computer, hub, the first frequency divider, the second frequency divider, the first control box, the helmet, right eye monitor, left eye monitor, the second control box, gloves, motion tracking sensor, the center-controlling computer is connected with hub, and hub, left eye tomography computer, the first frequency divider are sequentially connected.The motion tracking sensor measurement head of the utility model helmet and gloves and position and the posture of hand, the posture in each joint of gloves measurement hand;The sensor of gloves and the helmet is connected with respective receiver respectively, is transmitted by RS232 to center-controlling computer and is carried out information processing;Then the position and angle information of head and the sensor of hand are fed back to by two tomography computers by hub, thus the image of adjustment picture and virtual hand, improves the degree of accuracy of the information such as eye, hand, brain.
Description
Technical field
The utility model is related to field of virtual reality, and in particular to a kind of body-sensing it is interactive immerse type virtual reality system.
Background technology
In the virtual reality system based on Panorama Mosaic method, virtual scene is generated according to the following steps, sharp first
With data based on the discrete picture or consecutive image sequence collected, panoramic picture is formed by processing, then by closing
Several panoramic pictures are organized as virtual panoramic space by suitable spatial model.User can be advanced in this space, after
Move back, 360 ° the operation such as look around, look up, overlooking, closely seeing and far seeing.
In IBMR method for drafting, the split of panoramic picture is a more representational basic fundamental.At present widely
Applied to multiple fields such as virtual reality and computer visions, there is universal significance for whole IBMR technologies on its research.
But generally speaking, current virtual reality technology is also very immature.Although in recent years under the concern of all circles, the technology is
Significant progress is achieved, but is due to that its development time is limited, technical difficulty is high, therefore current state of development is apart from each row
The actual demand of each industry still differs greatly.Virtual reality technology will obtain sufficiently application and also have many needs of work to complete.
In existing virtual reality research, the degree of accuracy of the information such as eye, hand, brain is not enough.
Utility model content
Technical problem to be solved in the utility model is that the degree of accuracy of the information such as virtual reality system eye, hand brain is inadequate,
Purpose is to provide a kind of body-sensing the interactive degree of accuracy immersed type virtual reality system, improve the information such as eye, hand, brain, improved
Harmony between virtual reality system modules.
The utility model is achieved through the following technical solutions:
A kind of body-sensing it is interactive immerse type virtual reality system, including center-controlling computer, left eye tomography computer, the right side
Eye tomography computer, hub, the first frequency divider, the second frequency divider, the first control box, the helmet, right eye monitor, left eye monitoring
Device, the second control box, gloves, motion tracking sensor, the center-controlling computer are connected with hub, hub, left eye
Tomography computer, the first frequency divider are sequentially connected, and hub is also sequentially connected with right eye imagery computer, the second frequency divider;It is left
Eye monitor and the first control box are connected with the first frequency divider respectively, and right eye monitor and the first control box are divided with second respectively
Device is connected;First control box, the helmet, motion tracking sensor are sequentially connected;Motion tracking sensor and the second control box difference
It is connected with gloves, center-controlling computer is connected with the second control box;Motion tracking sensor is connected with center-controlling computer.
Control box is used to the output of right eye computer synthesize left eye computer to the helmet, by the slightly differentiated image of two width in the helmet
After being superimposed in display, the display effect of solid will be produced.
Further, left eye monitor and right eye monitor use general purpose PC display.
Further, motion tracking sensor includes gloves receiver, helmet receiver, serial interface cable, emitter, described
The helmet is connected by connecting helmet receiver with motion tracking sensor, and the gloves are passed by connecting gloves receiver with motion
Sensor is connected;Emitter, gloves receiver and helmet receiver are connected with serial interface cable respectively, and the center-controlling computer is led to
Connection serial interface cable is crossed to be connected with motion tracking sensor.Emitter is used to send electromagnetic field signal;Carried on gloves and the helmet
Sensor, gloves receiver and the helmet receive function and receive the signal that gloves and the helmet are sent respectively, the position where judging each
Put.
Further, serial interface cable uses RS232 serial port cable.
The utility model compared with prior art, has the following advantages and advantages:The utility model helmet and hand
The motion tracking sensor measurement head of set and position and the posture of hand, the posture in each joint of gloves measurement hand;Gloves and
The sensor of the helmet is connected with respective receiver respectively, is transmitted by RS232 to center-controlling computer and is carried out information processing;
Then the position and angle information of head and the sensor of hand are fed back to by two tomography computers by hub, thus adjusted
The image of picture and virtual hand, improves the degree of accuracy of the information such as eye, hand, brain.
Brief description of the drawings
Accompanying drawing described herein is used for providing further understanding the utility model embodiment, constitutes the one of the application
Part, does not constitute the restriction to the utility model embodiment.In the accompanying drawings:
Fig. 1 is the utility model structural representation.
Embodiment
For the purpose of this utility model, technical scheme and advantage is more clearly understood, with reference to embodiment and accompanying drawing,
The utility model is described in further detail, and exemplary embodiment of the present utility model and its explanation are only used for explaining this
Utility model, is not intended as limiting of the present utility model.
Embodiment
As shown in figure 1, a kind of body-sensing it is interactive immerse type virtual reality system, including center-controlling computer, left eye into
As computer, right eye imagery computer, hub, the first frequency divider, the second frequency divider, the first control box, the helmet, right eye monitoring
Device, left eye monitor, the second control box, gloves, motion tracking sensor, the center-controlling computer are connected with hub,
Hub, left eye tomography computer, the first frequency divider are sequentially connected, hub also with right eye imagery computer, the second frequency divider
It is sequentially connected;Left eye monitor and the first control box are connected with the first frequency divider respectively, right eye monitor and the first control box point
It is not connected with the second frequency divider;First control box, the helmet, motion tracking sensor are sequentially connected;Motion tracking sensor and
Two control box are connected with gloves respectively, and center-controlling computer is connected with the second control box;Motion tracking sensor is controlled with center
Computer connection processed.Left eye monitor and right eye monitor use general purpose PC display.
Motion tracking sensor includes gloves receiver, helmet receiver, serial interface cable, emitter, and the helmet passes through
Connection helmet receiver is connected with motion tracking sensor, and the gloves are connected by connecting gloves receiver with motion sensor
Connect;Emitter, gloves receiver and helmet receiver are connected with serial interface cable respectively, and the center-controlling computer passes through connection
Serial interface cable is connected with motion tracking sensor.Serial interface cable uses RS232 serial port cable.
Hub is the multiple mechanism of a receipts, center-controlling computer hair, left eye tomography computer and right eye imagery computer
Receive, center-controlling computer believes the motion of hand and attitude information, the positional information of positional information and two eyespots, direction of visual lines
Breath is sent to left eye tomography computer and right eye imagery computer;The video frequency output of computer is given left eye and shown by the first frequency divider
Show device and the helmet, the first control box exports the synthesis of left eye tomography computer and right eye imagery computer to the helmet, by two width
After slightly differentiated image in Helmet Mounted Display to being superimposed, the display effect of solid will be produced.Emitter produces electromagnetic field letter
Number, gloves receiver and helmet receiver receive the signal that the sensor disposed on gloves and the helmet is sent, and then pass through string
Mouth cable is sent to center-controlling computer.
The helmet is a helmet-mounted display (Head Mount Display, abbreviation HMD):Virtual Research
V6, LCD binary channels Helmet Mounted Display one, full color, resolution ratio is 640 × 480.Gloves use data glove (Data
Glove):One, the CyberGlove gloves of Virtual Technologies companies, the right hand, 18 sensors, without tactile,
Force feedback and gesture identification, the gloves are used for measuring the angle in 18 joints of the right hand.Motion tracking sensor:Ascension
The Flock of Birds six degree of freedom electromagnetic position trackers of Technology companies are a set of, equipped with two receivers, use
Host-guest architecture, wherein based on the side equipped with transmitter, the opposing party be from.Connected with a RS232 interface cable from main receiver
It is connected to center-controlling computer.
Above-described embodiment, is entered to the purpose of this utility model, technical scheme and beneficial effect
One step is described in detail, be should be understood that and be the foregoing is only embodiment of the present utility model, is not used to limit
Fixed protection domain of the present utility model, all any modifications within spirit of the present utility model and principle, made, is equally replaced
Change, improve, should be included within protection domain of the present utility model.
Claims (4)
1. a kind of body-sensing it is interactive immerse type virtual reality system, it is characterised in that be imaged including center-controlling computer, left eye
Computer, right eye imagery computer, hub, the first frequency divider, the second frequency divider, the first control box, the helmet, right eye monitoring
Device, left eye monitor, the second control box, gloves, motion tracking sensor, the center-controlling computer are connected with hub,
Hub, left eye tomography computer, the first frequency divider are sequentially connected, hub also with right eye imagery computer, the second frequency divider
It is sequentially connected;Left eye monitor and the first control box are connected with the first frequency divider respectively, right eye monitor and the first control box point
It is not connected with the second frequency divider;First control box, the helmet, motion tracking sensor are sequentially connected;Motion tracking sensor and
Two control box are connected with gloves respectively, and center-controlling computer is connected with the second control box;Motion tracking sensor is controlled with center
Computer connection processed.
2. a kind of body-sensing according to claim 1 it is interactive immerse type virtual reality system, it is characterised in that the left eye
Monitor and right eye monitor use general purpose PC display.
3. a kind of body-sensing according to claim 1 it is interactive immerse type virtual reality system, it is characterised in that the motion
Tracking transducer includes gloves receiver, helmet receiver, serial interface cable, emitter, and the helmet is received by connecting the helmet
Machine is connected with motion tracking sensor, and the gloves are connected by connecting gloves receiver with motion sensor;Emitter, gloves
Receiver and helmet receiver are connected with serial interface cable respectively, and the center-controlling computer is by connecting serial interface cable and motion
Tracking transducer is connected.
4. a kind of body-sensing according to claim 3 it is interactive immerse type virtual reality system, it is characterised in that the serial ports
Cable uses RS232 serial port cable.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201720067170.0U CN206451130U (en) | 2017-01-19 | 2017-01-19 | A kind of body-sensing it is interactive immerse type virtual reality system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201720067170.0U CN206451130U (en) | 2017-01-19 | 2017-01-19 | A kind of body-sensing it is interactive immerse type virtual reality system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN206451130U true CN206451130U (en) | 2017-08-29 |
Family
ID=59666810
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201720067170.0U Expired - Fee Related CN206451130U (en) | 2017-01-19 | 2017-01-19 | A kind of body-sensing it is interactive immerse type virtual reality system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN206451130U (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108388150A (en) * | 2018-04-11 | 2018-08-10 | 武汉灏存科技有限公司 | Simulation method, device, system based on data glove and storage medium |
-
2017
- 2017-01-19 CN CN201720067170.0U patent/CN206451130U/en not_active Expired - Fee Related
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108388150A (en) * | 2018-04-11 | 2018-08-10 | 武汉灏存科技有限公司 | Simulation method, device, system based on data glove and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Schneider et al. | Augmented reality based on edge computing using the example of remote live support | |
US7714895B2 (en) | Interactive and shared augmented reality system and method having local and remote access | |
CN111091611B (en) | An augmented reality system and method for workshop digital twin | |
EP1107052B1 (en) | Method and apparatus for image projection, and apparatus for controlling image projection | |
US6570566B1 (en) | Image processing apparatus, image processing method, and program providing medium | |
CN103977539B (en) | Cervical vertebra rehabilitation health care auxiliary training system | |
CN109804220A (en) | System and method for tracking the fortune dynamic posture on head and eyes | |
Wöhle et al. | Towards robust robot control in cartesian space using an infrastructureless head-and eye-gaze interface | |
CN105809689A (en) | Machine vision-based ship hull six-degree-of-freedom measurement method | |
CN107566793A (en) | Method, apparatus, system and electronic equipment for remote assistance | |
McDowall et al. | Implementation and integration of a counterbalanced CRT-based stereoscopic display for interactive viewpoint control in virtual-environment applications | |
CN113079136B (en) | Motion capture method, motion capture device, electronic equipment and computer-readable storage medium | |
CN112634379A (en) | Three-dimensional positioning measurement method based on mixed vision field light field | |
CN206451130U (en) | A kind of body-sensing it is interactive immerse type virtual reality system | |
CN112446347B (en) | Face direction determining method and device, storage medium and electronic equipment | |
CN113934389A (en) | Three-dimensional scanning processing method, system, processing device and storage medium | |
CN211349296U (en) | Interactive installation is caught in location based on CAVE projection | |
US11941851B2 (en) | Systems and methods for calibrating imaging and spatial orientation sensors | |
CN108115671A (en) | Tow-armed robot control method and system based on 3D visual sensors | |
CN118471050A (en) | Mixed reality helmet system suitable for simulated flight training | |
CN107807738B (en) | Head motion capturing system and method for VR display glasses | |
CN210109743U (en) | VR interactive system based on motion capture | |
CN209746614U (en) | Simulation interaction visualization system of virtual robot workstation | |
CN109214295A (en) | The gesture identification method of data fusion based on Kinect v2 and Leap Motion | |
CN116149465A (en) | System and method for carrying out multi-person bare-hand interaction in virtual reality |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20170829 Termination date: 20180119 |