US20040021764A1 - Visual teleconferencing apparatus - Google Patents
Visual teleconferencing apparatus Download PDFInfo
- Publication number
- US20040021764A1 US20040021764A1 US10/336,244 US33624403A US2004021764A1 US 20040021764 A1 US20040021764 A1 US 20040021764A1 US 33624403 A US33624403 A US 33624403A US 2004021764 A1 US2004021764 A1 US 2004021764A1
- Authority
- US
- United States
- Prior art keywords
- visual
- conference
- station
- information
- conference station
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000000007 visual effect Effects 0.000 title claims abstract description 275
- 238000004891 communication Methods 0.000 claims abstract description 56
- 230000007246 mechanism Effects 0.000 claims abstract description 34
- 230000005540 biological transmission Effects 0.000 claims abstract 4
- 238000000034 method Methods 0.000 claims description 88
- 238000004590 computer program Methods 0.000 claims description 7
- 230000009466 transformation Effects 0.000 claims description 6
- 230000003287 optical effect Effects 0.000 claims description 3
- 230000000694 effects Effects 0.000 claims 9
- 230000000977 initiatory effect Effects 0.000 claims 2
- 230000008531 maintenance mechanism Effects 0.000 claims 2
- 230000001131 transforming effect Effects 0.000 claims 1
- 238000012545 processing Methods 0.000 abstract description 3
- 230000008569 process Effects 0.000 description 21
- 230000006835 compression Effects 0.000 description 4
- 238000007906 compression Methods 0.000 description 4
- 230000001755 vocal effect Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000006978 adaptation Effects 0.000 description 2
- 238000013475 authorization Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000006855 networking Effects 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 230000004075 alteration Effects 0.000 description 1
- 230000001143 conditioned effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000012358 sourcing Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
- 230000004382 visual function Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B13/00—Optical objectives specially designed for the purposes specified below
- G02B13/06—Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces
Definitions
- This invention relates to the field of video conferencing.
- Video conferencing systems have been difficult to use and setup, and usually require special configurations and multiple cameras. In comparison, even high-quality audio conference telephones have a very small footprint and are simple to use.
- a major problem with conventional (audio-only) teleconferencing systems is that it is difficult to determine who is on the other end of the line and who is speaking or interjecting words. Voices are identifiable only by their sound qualities (accent, pitch, inflection). In addition, the presence of completely silent parties cannot be determined or verified. Brief interjections can even complicate verbal identity determination because they are so short.
- video conferencing systems are generally not very useful in a conference room setting.
- a typical meeting includes a number of people, generally sitting around a table. Each of the people at the meeting can observe all of the other participants, facial expressions, secondary conversations etc. Much of this participation is lost using prior art video-conferencing systems.
- One aspect of the invention is a visual conference station that includes the facilities of the prior art teleconferencing devices along with a visual component.
- the lens of the visual component is mounted on the device.
- the lens captures a panoramic image of the surrounding scene.
- the captured image is compressed and sent over a network connection to a compatible remote visual conference station (possibly via a conference server) where the panoramic image is presented to the meeting participants at the remote location.
- Other aspects include a device that communicates the visual information that cooperates with an existing audio teleconferencing station.
- One aspect of the invention initializes the visual data communication link from information encoded over a telephone network.
- Another aspect of the invention is a conference server system (and method and program product therefore) that receives visual information from one of the visual conference stations in a conference and distributes that information to other visual conference stations (optionally including the sourcing station).
- FIG. 1A illustrates a side view of a visual conference station in accordance with a preferred embodiment
- FIG. 1B illustrates a top view of the visual conference station of FIG. 1A
- FIG. 2A illustrates a side view of the visual conference station of FIG. 1A in use in accordance with a preferred embodiment
- FIG. 2B illustrates a top view of FIG. 2A
- FIG. 3A illustrates the communications environment of the visual conference station in accordance with a preferred embodiment
- FIG. 3B illustrates the communications environment of the visual conference station in accordance with a preferred embodiment
- FIG. 4 illustrates the visual conference station system architecture in accordance with a preferred embodiment
- FIG. 5 illustrates an initialization procedure in accordance with a preferred embodiment
- FIG. 6 illustrates a visual receive initialization procedure in accordance with a preferred embodiment
- FIG. 7 illustrates a visual send thread procedure in accordance with a preferred embodiment
- FIG. 8 illustrates a visual display thread procedure in accordance with a preferred embodiment
- FIG. 9A illustrates a conference registration process in accordance with a preferred embodiment
- FIG. 9B illustrates a visual information distribution process in accordance with a preferred embodiment.
- FIG. 1A illustrates a side view of a visual conference station 100 that includes a panoramic lens 101 that captures light from substantially 360-degrees around the axis of the lens and that has a vertical field-of-view 103 throughout.
- the panoramic lens 101 is mounted on the top of a housing 105 .
- the housing 105 includes a speaker 107 , a microphone 109 , a control unit (keypad) 111 , and a visual display 113 .
- the housing 105 is shaped and the panoramic lens 101 placed such that the housing 105 does not interfere with the field-of-view of the panoramic lens 101 .
- the speaker 107 , microphone 109 , and control unit (keypad) 111 have similar function to a traditional speakerphone.
- One preferred embodiment uses the panoramic lens 101 (a micro-panoramic lens) such as disclosed in U.S. Pat. No. 6,175,454 by Hoogland and assigned to Be Here Corporation.
- Another preferred embodiment uses a panoramic lens such as disclosed in U.S. Pat. Nos. 6,341,044 or 6,373,642 by Driscoll and assigned to Be Here Corporation.
- These lenses generate annular images of the surrounding panoramic scene.
- other types of wide-angle lenses or combination of lenses can also be used (for example fish-eye lenses, 220-degree lenses, or other lenses that can gather light to illuminate a circle).
- a micro-panoramic lens provides benefits due to its small size.
- the subsequent description is primarily directed towards a panoramic lens that generates an annular image, the invention encompasses the use of wide-angle lenses (such as fish-eye lenses or very-wide angle lenses (for example a 220-degree wide-angle lens)).
- the visual conference station 100 includes communication ports for connection to the telephone network and/or a high-speed communication network (for example, the Internet).
- the visual conference station 100 can include connections for separate speakers, microphones, displays, and/or computer input/output busses.
- FIG. 1B illustrates a top view of the visual conference station 100 of FIG. 1A. Note that the visual display 113 is preferably tilted.
- the panoramic lens 101 is optically connected to an optical sensor integrated circuit (such as a CCD or MOS device).
- an optical sensor integrated circuit such as a CCD or MOS device.
- the visual display 113 can be a liquid crystal display, or any other display that can present a sequence of images (for example, but not limited to, cathode ray tubes and plasma displays).
- the visual conference station 100 is placed in the middle of a table around which the people participating in a conference.
- One visual conference station 100 communicates with another visual conference station 100 to exchange audio information acquired through the microphone 109 and panoramic image information captured by the panoramic lens 101 .
- the audio information is reproduced using the speaker 107 and the image information is presented using the visual display 113 .
- FIG. 2A illustrates a side view of the visual conference station 100 in use on a table with two shown people. Note that the vertical field-of-view 103 captures the head and torso or the meeting participants. In some embodiments, the vertical field-of-view 103 can be such that a portion of the table is also captured.
- FIG. 2B illustrates the placement of the visual conference station 100 . Each of the people around the table is captured by the 360-degree view of the panoramic lens 101 .
- FIG. 3A illustrates a first communications environment 300 for a local visual conference station 301 and a remote visual conference station 303 .
- the local visual conference station 301 and the remote visual conference station 303 communicate using both a telephone network 305 and a high-speed network 307 .
- the telephone network 305 can be used to communicate audio information while the high-speed network 307 can be used to communicate visual information.
- both the visual and audio information is communicated over the high-speed network 307 .
- both the visual and audio information is communicated over the telephone network 305 .
- the conference participants at the one site can view the conference participants at the other site while the conference participants at the other site can also view the conference participants at the one site.
- the telephone network 305 can be used to send sufficient information from the local visual conference station 301 to the remote visual conference station 303 such that the remote visual conference station 303 can make a connection to the local visual conference station 301 using the high-speed network 307 .
- FIG. 3B illustrates a second communications environment 308 wherein the remote visual conference station 303 and the local visual conference station 301 communicate with a visual conferencing server 309 over a network.
- the visual conferencing server 309 connects a multiple of the visual conference station 100 together.
- the local visual conference station 301 sends its annular (or circular) image to the visual conferencing server 309 .
- the visual conferencing server 309 then transforms the annular image into a panoramic image and supplies the panoramic image to the appropriate stations in the conference (such as at least one of the remote visual conference station 303 and/or the local visual conference station 301 ).
- the visual conferencing server 309 can offload the image processing computation from the stations to the visual conferencing server 309 .
- the local visual conference station 301 also provides the visual conferencing server 309 with information about the characteristics of the sent image. This information can be sent with each image, with the image stream, and/or when the local visual conference station 301 registers with the visual conferencing server 309 .
- the conference participants at the one site can view the conference participants at the other site while the conference participants at the other site can also view the conference participants at the one site.
- Another capability of the system shown in FIG. 3B is that it allows one-way participation. That is, participants from the one site can be viewed by a multitude of other sites (the station at the one site sending audio/visual information to the server that redistributes the information to the remote visual conference station 303 at each of the other sites). This allows many observer sites to monitor a meeting at the one site.
- the network transmits information (such as data that defines a panoramic image as well as data that defines a computer program).
- information is embodied within a carrier-wave.
- carrier-wave includes electromagnetic signals, visible or invisible light pulses, signals on a data bus, or signals transmitted over any wire, wireless, or optical fiber technology that allows information to be transmitted over a network.
- Programs and data are commonly read from both tangible physical media (such as a compact, floppy, or magnetic disk) and from a network.
- the network like a tangible physical media, is a computer usable data carrier.
- FIG. 4 illustrates a visual conference station system architecture 400 that includes an image sensor 401 on which the panoramic lens 101 is optically (and in a preferred embodiment also physically) attached.
- the panoramic lens 101 captures light from a 360-degree panoramic scene around the lens that is within the vertical field-of-view 103 . This light from the panoramic scene is focused on the image sensor 401 where an annular or wide-angle image of the panoramic scene is captured.
- the image sensor 401 can be any of the commercially available image sensors (such as a CCD or CMOS sensor).
- the visual conference station system architecture 400 also includes a memory 403 , a control processor 405 , a communication processor 407 , one or more communication ports 409 , a visual display processor 411 , a visual display 413 , a user control interface 415 , a user control input 417 , an audio processor 419 , a telephone line interface 420 and an electronic data bus system 421 .
- a memory 403 a control processor 405 , a communication processor 407 , one or more communication ports 409 , a visual display processor 411 , a visual display 413 , a user control interface 415 , a user control input 417 , an audio processor 419 , a telephone line interface 420 and an electronic data bus system 421 .
- this architecture can be implemented on a single integrated circuit as well as by using multiple integrated circuits and/or a computer.
- the panoramic lens can be a wide-angle lens or a catadioptric lens and in a preferred embodiment is a miniature lens.
- the field-of-view of the panoramic lens extends through the horizon line.
- the memory 403 and the control processor 405 can communicate through the electronic data bus system 421 and/or through a specialized memory bus.
- the control processor 405 can be a general or special purpose programmed processor, an ASIC or other specialized circuitry, or some combination thereof.
- the control processor 405 communicates to the image sensor 401 to cause a digitized representation of the captured panoramic image (the captured visual information) to be transferred to the memory 403 .
- the control processor 405 can then cause all or part of the panoramic image to be transferred (via the communication processor 407 and the one or more communication ports 409 or the telephone line interface 420 ) and/or presented using the visual display processor 411 as conditioned by the user control input 417 through the user control interface 415 .
- a panoramic image can be received by the one or more communication ports 409 and/or the telephone line interface 420 , stored in the memory 403 and presented using the visual display processor 411 and the visual display 413 .
- the local visual conference station 301 and the remote visual conference station 303 directly exchange their respective panoramic images (either as an annular representation or as a rectangular representation) as well as the captured audio information.
- the remote visual conference station 303 and the local visual conference station 301 communicate with the visual conferencing server 309 as previously discussed.
- the visual conference station 100 illustrated in FIG. 1A incorporates the speaker 107 , the microphone 109 , and the visual display 113 , other preferred embodiments need only provide interfaces to one or more of these devices such that the audio and visual information is provided to the audio/visual devices through wire, wireless, and/or optical means.
- the functions of the control unit (keypad) 111 can be provided by many different control mechanisms including (but not limited) to hand-held remote controls, network control programs (such as a browser), voice recognition controls and other control mechanisms.
- the audio processor 419 typically is configured to include both an audio output processor used to drive a speaker and an audio input processor used to receive information from a microphone.
- the video information from the image sensor 401 can be communicated to a computer (for example using a computer peripheral interface such as a SCSI, Firewire®, or USB interface).
- a computer for example using a computer peripheral interface such as a SCSI, Firewire®, or USB interface.
- a computer peripheral interface such as a SCSI, Firewire®, or USB interface.
- one preferred embodiment includes an assembly comprising the panoramic lens 101 and the image sensor 401 where the assembly is in communication with a computer system that provides the communication, audio/visual, user, and networking functionality.
- the visual conference station 100 can include a general-purpose computer capable of being configured to send presentations and other information to the remote stations as well as providing the audio/visual functions previously described.
- a system can also include (or include an interface to) a video projector system.
- FIG. 5 illustrates an ‘initialization’ procedure 500 that can be invoked when the visual conference station 100 is directed to place a visual conference call.
- the ‘initialization’ procedure 500 initiates at a ‘start’ terminal 501 and continues to an ‘establish audio communication’ procedure 503 that receives operator input.
- the visual conference station 100 uses an operator input mechanism (for example, a keypad, a PDA, a web browser, etc.) to input the telephone number of the visual conference station 100 at the remote site.
- the ‘establish audio communication’ procedure 503 uses the operator input to make a connection with the remote visual conference station. This connection can be made over the traditional telephone network or can be established using network telephony.
- the ‘initialization’ procedure 500 continues to a ‘start visual receive initialization thread’ procedure 505 that starts the visual receive initialization thread that is subsequently described with respect to FIG. 6.
- audio information can be exchanged between the stations over the telephone line or the high-speed link.
- captured audio information captured by a microphone at the local site is sent to the remote site where it is received as received audio information and reproduced through a speaker.
- a ‘send visual communication burst information’ procedure 507 encodes the Internet address of the local visual conference station along with additional communication parameters (such as service requirements, encryption keys etc.) and, if desired, textual information such as the names of the people in attendance at the local visual conference station, and/or information that identifies the local visual conference station. Then a ‘delay’ procedure 509 waits for a period of time (usually 1 - 5 seconds). After the delay, a ‘visual communication established’ decision procedure 511 determines whether the remote visual conference station has established visual communication over a high-speed network with the local visual conference station.
- additional communication parameters such as service requirements, encryption keys etc.
- the ‘initialization’ procedure 500 returns to the ‘send visual communication burst information’ procedure 507 to resend the visual communication information.
- the visual conference station operates as a traditional audio conference phone.
- the ‘visual communication established’ decision procedure 511 determines that visual communication has been established with the remote visual conference station
- the ‘initialization’ procedure 500 continues to a ‘start display thread’ procedure 513 that initiates the display thread process as is subsequently described with respect to FIG. 8.
- the ‘initialization’ procedure 500 exits at an ‘end’ terminal 515 .
- FIG. 6 illustrates a visual receive initialization procedure 600 that is invoked by the ‘start visual receive initialization thread’ procedure 505 of FIG. 5 and that initiates at a ‘start’ terminal 601 .
- the visual receive initialization procedure 600 waits at a ‘receive visual communication burst’ procedure 603 for receipt of the visual communication burst information sent by the other visual conference station. Once the visual communication burst information is received, it is parsed and the information made available as needed.
- An establish visual communication procedure 605 uses information received from the ‘receive visual communication burst’ procedure 603 to initiate communication of visual information with the visual conference station that sent the visual communication burst information. This establishment of communication between the visual conference stations can be accomplished by many protocols (such as by exchange of UDP packets or by establishment of a connection using an error correcting protocol and can use well-established Internet streaming protocols).
- the visual receive initialization procedure 600 continues to a ‘start visual send thread’ procedure 607 that initiates the visual send thread that is subsequently described with respect to FIG. 7. Then the visual receive initialization procedure 600 completes through the ‘end’ terminal 609 .
- FIG. 7 illustrates a visual send thread 700 that initiates at a ‘start’ terminal 701 after being invoked by the ‘start visual send thread’ procedure 607 of FIG. 6.
- a ‘receive annular image’ procedure 703 reads the annular (or wide angle) image captured by the panoramic lens 101 from the image sensor 401 into the memory 403 . Then an ‘unwrap annular image’ procedure 705 transforms the captured visual information (the annular or wide-angle image) into a panoramic image (generally, rectangular in shape).
- a ‘compress panoramic image’ procedure 707 then compresses the panoramic image or the captured visual information (either by itself, or with respect to previously compressed panoramic images).
- a ‘send compressed panoramic’ procedure 709 then sends the compressed visual information to the other visual conference station for display (as is subsequently described with respect to FIG. 8.
- a ‘delay’ procedure 711 then waits for a period.
- the visual send thread 700 returns to the ‘receive annular image’ procedure 703 and repeats until the visual portion of the conference call is terminated (for example, by ending the call, by explicit instruction by an operator etc.)
- an operator at the local visual conference station can pause the sending of visual images (for example, using a control analogous to a visual mute button).
- the ‘unwrap annular image’ procedure 705 need not be performed (hence the dashed procedure box in FIG. 7) if this function is provided by a server (such as the visual conferencing server 309 ).
- the ‘compress panoramic image’ procedure 707 can compress the panoramic image using MPEG compression, JPEG compression, JPEG compression with difference information or any techniques well known in the art to compress a stream of images.
- MPEG compression JPEG compression
- JPEG compression JPEG compression with difference information
- any techniques well known in the art to compress a stream of images.
- the ‘unwrap annular image’ procedure 705 and the ‘compress panoramic image’ procedure 707 can be combined into a single step.
- FIG. 8 illustrates a display thread 800 used to display the visual information sent by the ‘send compressed panoramic’ procedure 709 of FIG. 7.
- the display thread 800 is invoked by the ‘start display thread’ procedure 513 of FIG. 5 and initiates at a ‘start’ terminal 801 .
- a ‘receive compressed panorama’ procedure 803 then receives the compressed panorama information (the received visual information) sent by the other visual conference station. Once the panorama information is received, the display thread 800 continues to a ‘present panorama’ procedure 805 that expands the compressed information and displays the resulting visual on the visual display 413 .
- FIG. 5 through FIG. 8 describe aspects of the embodiment shown in FIG. 3A. Such a one would also understand how to adapt these aspects for the embodiment shown in FIG. 3B.
- One adaptation is that the local visual conference station 301 and the remote visual conference station 303 do not communicate directly but instead each communicates with the visual conferencing server 309 .
- Another adaptation can be that neither the local visual conference station 301 nor the remote visual conference station 303 transform the annular or wide-angle image to a panoramic image. Instead, the annular or wide-angle image is compressed and sent to the visual conferencing server 309 where the image is decompressed and transformed into a panoramic image.
- the visual conferencing server 309 then compresses the panoramic image and sends it to the remote visual conference station 303 (or more than one remote station).
- the visual information exchanged between the visual conference stations can include computer-generated visual information (for example, a computer-generated presentation that generates images corresponding to that projected onto a screen).
- FIG. 9A illustrates a ‘conference registration’ process 900 that can be used by the visual conferencing server 309 to establish a conference.
- the ‘conference registration’ process 900 can be used with Internet, local area network, telephone or other protocols.
- the ‘conference registration’ process 900 initiates at a ‘start’ terminal 901 and continues to a ‘receive conference join request’ procedure 903 that receives and validates (verifies that the provided information is in the correct format) a request from a visual conference station to establish or join a conference.
- the information in the request includes a conference identifier and an authorization code (along with sufficient information needed to address the visual conference station making the request).
- a ‘conference established’ decision procedure 905 determines whether the provided information identifies an existing conference. If the identified conference is not already established, the ‘conference registration’ process 900 continues to an ‘establish conference’ procedure 907 that examines the previously validated join request and verifies that the visual conference station making the join request has the capability of establishing the conference. The ‘establish conference’ procedure 907 also determines the properties required for others to join the conference.
- One skilled in the art will understand that there are many ways that a conference can be established. These include, but are not limited to, the conference organizer including a list of authorized visual conference station addresses, providing a conference name and password, and other validation schemas known in the art. If this verification fails, the ‘conference registration’ process 900 processes the next join request (not shown).
- the ‘conference registration’ process 900 continues to a ‘verify authorization’ procedure 909 that examines the previously validated information in the join request to determine whether the visual conference station making the join request is authorized to join the identified conference. If this verification fails, the ‘conference registration’ process 900 processes the next join request (not shown).
- the ‘conference registration’ process 900 continues to an ‘add VCS to conference’ procedure 911 that adds the visual conference station making the request to the conference. Then the ‘conference registration’ process 900 loops back to the ‘receive conference join request’ procedure 903 to handle the next join request.
- FIG. 9B illustrates a ‘distribute visual information’ process 940 can be used to receive visual information from each visual conference station in the conference and to distribute the visual information to each of the member conference stations.
- the ‘distribute visual information’ process 940 can be used, without limitation, to receive the visual information from one member conference station and distribute that information to all the other member conference stations, or all the other member conference stations as well as the one member conference station; to exchange visual information between two member conference stations; and/or to exchange visual information between all member conference stations (subject to the amount of visual information that can be displayed, or operator parameters at a particular visual conference station).
- the ‘distribute visual information’ process 940 initiates at a ‘start’ terminal 941 and continues to a ‘receive visual information from VCS’ procedure 943 that receives visual information from a visual conference station.
- the visual information is examined at a ‘transformation required’ decision procedure 945 to determine whether the visual information is in a rectangular panoramic form and need not be transformed. If the visual information is not in a rectangular panoramic form (thus, the server is to perform the transformation) the ‘distribute visual information’ process 940 continues to a ‘transform visual information’ procedure 947 provides the transformation from the annular or wide-angle format into a rectangular panoramic image and performs any required compression.
- the ‘distribute visual information’ process 940 continues to a ‘send visual information to conference’ procedure 949 where the panoramic image is selectively sent to each of the member conference stations (possibly including the visual conference station that sent the visual information) based on the conference parameters.
- the ‘distribute visual information’ process 940 then continues to a ‘reset active timer’ procedure 951 that resets a timeout timer.
- the timeout timer is used to detect when the conference is completed (that is, when no visual information is being sent to the visual conferencing server 309 for a particular conference).
- the ‘distribute visual information’ process 940 loops back to the ‘receive visual information from VCS’ procedure 943 to receive the next visual information for distribution.
- visual information includes video information of any frame rate, sequences of still images, and computer generated images.
- the described procedures can be implemented as computer programs executed by a computer, by specialized circuitry, or some combination thereof.
- a standard speakerphone or audio conferencing device through, for example, but without limitation, a phone line, an infrared communication mechanism or other a wireless communication mechanism.
- a configuration where the visual conference station 100 includes wire or wireless connections for external computer/video monitors and/or computers (such that computer presentation at one conference station can be made available to each of the visual conference stations; and such that the panoramic image can be presented on projection monitors or on a personal computer in communication with the visual conference station.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Telephonic Communication Services (AREA)
Abstract
An audio/visual conference station that includes a panoramic lens to capture an image of the panoramic scene surrounding the lens. The station also includes communication mechanisms to compress the panoramic image for transmission to a remote audio/visual conference station for display. Thus, people around the remote audio/visual conference station are able to both hear and see those at the local audio/visual conference station and vice versa. The audio/visual conference stations can also communicate through a server system to increase the number of visual conference stations exchanging or sharing images. In addition the server system can off-load some of the image processing steps from the visual conference stations.
Description
- This application claims the benefit of United States Provisional Patent Application serial number: 60/352,779 by Driscoll, filed Jan. 28, 2002.
- 1. Field of the Invention
- This invention relates to the field of video conferencing.
- 2. Background
- Video conferencing systems have been difficult to use and setup, and usually require special configurations and multiple cameras. In comparison, even high-quality audio conference telephones have a very small footprint and are simple to use.
- A major problem with conventional (audio-only) teleconferencing systems is that it is difficult to determine who is on the other end of the line and who is speaking or interjecting words. Voices are identifiable only by their sound qualities (accent, pitch, inflection). In addition, the presence of completely silent parties cannot be determined or verified. Brief interjections can even complicate verbal identity determination because they are so short.
- One reason for the slow adoption of video conferencing systems is that these systems are generally not very useful in a conference room setting. For example, a typical meeting includes a number of people, generally sitting around a table. Each of the people at the meeting can observe all of the other participants, facial expressions, secondary conversations etc. Much of this participation is lost using prior art video-conferencing systems.
- One major problem with prior art videoconferencing systems is that they convert a meeting taking place over a table into a theatre event. That is, a meeting where everyone is facing a large television at the end of the room that has a distracting robotic camera on top of it. This is also true of the remote site where another “theatre” environment is set up. Thus, both the local and remote sites seem to be sitting on a stage looking out at the other audience. This arrangement inhibits and/or masks ordinary meeting behavior, where body language, brief rapid-fire verbal exchanges and other non-verbal behavior are critical. It also prevents the parties in each “theatre” from effectively meeting among their own local peers, because they are all forced to keep their attention at the television at the end of the room.
- It would be advantageous to have a visual conferencing system that is simple to use, has only one lens, has a small footprint and can be positioned in the middle of a conference table.
- One aspect of the invention is a visual conference station that includes the facilities of the prior art teleconferencing devices along with a visual component. The lens of the visual component is mounted on the device. The lens captures a panoramic image of the surrounding scene. The captured image is compressed and sent over a network connection to a compatible remote visual conference station (possibly via a conference server) where the panoramic image is presented to the meeting participants at the remote location.
- Other aspects include a device that communicates the visual information that cooperates with an existing audio teleconferencing station.
- One aspect of the invention initializes the visual data communication link from information encoded over a telephone network.
- Another aspect of the invention is a conference server system (and method and program product therefore) that receives visual information from one of the visual conference stations in a conference and distributes that information to other visual conference stations (optionally including the sourcing station).
- The foregoing and many other aspects of the present invention will no doubt become obvious to those of ordinary skill in the art after having read the following detailed description of the preferred embodiments that are illustrated in the various drawing figures.
- FIG. 1A illustrates a side view of a visual conference station in accordance with a preferred embodiment;
- FIG. 1B illustrates a top view of the visual conference station of FIG. 1A;
- FIG. 2A illustrates a side view of the visual conference station of FIG. 1A in use in accordance with a preferred embodiment;
- FIG. 2B illustrates a top view of FIG. 2A;
- FIG. 3A illustrates the communications environment of the visual conference station in accordance with a preferred embodiment;
- FIG. 3B illustrates the communications environment of the visual conference station in accordance with a preferred embodiment
- FIG. 4 illustrates the visual conference station system architecture in accordance with a preferred embodiment;
- FIG. 5 illustrates an initialization procedure in accordance with a preferred embodiment;
- FIG. 6 illustrates a visual receive initialization procedure in accordance with a preferred embodiment;
- FIG. 7 illustrates a visual send thread procedure in accordance with a preferred embodiment;
- FIG. 8 illustrates a visual display thread procedure in accordance with a preferred embodiment;
- FIG. 9A illustrates a conference registration process in accordance with a preferred embodiment; and
- FIG. 9B illustrates a visual information distribution process in accordance with a preferred embodiment.
- FIG. 1A illustrates a side view of a
visual conference station 100 that includes apanoramic lens 101 that captures light from substantially 360-degrees around the axis of the lens and that has a vertical field-of-view 103 throughout. Thepanoramic lens 101 is mounted on the top of ahousing 105. Thehousing 105 includes aspeaker 107, amicrophone 109, a control unit (keypad) 111, and avisual display 113. - The
housing 105 is shaped and thepanoramic lens 101 placed such that thehousing 105 does not interfere with the field-of-view of thepanoramic lens 101. - The
speaker 107,microphone 109, and control unit (keypad) 111 have similar function to a traditional speakerphone. - One preferred embodiment uses the panoramic lens101 (a micro-panoramic lens) such as disclosed in U.S. Pat. No. 6,175,454 by Hoogland and assigned to Be Here Corporation. Another preferred embodiment uses a panoramic lens such as disclosed in U.S. Pat. Nos. 6,341,044 or 6,373,642 by Driscoll and assigned to Be Here Corporation. These lenses generate annular images of the surrounding panoramic scene. However, other types of wide-angle lenses or combination of lenses can also be used (for example fish-eye lenses, 220-degree lenses, or other lenses that can gather light to illuminate a circle). A micro-panoramic lens provides benefits due to its small size. Although the subsequent description is primarily directed towards a panoramic lens that generates an annular image, the invention encompasses the use of wide-angle lenses (such as fish-eye lenses or very-wide angle lenses (for example a 220-degree wide-angle lens)).
- Although not shown in the figure, the
visual conference station 100 includes communication ports for connection to the telephone network and/or a high-speed communication network (for example, the Internet). In addition, thevisual conference station 100 can include connections for separate speakers, microphones, displays, and/or computer input/output busses. - FIG. 1B illustrates a top view of the
visual conference station 100 of FIG. 1A. Note that thevisual display 113 is preferably tilted. - As is subsequently described, the
panoramic lens 101 is optically connected to an optical sensor integrated circuit (such as a CCD or MOS device). - The
visual display 113 can be a liquid crystal display, or any other display that can present a sequence of images (for example, but not limited to, cathode ray tubes and plasma displays). - Like a traditional high-quality conference phone, the
visual conference station 100 is placed in the middle of a table around which the people participating in a conference. Onevisual conference station 100 communicates with anothervisual conference station 100 to exchange audio information acquired through themicrophone 109 and panoramic image information captured by thepanoramic lens 101. When received, the audio information is reproduced using thespeaker 107 and the image information is presented using thevisual display 113. - FIG. 2A illustrates a side view of the
visual conference station 100 in use on a table with two shown people. Note that the vertical field-of-view 103 captures the head and torso or the meeting participants. In some embodiments, the vertical field-of-view 103 can be such that a portion of the table is also captured. FIG. 2B illustrates the placement of thevisual conference station 100. Each of the people around the table is captured by the 360-degree view of thepanoramic lens 101. - FIG. 3A illustrates a
first communications environment 300 for a localvisual conference station 301 and a remotevisual conference station 303. In one preferred embodiment, the localvisual conference station 301 and the remotevisual conference station 303 communicate using both atelephone network 305 and a high-speed network 307. Thetelephone network 305 can be used to communicate audio information while the high-speed network 307 can be used to communicate visual information. In another preferred embodiment, both the visual and audio information is communicated over the high-speed network 307. In yet another preferred embodiment, both the visual and audio information is communicated over thetelephone network 305. Thus, the conference participants at the one site can view the conference participants at the other site while the conference participants at the other site can also view the conference participants at the one site. - As is subsequently described the
telephone network 305 can be used to send sufficient information from the localvisual conference station 301 to the remotevisual conference station 303 such that the remotevisual conference station 303 can make a connection to the localvisual conference station 301 using the high-speed network 307. - FIG. 3B illustrates a
second communications environment 308 wherein the remotevisual conference station 303 and the localvisual conference station 301 communicate with avisual conferencing server 309 over a network. Thevisual conferencing server 309 connects a multiple of thevisual conference station 100 together. The localvisual conference station 301 sends its annular (or circular) image to thevisual conferencing server 309. Thevisual conferencing server 309 then transforms the annular image into a panoramic image and supplies the panoramic image to the appropriate stations in the conference (such as at least one of the remotevisual conference station 303 and/or the local visual conference station 301). Thus, thevisual conferencing server 309 can offload the image processing computation from the stations to thevisual conferencing server 309. The localvisual conference station 301 also provides thevisual conferencing server 309 with information about the characteristics of the sent image. This information can be sent with each image, with the image stream, and/or when the localvisual conference station 301 registers with thevisual conferencing server 309. Thus, the conference participants at the one site can view the conference participants at the other site while the conference participants at the other site can also view the conference participants at the one site. - Another capability of the system shown in FIG. 3B is that it allows one-way participation. That is, participants from the one site can be viewed by a multitude of other sites (the station at the one site sending audio/visual information to the server that redistributes the information to the remote
visual conference station 303 at each of the other sites). This allows many observer sites to monitor a meeting at the one site. - One skilled in the art will understand that the network transmits information (such as data that defines a panoramic image as well as data that defines a computer program). Generally, the information is embodied within a carrier-wave. The term “carrier-wave” includes electromagnetic signals, visible or invisible light pulses, signals on a data bus, or signals transmitted over any wire, wireless, or optical fiber technology that allows information to be transmitted over a network. Programs and data are commonly read from both tangible physical media (such as a compact, floppy, or magnetic disk) and from a network. Thus, the network, like a tangible physical media, is a computer usable data carrier.
- FIG. 4 illustrates a visual conference
station system architecture 400 that includes animage sensor 401 on which thepanoramic lens 101 is optically (and in a preferred embodiment also physically) attached. Thepanoramic lens 101 captures light from a 360-degree panoramic scene around the lens that is within the vertical field-of-view 103. This light from the panoramic scene is focused on theimage sensor 401 where an annular or wide-angle image of the panoramic scene is captured. Theimage sensor 401 can be any of the commercially available image sensors (such as a CCD or CMOS sensor). The visual conferencestation system architecture 400 also includes amemory 403, acontrol processor 405, acommunication processor 407, one ormore communication ports 409, avisual display processor 411, avisual display 413, auser control interface 415, auser control input 417, anaudio processor 419, atelephone line interface 420 and an electronicdata bus system 421. One skilled in the art will understand that this architecture can be implemented on a single integrated circuit as well as by using multiple integrated circuits and/or a computer. - The panoramic lens can be a wide-angle lens or a catadioptric lens and in a preferred embodiment is a miniature lens. In a preferred embodiment, the field-of-view of the panoramic lens extends through the horizon line.
- The
memory 403 and thecontrol processor 405 can communicate through the electronicdata bus system 421 and/or through a specialized memory bus. Thecontrol processor 405 can be a general or special purpose programmed processor, an ASIC or other specialized circuitry, or some combination thereof. - The
control processor 405 communicates to theimage sensor 401 to cause a digitized representation of the captured panoramic image (the captured visual information) to be transferred to thememory 403. Thecontrol processor 405 can then cause all or part of the panoramic image to be transferred (via thecommunication processor 407 and the one ormore communication ports 409 or the telephone line interface 420) and/or presented using thevisual display processor 411 as conditioned by theuser control input 417 through theuser control interface 415. - In addition, a panoramic image can be received by the one or
more communication ports 409 and/or thetelephone line interface 420, stored in thememory 403 and presented using thevisual display processor 411 and thevisual display 413. - In one preferred embodiment of the visual conference
station system architecture 400, the localvisual conference station 301 and the remotevisual conference station 303 directly exchange their respective panoramic images (either as an annular representation or as a rectangular representation) as well as the captured audio information. - In another preferred embodiment, the remote
visual conference station 303 and the localvisual conference station 301 communicate with thevisual conferencing server 309 as previously discussed. - One skilled in the art would understand that although the
visual conference station 100 illustrated in FIG. 1A incorporates thespeaker 107, themicrophone 109, and thevisual display 113, other preferred embodiments need only provide interfaces to one or more of these devices such that the audio and visual information is provided to the audio/visual devices through wire, wireless, and/or optical means. Further, that the functions of the control unit (keypad) 111 can be provided by many different control mechanisms including (but not limited) to hand-held remote controls, network control programs (such as a browser), voice recognition controls and other control mechanisms. Furthermore, such a one would understand that theaudio processor 419 typically is configured to include both an audio output processor used to drive a speaker and an audio input processor used to receive information from a microphone. - In yet another preferred embodiment, the video information from the
image sensor 401 can be communicated to a computer (for example using a computer peripheral interface such as a SCSI, Firewire®, or USB interface). Thus, one preferred embodiment includes an assembly comprising thepanoramic lens 101 and theimage sensor 401 where the assembly is in communication with a computer system that provides the communication, audio/visual, user, and networking functionality. - In still another embodiment, the
visual conference station 100 can include a general-purpose computer capable of being configured to send presentations and other information to the remote stations as well as providing the audio/visual functions previously described. Such a system can also include (or include an interface to) a video projector system. - FIG. 5 illustrates an ‘initialization’
procedure 500 that can be invoked when thevisual conference station 100 is directed to place a visual conference call. The ‘initialization’procedure 500 initiates at a ‘start’terminal 501 and continues to an ‘establish audio communication’procedure 503 that receives operator input. Thevisual conference station 100 uses an operator input mechanism (for example, a keypad, a PDA, a web browser, etc.) to input the telephone number of thevisual conference station 100 at the remote site. The ‘establish audio communication’procedure 503 uses the operator input to make a connection with the remote visual conference station. This connection can be made over the traditional telephone network or can be established using network telephony. - Once audio communication is established, the ‘initialization’
procedure 500 continues to a ‘start visual receive initialization thread’procedure 505 that starts the visual receive initialization thread that is subsequently described with respect to FIG. 6. - Once audio communication is established, audio information can be exchanged between the stations over the telephone line or the high-speed link. Thus, captured audio information captured by a microphone at the local site is sent to the remote site where it is received as received audio information and reproduced through a speaker.
- A ‘send visual communication burst information’
procedure 507 encodes the Internet address of the local visual conference station along with additional communication parameters (such as service requirements, encryption keys etc.) and, if desired, textual information such as the names of the people in attendance at the local visual conference station, and/or information that identifies the local visual conference station. Then a ‘delay’procedure 509 waits for a period of time (usually 1-5 seconds). After the delay, a ‘visual communication established’decision procedure 511 determines whether the remote visual conference station has established visual communication over a high-speed network with the local visual conference station. If the visual communication has not been established, the ‘initialization’procedure 500 returns to the ‘send visual communication burst information’procedure 507 to resend the visual communication information. Although not specifically shown in FIG. 5, if the visual communication is not established after some time period, this loop ends, and the visual conference station operates as a traditional audio conference phone. - However, if the ‘visual communication established’
decision procedure 511 determines that visual communication has been established with the remote visual conference station, the ‘initialization’procedure 500 continues to a ‘start display thread’procedure 513 that initiates the display thread process as is subsequently described with respect to FIG. 8. - The ‘initialization’
procedure 500 exits at an ‘end’terminal 515. - One skilled in the art will understand that there exist other protocols for establishing communication between the local
visual conference station 301 and the remotevisual conference station 303 other than the one just described. These other protocols will be useful in homogeneous networking environments where both audio and visual information are transmitted over the same network (for example, the internet or the telephone network). - FIG. 6 illustrates a visual receive
initialization procedure 600 that is invoked by the ‘start visual receive initialization thread’procedure 505 of FIG. 5 and that initiates at a ‘start’terminal 601. The visual receiveinitialization procedure 600 waits at a ‘receive visual communication burst’procedure 603 for receipt of the visual communication burst information sent by the other visual conference station. Once the visual communication burst information is received, it is parsed and the information made available as needed. An establishvisual communication procedure 605 uses information received from the ‘receive visual communication burst’procedure 603 to initiate communication of visual information with the visual conference station that sent the visual communication burst information. This establishment of communication between the visual conference stations can be accomplished by many protocols (such as by exchange of UDP packets or by establishment of a connection using an error correcting protocol and can use well-established Internet streaming protocols). - Once the visual communication between the visual conference stations is established, the visual receive
initialization procedure 600 continues to a ‘start visual send thread’procedure 607 that initiates the visual send thread that is subsequently described with respect to FIG. 7. Then the visual receiveinitialization procedure 600 completes through the ‘end’terminal 609. - FIG. 7 illustrates a
visual send thread 700 that initiates at a ‘start’ terminal 701 after being invoked by the ‘start visual send thread’procedure 607 of FIG. 6. A ‘receive annular image’ procedure 703 reads the annular (or wide angle) image captured by thepanoramic lens 101 from theimage sensor 401 into thememory 403. Then an ‘unwrap annular image’ procedure 705 transforms the captured visual information (the annular or wide-angle image) into a panoramic image (generally, rectangular in shape). A ‘compress panoramic image’procedure 707 then compresses the panoramic image or the captured visual information (either by itself, or with respect to previously compressed panoramic images). A ‘send compressed panoramic’procedure 709 then sends the compressed visual information to the other visual conference station for display (as is subsequently described with respect to FIG. 8. A ‘delay’procedure 711 then waits for a period. Thevisual send thread 700 returns to the ‘receive annular image’ procedure 703 and repeats until the visual portion of the conference call is terminated (for example, by ending the call, by explicit instruction by an operator etc.) In addition, an operator at the local visual conference station can pause the sending of visual images (for example, using a control analogous to a visual mute button). - The ‘unwrap annular image’ procedure705 need not be performed (hence the dashed procedure box in FIG. 7) if this function is provided by a server (such as the visual conferencing server 309).
- The ‘compress panoramic image’
procedure 707 can compress the panoramic image using MPEG compression, JPEG compression, JPEG compression with difference information or any techniques well known in the art to compress a stream of images. In addition, one skilled in the art will understand that the ‘unwrap annular image’ procedure 705 and the ‘compress panoramic image’procedure 707 can be combined into a single step. - FIG. 8 illustrates a
display thread 800 used to display the visual information sent by the ‘send compressed panoramic’procedure 709 of FIG. 7. Thedisplay thread 800 is invoked by the ‘start display thread’procedure 513 of FIG. 5 and initiates at a ‘start’terminal 801. A ‘receive compressed panorama’procedure 803 then receives the compressed panorama information (the received visual information) sent by the other visual conference station. Once the panorama information is received, thedisplay thread 800 continues to a ‘present panorama’procedure 805 that expands the compressed information and displays the resulting visual on thevisual display 413. - One skilled in the art will understand that FIG. 5 through FIG. 8 describe aspects of the embodiment shown in FIG. 3A. Such a one would also understand how to adapt these aspects for the embodiment shown in FIG. 3B. One adaptation is that the local
visual conference station 301 and the remotevisual conference station 303 do not communicate directly but instead each communicates with thevisual conferencing server 309. Another adaptation can be that neither the localvisual conference station 301 nor the remotevisual conference station 303 transform the annular or wide-angle image to a panoramic image. Instead, the annular or wide-angle image is compressed and sent to thevisual conferencing server 309 where the image is decompressed and transformed into a panoramic image. Thevisual conferencing server 309 then compresses the panoramic image and sends it to the remote visual conference station 303 (or more than one remote station). Such a one will also understand how to automatically determine whether the localvisual conference station 301 is connecting directly with the remotevisual conference station 303 or to avisual conferencing server 309 and appropriately condition the procedures. Further, one skilled in the art after reading the forgoing will understand that the visual information exchanged between the visual conference stations can include computer-generated visual information (for example, a computer-generated presentation that generates images corresponding to that projected onto a screen). - FIG. 9A illustrates a ‘conference registration’
process 900 that can be used by thevisual conferencing server 309 to establish a conference. The ‘conference registration’process 900 can be used with Internet, local area network, telephone or other protocols. The ‘conference registration’process 900 initiates at a ‘start’terminal 901 and continues to a ‘receive conference join request’procedure 903 that receives and validates (verifies that the provided information is in the correct format) a request from a visual conference station to establish or join a conference. Generally, the information in the request includes a conference identifier and an authorization code (along with sufficient information needed to address the visual conference station making the request). - Next, a ‘conference established’
decision procedure 905 determines whether the provided information identifies an existing conference. If the identified conference is not already established, the ‘conference registration’process 900 continues to an ‘establish conference’procedure 907 that examines the previously validated join request and verifies that the visual conference station making the join request has the capability of establishing the conference. The ‘establish conference’procedure 907 also determines the properties required for others to join the conference. One skilled in the art will understand that there are many ways that a conference can be established. These include, but are not limited to, the conference organizer including a list of authorized visual conference station addresses, providing a conference name and password, and other validation schemas known in the art. If this verification fails, the ‘conference registration’process 900 processes the next join request (not shown). - Once the conference is established, or if the conference was already established, the ‘conference registration’
process 900 continues to a ‘verify authorization’procedure 909 that examines the previously validated information in the join request to determine whether the visual conference station making the join request is authorized to join the identified conference. If this verification fails, the ‘conference registration’process 900 processes the next join request (not shown). - If the join request is verified, the ‘conference registration’
process 900 continues to an ‘add VCS to conference’procedure 911 that adds the visual conference station making the request to the conference. Then the ‘conference registration’process 900 loops back to the ‘receive conference join request’procedure 903 to handle the next join request. - One skilled in the art will understand that there are many ways, equivalent to the one illustrated in FIG. 9A, for establishing a conference.
- FIG. 9B illustrates a ‘distribute visual information’
process 940 can be used to receive visual information from each visual conference station in the conference and to distribute the visual information to each of the member conference stations. The ‘distribute visual information’process 940 can be used, without limitation, to receive the visual information from one member conference station and distribute that information to all the other member conference stations, or all the other member conference stations as well as the one member conference station; to exchange visual information between two member conference stations; and/or to exchange visual information between all member conference stations (subject to the amount of visual information that can be displayed, or operator parameters at a particular visual conference station). - The ‘distribute visual information’
process 940 initiates at a ‘start’terminal 941 and continues to a ‘receive visual information from VCS’procedure 943 that receives visual information from a visual conference station. The visual information is examined at a ‘transformation required’decision procedure 945 to determine whether the visual information is in a rectangular panoramic form and need not be transformed. If the visual information is not in a rectangular panoramic form (thus, the server is to perform the transformation) the ‘distribute visual information’process 940 continues to a ‘transform visual information’procedure 947 provides the transformation from the annular or wide-angle format into a rectangular panoramic image and performs any required compression. Regardless of the branch taken at the ‘transformation required’decision procedure 945, the ‘distribute visual information’process 940 continues to a ‘send visual information to conference’procedure 949 where the panoramic image is selectively sent to each of the member conference stations (possibly including the visual conference station that sent the visual information) based on the conference parameters. - The ‘distribute visual information’
process 940 then continues to a ‘reset active timer’procedure 951 that resets a timeout timer. The timeout timer is used to detect when the conference is completed (that is, when no visual information is being sent to thevisual conferencing server 309 for a particular conference). One skilled in the art will understand that there exist many other ways to detect when the conference terminates extending from explicit ‘leave’ commands to time constraints. After the timer is reset, the ‘distribute visual information’process 940 loops back to the ‘receive visual information from VCS’procedure 943 to receive the next visual information for distribution. - One skilled in the art after reading the forgoing will understand that visual information includes video information of any frame rate, sequences of still images, and computer generated images. In addition, such a one will understand that the described procedures can be implemented as computer programs executed by a computer, by specialized circuitry, or some combination thereof.
- One skilled in the art after reading the forgoing will understand that there are many configurations of the invention. These include, but are not limited to:
- A configuration where a device containing the visual processing portion of the invention is in communication with a standard speakerphone or audio conferencing device (through, for example, but without limitation, a phone line, an infrared communication mechanism or other a wireless communication mechanism). Thus, this configuration can be viewed as an enhancement to an existing audio conference phone.
- A configuration where a separate computer reads the image sensor and provides the necessary visual information processing and communication.
- A configuration where the
visual conference station 100 includes wire or wireless connections for external computer/video monitors and/or computers (such that computer presentation at one conference station can be made available to each of the visual conference stations; and such that the panoramic image can be presented on projection monitors or on a personal computer in communication with the visual conference station. - A configuration where the
visual conference station 100 includes a general-purpose computer. - From the foregoing, it will be appreciated that the invention has (without limitation) the following advantages:
- It returns the “videoconference” format to the natural “people-around-a-table arrangement.” All of the participants at the remote site are now arrayed in front of the participants at the local site (in miniature). Thus, the peopled attending the conference look across the table at each other, and interact in a natural manner.
- It is simpler and cheaper than the prior art videoconferencing systems. It also has a smaller, more acceptable footprint (equivalent to the ubiquitous teleconferencing phones in most meeting rooms).
- It answers the basic question of most meetings: who is attending the meeting, who is speaking, and what the body language and other non-verbal cues are being made by the other participants.
- Unlike the use of robotic cameras, it has no moving parts, makes no noise and thus does not distract the meeting participants.
- It is completely automatic and thus, requires no manual or assisted steering, zooming or adjustment of the camera or lens.
- It gracefully recovers from network problems in that it naturally degrades back to conventional teleconferencing, as opposed to having the meeting collapse because of a lost network connection.
- It can use well-developed video streaming protocols when using IP network environments.
- Although the present invention has been described in terms of the presently preferred embodiments, one skilled in the art will understand that various modifications and alterations may be made without departing from the scope of the invention. Accordingly, the scope of the invention is not to be limited to the particular invention embodiments discussed herein.
Claims (27)
1. A visual conference station comprising:
a housing containing:
a control processor;
a memory coupled to said control processor;
a communications processor in communication with said control processor;
an audio output processor, in communication with said control processor, configured to prepare received audio information for presentation;
a display processor, in communication with said control processor, configured to prepare received visual information for presentation;
an audio input processor, in communication with said control processor, configured to prepare captured audio information captured near said housing for transmission;
an image sensor, in communication with said control processor, said control processor further configured to prepare captured visual information from said image sensor for transmission; and
a panoramic lens in optical communication with said image sensor and configured to capture a scene around the panoramic lens.
2. The visual conference station of claim 1 further comprising:
at least one visual display in communication with said display processor configured to present said received visual information;
at least one speaker in communication with said audio output processor; and
at least one microphone in communication with said audio input processor.
3. The visual conference station of claim 1 wherein said captured visual information is transmitted using said communications processor.
4. The visual conference station of claim 1 wherein said received visual information is received using said communications processor.
5. The visual conference station of claim 1 further comprising a telephone line interface and said received visual information is received using said telephone line interface.
6. The visual conference station of claim 1 wherein said control processor is configured to transform said captured visual information into a rectangular panoramic image prior to transmission.
7. The visual conference station of claim 1 wherein the panoramic lens is a wide-angle lens.
8. The visual conference station of claim 1 wherein the panoramic lens is a catadioptric lens.
9. The visual conference station of claim 1 wherein the panoramic lens has a field-of-view that extends through a horizon line.
10. The visual conference station of claim 1 wherein the panoramic lens is mounted on the housing.
11. The visual conference station of claim 1 wherein said received visual information has been sent from a second visual conference station.
12. The visual conference station of claim 1 wherein said received visual information has been sent from a visual conference server.
13. A method comprising steps of:
receiving visual information from a first visual conference station;
identifying a conference to which said first visual conference station belongs;
identifying a second visual conference station in said conference; and
distributing said visual information to said second visual conference station.
14. The method of claim 13 further comprising transforming said visual information from a non-rectangular image into a panoramic image for distribution.
15. The method of claim 14 further comprising distributing said panoramic image to said first visual conference station.
16. The method of claim 13 further comprising steps of:
establishing said conference responsive to one of said first visual conference station and said second visual conference station;
maintaining said conference; and
terminating said conference.
17. The method of claim 13 further comprising steps of:
receiving audio information from said first visual conference station; and
distributing said audio information to said second visual conference station.
18. A visual conference server comprising:
a receiving mechanism configured to receive visual information from a first visual conference station;
a first identification mechanism configured to identify a conference to which said first visual conference station belongs;
a second identification mechanism, responsive to the first identification mechanism, configured to identify a second visual conference station in said conference; and
a distribution mechanism, responsive to the second identification mechanism, configured to distribute said visual information to said second visual conference station.
19. The visual conference server of claim 18 further comprising a transformation mechanism, responsive to the receiving mechanism, configured to transform said visual information from a non-rectangular image into a panoramic image for distribution.
20. The visual conference server of claim 19 wherein the distribution mechanism is further configured to distribute said panoramic image to said first visual conference station.
21. The visual conference server of claim 18 further comprising:
a conference initiation mechanism configured to establish said conference responsive to one of said first visual conference station and said second visual conference station;
a conference maintenance mechanism configured to maintain said conference; and
a conference termination mechanism configured to terminate said conference.
22. The visual conference server of claim 18 further comprising:
an audio reception mechanism configured to receive audio information from said first visual conference station; and
an audio distribution mechanism, responsive to the audio reception mechanism, configured to distribute said audio information to said second visual conference station.
23. A computer program product comprising:
a computer usable data carrier having computer readable code embodied therein for causing a computer to operate as a visual conference server, said computer readable code including:
computer readable program code configured to cause said computer to effect a receiving mechanism configured to receive visual information from a first visual conference station;
computer readable program code configured to cause said computer to effect a first identification mechanism configured to identify a conference to which said first visual conference station belongs;
a second identification mechanism, responsive to the first identification mechanism, configured to identify a second visual conference station in said conference; and
computer readable program code configured to cause said computer to effect a distribution mechanism, responsive to the second identification mechanism, configured to distribute said visual information to said second visual conference station.
24. The computer program product of claim 23 further comprising computer readable program code configured to cause said computer to effect a transformation mechanism, responsive to the receiving mechanism, configured to transform said visual information from a non-rectangular image into a panoramic image for distribution.
25. The computer program product of claim 24 wherein the distribution mechanism is further configured to distribute said panoramic image to said first visual conference station.
26. The computer program product of claim 23 further comprising:
computer readable program code configured to cause said computer to effect a conference initiation mechanism configured to establish said conference responsive to one of said first visual conference station and said second visual conference station;
computer readable program code configured to cause said computer to effect a conference maintenance mechanism configured to maintain said conference; and
computer readable program code configured to cause said computer to effect a conference termination mechanism configured to terminate said conference.
27. The computer program product of claim 23 further comprising:
computer readable program code configured to cause said computer to effect an audio reception mechanism configured to receive audio information from said first visual conference station; and
a computer readable program code configured to cause said computer to effect n audio distribution mechanism, responsive to the audio reception mechanism, configured to distribute said audio information to said second visual conference station.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/336,244 US20040021764A1 (en) | 2002-01-28 | 2003-01-03 | Visual teleconferencing apparatus |
US10/462,217 US20040008423A1 (en) | 2002-01-28 | 2003-06-12 | Visual teleconferencing apparatus |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US35277902P | 2002-01-28 | 2002-01-28 | |
US10/336,244 US20040021764A1 (en) | 2002-01-28 | 2003-01-03 | Visual teleconferencing apparatus |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/462,217 Continuation-In-Part US20040008423A1 (en) | 2002-01-28 | 2003-06-12 | Visual teleconferencing apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
US20040021764A1 true US20040021764A1 (en) | 2004-02-05 |
Family
ID=46204703
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/336,244 Abandoned US20040021764A1 (en) | 2002-01-28 | 2003-01-03 | Visual teleconferencing apparatus |
Country Status (1)
Country | Link |
---|---|
US (1) | US20040021764A1 (en) |
Cited By (76)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020194613A1 (en) * | 2001-06-06 | 2002-12-19 | Unger Robert Allan | Reconstitution of program streams split across multiple program identifiers |
US20020196327A1 (en) * | 2001-06-14 | 2002-12-26 | Yong Rui | Automated video production system and method using expert video production rules for online publishing of lectures |
US20030152224A1 (en) * | 2002-01-02 | 2003-08-14 | Candelore Brant L. | Video scene change detection |
US20030159139A1 (en) * | 2002-01-02 | 2003-08-21 | Candelore Brant L. | Video slice and active region based dual partial encryption |
US20030174837A1 (en) * | 2002-01-02 | 2003-09-18 | Candelore Brant L. | Content replacement by PID mapping |
US20030234866A1 (en) * | 2002-06-21 | 2003-12-25 | Ross Cutler | System and method for camera color calibration and image stitching |
US20040001137A1 (en) * | 2002-06-27 | 2004-01-01 | Ross Cutler | Integrated design for omni-directional camera and microphone array |
US20040047470A1 (en) * | 2002-09-09 | 2004-03-11 | Candelore Brant L. | Multiple partial encryption using retuning |
US20040049688A1 (en) * | 2001-06-06 | 2004-03-11 | Candelore Brant L. | Upgrading of encryption |
US20040086127A1 (en) * | 2002-11-05 | 2004-05-06 | Candelore Brant L. | Mechanism for protecting the transfer of digital content |
US20040151314A1 (en) * | 1999-03-30 | 2004-08-05 | Candelore Brant L. | Method and apparatus for securing control words |
US20040158721A1 (en) * | 1999-03-30 | 2004-08-12 | Candelore Brant L. | System, method and apparatus for secure digital content transmission |
US20040181666A1 (en) * | 2001-06-06 | 2004-09-16 | Candelore Brant L. | IP delivery of secure digital content |
US20040187161A1 (en) * | 2003-03-20 | 2004-09-23 | Cao Adrean T. | Auxiliary program association table |
US20040240668A1 (en) * | 2003-03-25 | 2004-12-02 | James Bonan | Content scrambling with minimal impact on legacy devices |
US20040254982A1 (en) * | 2003-06-12 | 2004-12-16 | Hoffman Robert G. | Receiving system for video conferencing system |
US20040267521A1 (en) * | 2003-06-25 | 2004-12-30 | Ross Cutler | System and method for audio/video speaker detection |
US20040263646A1 (en) * | 2003-06-24 | 2004-12-30 | Microsoft Corporation | Whiteboard view camera |
US20040263636A1 (en) * | 2003-06-26 | 2004-12-30 | Microsoft Corporation | System and method for distributed meetings |
US20040263611A1 (en) * | 2003-06-26 | 2004-12-30 | Ross Cutler | Omni-directional camera design for video conferencing |
US20050028193A1 (en) * | 2002-01-02 | 2005-02-03 | Candelore Brant L. | Macro-block based content replacement by PID mapping |
US20050036067A1 (en) * | 2003-08-05 | 2005-02-17 | Ryal Kim Annon | Variable perspective view of video images |
US20050046703A1 (en) * | 2002-06-21 | 2005-03-03 | Cutler Ross G. | Color calibration in photographic devices |
US20050063541A1 (en) * | 2002-11-05 | 2005-03-24 | Candelore Brant L. | Digital rights management of a digital device |
US20050066357A1 (en) * | 2003-09-22 | 2005-03-24 | Ryal Kim Annon | Modifying content rating |
US20050097596A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Re-encrypted delivery of video-on-demand content |
US20050094809A1 (en) * | 2003-11-03 | 2005-05-05 | Pedlow Leo M.Jr. | Preparation of content for multiple conditional access methods in video on demand |
US20050097598A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Batch mode session-based encryption of video on demand content |
US20050097614A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Bi-directional indices for trick mode video-on-demand |
US20050102702A1 (en) * | 2003-11-12 | 2005-05-12 | Candelore Brant L. | Cablecard with content manipulation |
US20050117034A1 (en) * | 2002-06-21 | 2005-06-02 | Microsoft Corp. | Temperature compensation in multi-camera photographic devices |
US20050117015A1 (en) * | 2003-06-26 | 2005-06-02 | Microsoft Corp. | Foveated panoramic camera system |
US20050129233A1 (en) * | 2003-12-16 | 2005-06-16 | Pedlow Leo M.Jr. | Composite session-based encryption of Video On Demand content |
US20050151837A1 (en) * | 2002-06-21 | 2005-07-14 | Microsoft Corp. | Minimizing dead zones in panoramic images |
US20050169473A1 (en) * | 2004-02-03 | 2005-08-04 | Candelore Brant L. | Multiple selective encryption with DRM |
US20050192904A1 (en) * | 2002-09-09 | 2005-09-01 | Candelore Brant L. | Selective encryption with coverage encryption |
US20050206659A1 (en) * | 2002-06-28 | 2005-09-22 | Microsoft Corporation | User interface for a system and method for head size equalization in 360 degree panoramic images |
US20050243168A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using audio watermarking techniques |
US20050243166A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video |
US20050243167A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using delta frames |
US20050280700A1 (en) * | 2001-06-14 | 2005-12-22 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
US20050285943A1 (en) * | 2002-06-21 | 2005-12-29 | Cutler Ross G | Automatic face extraction for use in recorded meetings timelines |
US20060018627A1 (en) * | 2004-07-20 | 2006-01-26 | Canon Kabushiki Kaisha | Image reproducing apparatus and image reproducing method |
US20060023075A1 (en) * | 2004-07-28 | 2006-02-02 | Microsoft Corp. | Maintenance of panoramic camera orientation |
US20060023106A1 (en) * | 2004-07-28 | 2006-02-02 | Microsoft Corporation | Multi-view integrated camera system |
US20060146177A1 (en) * | 2004-12-30 | 2006-07-06 | Microsoft Corp. | Camera lens shuttering mechanism |
US20060153379A1 (en) * | 2001-06-06 | 2006-07-13 | Candelore Brant L | Partial encryption and PID mapping |
US20060164552A1 (en) * | 2005-01-21 | 2006-07-27 | Microsoft Corp. | Embedding a panoramic image in a video stream |
US20060184626A1 (en) * | 2005-02-11 | 2006-08-17 | International Business Machines Corporation | Client / server application task allocation based upon client resources |
US20060271492A1 (en) * | 2000-02-15 | 2006-11-30 | Candelore Brant L | Method and apparatus for implementing revocation in broadcast networks |
US20060291478A1 (en) * | 2005-06-24 | 2006-12-28 | Microsoft Corporation | Audio/video synchronization using audio hashing |
US7184609B2 (en) | 2002-06-28 | 2007-02-27 | Microsoft Corp. | System and method for head size equalization in 360 degree panoramic images |
US20070058879A1 (en) * | 2005-09-15 | 2007-03-15 | Microsoft Corporation | Automatic detection of panoramic camera position and orientation table parameters |
US20070098166A1 (en) * | 2002-01-02 | 2007-05-03 | Candelore Brant L | Slice mask and moat pattern partial encryption |
US20070189710A1 (en) * | 2004-12-15 | 2007-08-16 | Pedlow Leo M Jr | Content substitution editor |
US7260257B2 (en) | 2002-06-19 | 2007-08-21 | Microsoft Corp. | System and method for whiteboard and audio capture |
US20070300165A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation, Corporation In The State Of Washington | User interface for sub-conferencing |
US20070299912A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation, Corporation In The State Of Washington | Panoramic video in a live meeting client |
US20070299710A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation | Full collaboration breakout rooms for conferencing |
US20080008458A1 (en) * | 2006-06-26 | 2008-01-10 | Microsoft Corporation | Interactive Recording and Playback for Network Conferencing |
US20080118053A1 (en) * | 2006-11-21 | 2008-05-22 | Beam J Wade | Speak-up |
US20080255840A1 (en) * | 2007-04-16 | 2008-10-16 | Microsoft Corporation | Video Nametags |
US20090003678A1 (en) * | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Automatic gain and exposure control using region of interest detection |
US20090002477A1 (en) * | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Capture device movement compensation for speaker indexing |
US20090002476A1 (en) * | 2007-06-28 | 2009-01-01 | Microsoft Corporation | Microphone array for a camera speakerphone |
US7525928B2 (en) | 2003-06-16 | 2009-04-28 | Microsoft Corporation | System and process for discovery of network-connected devices at remote sites using audio-based discovery techniques |
US20090180025A1 (en) * | 2002-05-28 | 2009-07-16 | Sony Corporation | Method and apparatus for overlaying graphics on video |
US20090222729A1 (en) * | 2008-02-29 | 2009-09-03 | Deshpande Sachin G | Methods and Systems for Audio-Device Activation |
US7730300B2 (en) | 1999-03-30 | 2010-06-01 | Sony Corporation | Method and apparatus for protecting the transfer of data |
US20100183149A1 (en) * | 1999-11-09 | 2010-07-22 | Sony Corporation | Method for simulcrypting scrambled data to a plurality of conditional access devices |
US20110123055A1 (en) * | 2009-11-24 | 2011-05-26 | Sharp Laboratories Of America, Inc. | Multi-channel on-display spatial audio system |
US8041190B2 (en) | 2004-12-15 | 2011-10-18 | Sony Corporation | System and method for the creation, synchronization and delivery of alternate content |
US8185921B2 (en) | 2006-02-28 | 2012-05-22 | Sony Corporation | Parental control of displayed content using closed captioning |
US20160286119A1 (en) * | 2011-04-18 | 2016-09-29 | 360fly, Inc. | Mobile Device-Mountable Panoramic Camera System and Method of Displaying Images Captured Therefrom |
US20190020748A1 (en) * | 2017-05-07 | 2019-01-17 | Compal Electronics, Inc. | Electronic device |
US10951859B2 (en) | 2018-05-30 | 2021-03-16 | Microsoft Technology Licensing, Llc | Videoconferencing device and method |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5686957A (en) * | 1994-07-27 | 1997-11-11 | International Business Machines Corporation | Teleconferencing imaging system with automatic camera steering |
US5808663A (en) * | 1997-01-21 | 1998-09-15 | Dell Computer Corporation | Multimedia carousel for video conferencing and multimedia presentation applications |
US5867209A (en) * | 1994-11-18 | 1999-02-02 | Casio Computer Co., Ltd | Television telephone which displays image data having a first precision degree and image data having a second precision degree on a respective display region of a display screen |
US6356294B1 (en) * | 1998-08-11 | 2002-03-12 | 8×8, Inc. | Multi-point communication arrangement and method |
US6449103B1 (en) * | 1997-04-16 | 2002-09-10 | Jeffrey R. Charles | Solid catadioptric omnidirectional optical system having central coverage means which is associated with a camera, projector, medical instrument, or similar article |
-
2003
- 2003-01-03 US US10/336,244 patent/US20040021764A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5686957A (en) * | 1994-07-27 | 1997-11-11 | International Business Machines Corporation | Teleconferencing imaging system with automatic camera steering |
US5867209A (en) * | 1994-11-18 | 1999-02-02 | Casio Computer Co., Ltd | Television telephone which displays image data having a first precision degree and image data having a second precision degree on a respective display region of a display screen |
US5808663A (en) * | 1997-01-21 | 1998-09-15 | Dell Computer Corporation | Multimedia carousel for video conferencing and multimedia presentation applications |
US6449103B1 (en) * | 1997-04-16 | 2002-09-10 | Jeffrey R. Charles | Solid catadioptric omnidirectional optical system having central coverage means which is associated with a camera, projector, medical instrument, or similar article |
US6356294B1 (en) * | 1998-08-11 | 2002-03-12 | 8×8, Inc. | Multi-point communication arrangement and method |
Cited By (128)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7730300B2 (en) | 1999-03-30 | 2010-06-01 | Sony Corporation | Method and apparatus for protecting the transfer of data |
US20040158721A1 (en) * | 1999-03-30 | 2004-08-12 | Candelore Brant L. | System, method and apparatus for secure digital content transmission |
US20040151314A1 (en) * | 1999-03-30 | 2004-08-05 | Candelore Brant L. | Method and apparatus for securing control words |
US8488788B2 (en) | 1999-11-09 | 2013-07-16 | Sony Corporation | Method for simulcrypting scrambled data to a plurality of conditional access devices |
US20100183149A1 (en) * | 1999-11-09 | 2010-07-22 | Sony Corporation | Method for simulcrypting scrambled data to a plurality of conditional access devices |
US20060271492A1 (en) * | 2000-02-15 | 2006-11-30 | Candelore Brant L | Method and apparatus for implementing revocation in broadcast networks |
US7747853B2 (en) | 2001-06-06 | 2010-06-29 | Sony Corporation | IP delivery of secure digital content |
US20040181666A1 (en) * | 2001-06-06 | 2004-09-16 | Candelore Brant L. | IP delivery of secure digital content |
US20040049688A1 (en) * | 2001-06-06 | 2004-03-11 | Candelore Brant L. | Upgrading of encryption |
US20060153379A1 (en) * | 2001-06-06 | 2006-07-13 | Candelore Brant L | Partial encryption and PID mapping |
US20060269060A1 (en) * | 2001-06-06 | 2006-11-30 | Candelore Brant L | Partial encryption and PID mapping |
US20060262926A1 (en) * | 2001-06-06 | 2006-11-23 | Candelore Brant L | Time division partial encryption |
US20020194613A1 (en) * | 2001-06-06 | 2002-12-19 | Unger Robert Allan | Reconstitution of program streams split across multiple program identifiers |
US7751560B2 (en) | 2001-06-06 | 2010-07-06 | Sony Corporation | Time division partial encryption |
US20070271470A9 (en) * | 2001-06-06 | 2007-11-22 | Candelore Brant L | Upgrading of encryption |
US7895616B2 (en) | 2001-06-06 | 2011-02-22 | Sony Corporation | Reconstitution of program streams split across multiple packet identifiers |
US20050280700A1 (en) * | 2001-06-14 | 2005-12-22 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
US7349005B2 (en) | 2001-06-14 | 2008-03-25 | Microsoft Corporation | Automated video production system and method using expert video production rules for online publishing of lectures |
US7515172B2 (en) | 2001-06-14 | 2009-04-07 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
US7580054B2 (en) | 2001-06-14 | 2009-08-25 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
US20050285933A1 (en) * | 2001-06-14 | 2005-12-29 | Microsoft Corporation | Automated online broadcasting system and method using an omni-directional camera system for viewing meetings over a computer network |
US20020196327A1 (en) * | 2001-06-14 | 2002-12-26 | Yong Rui | Automated video production system and method using expert video production rules for online publishing of lectures |
US7823174B2 (en) | 2002-01-02 | 2010-10-26 | Sony Corporation | Macro-block based content replacement by PID mapping |
US7751563B2 (en) | 2002-01-02 | 2010-07-06 | Sony Corporation | Slice mask and moat pattern partial encryption |
US20050028193A1 (en) * | 2002-01-02 | 2005-02-03 | Candelore Brant L. | Macro-block based content replacement by PID mapping |
US20030152224A1 (en) * | 2002-01-02 | 2003-08-14 | Candelore Brant L. | Video scene change detection |
US20030159139A1 (en) * | 2002-01-02 | 2003-08-21 | Candelore Brant L. | Video slice and active region based dual partial encryption |
US20070098166A1 (en) * | 2002-01-02 | 2007-05-03 | Candelore Brant L | Slice mask and moat pattern partial encryption |
US20030174837A1 (en) * | 2002-01-02 | 2003-09-18 | Candelore Brant L. | Content replacement by PID mapping |
US7765567B2 (en) | 2002-01-02 | 2010-07-27 | Sony Corporation | Content replacement by PID mapping |
US20090180025A1 (en) * | 2002-05-28 | 2009-07-16 | Sony Corporation | Method and apparatus for overlaying graphics on video |
US7260257B2 (en) | 2002-06-19 | 2007-08-21 | Microsoft Corp. | System and method for whiteboard and audio capture |
US7936374B2 (en) | 2002-06-21 | 2011-05-03 | Microsoft Corporation | System and method for camera calibration and images stitching |
US7602412B2 (en) | 2002-06-21 | 2009-10-13 | Microsoft Corporation | Temperature compensation in multi-camera photographic devices |
US20050151837A1 (en) * | 2002-06-21 | 2005-07-14 | Microsoft Corp. | Minimizing dead zones in panoramic images |
US20030234866A1 (en) * | 2002-06-21 | 2003-12-25 | Ross Cutler | System and method for camera color calibration and image stitching |
US7782357B2 (en) * | 2002-06-21 | 2010-08-24 | Microsoft Corporation | Minimizing dead zones in panoramic images |
US20050117034A1 (en) * | 2002-06-21 | 2005-06-02 | Microsoft Corp. | Temperature compensation in multi-camera photographic devices |
US7259784B2 (en) | 2002-06-21 | 2007-08-21 | Microsoft Corporation | System and method for camera color calibration and image stitching |
US20050046703A1 (en) * | 2002-06-21 | 2005-03-03 | Cutler Ross G. | Color calibration in photographic devices |
US20050285943A1 (en) * | 2002-06-21 | 2005-12-29 | Cutler Ross G | Automatic face extraction for use in recorded meetings timelines |
US7598975B2 (en) | 2002-06-21 | 2009-10-06 | Microsoft Corporation | Automatic face extraction for use in recorded meetings timelines |
US20040001137A1 (en) * | 2002-06-27 | 2004-01-01 | Ross Cutler | Integrated design for omni-directional camera and microphone array |
US7852369B2 (en) | 2002-06-27 | 2010-12-14 | Microsoft Corp. | Integrated design for omni-directional camera and microphone array |
US20050206659A1 (en) * | 2002-06-28 | 2005-09-22 | Microsoft Corporation | User interface for a system and method for head size equalization in 360 degree panoramic images |
US7184609B2 (en) | 2002-06-28 | 2007-02-27 | Microsoft Corp. | System and method for head size equalization in 360 degree panoramic images |
US7149367B2 (en) | 2002-06-28 | 2006-12-12 | Microsoft Corp. | User interface for a system and method for head size equalization in 360 degree panoramic images |
US20050192904A1 (en) * | 2002-09-09 | 2005-09-01 | Candelore Brant L. | Selective encryption with coverage encryption |
US20040047470A1 (en) * | 2002-09-09 | 2004-03-11 | Candelore Brant L. | Multiple partial encryption using retuning |
US8818896B2 (en) | 2002-09-09 | 2014-08-26 | Sony Corporation | Selective encryption with coverage encryption |
US7711115B2 (en) | 2002-11-05 | 2010-05-04 | Sony Corporation | Descrambler |
US20050063541A1 (en) * | 2002-11-05 | 2005-03-24 | Candelore Brant L. | Digital rights management of a digital device |
US20060198519A9 (en) * | 2002-11-05 | 2006-09-07 | Candelore Brant L | Digital rights management of a digital device |
US20040086127A1 (en) * | 2002-11-05 | 2004-05-06 | Candelore Brant L. | Mechanism for protecting the transfer of digital content |
US8572408B2 (en) | 2002-11-05 | 2013-10-29 | Sony Corporation | Digital rights management of a digital device |
US20040088558A1 (en) * | 2002-11-05 | 2004-05-06 | Candelore Brant L. | Descrambler |
US7724907B2 (en) | 2002-11-05 | 2010-05-25 | Sony Corporation | Mechanism for protecting the transfer of digital content |
US20040187161A1 (en) * | 2003-03-20 | 2004-09-23 | Cao Adrean T. | Auxiliary program association table |
US20040240668A1 (en) * | 2003-03-25 | 2004-12-02 | James Bonan | Content scrambling with minimal impact on legacy devices |
US20040254982A1 (en) * | 2003-06-12 | 2004-12-16 | Hoffman Robert G. | Receiving system for video conferencing system |
US7525928B2 (en) | 2003-06-16 | 2009-04-28 | Microsoft Corporation | System and process for discovery of network-connected devices at remote sites using audio-based discovery techniques |
US7397504B2 (en) | 2003-06-24 | 2008-07-08 | Microsoft Corp. | Whiteboard view camera |
US20040263646A1 (en) * | 2003-06-24 | 2004-12-30 | Microsoft Corporation | Whiteboard view camera |
US7343289B2 (en) | 2003-06-25 | 2008-03-11 | Microsoft Corp. | System and method for audio/video speaker detection |
US20040267521A1 (en) * | 2003-06-25 | 2004-12-30 | Ross Cutler | System and method for audio/video speaker detection |
US20040263611A1 (en) * | 2003-06-26 | 2004-12-30 | Ross Cutler | Omni-directional camera design for video conferencing |
US20040263636A1 (en) * | 2003-06-26 | 2004-12-30 | Microsoft Corporation | System and method for distributed meetings |
US7428000B2 (en) | 2003-06-26 | 2008-09-23 | Microsoft Corp. | System and method for distributed meetings |
US20050117015A1 (en) * | 2003-06-26 | 2005-06-02 | Microsoft Corp. | Foveated panoramic camera system |
US20050036067A1 (en) * | 2003-08-05 | 2005-02-17 | Ryal Kim Annon | Variable perspective view of video images |
US20050066357A1 (en) * | 2003-09-22 | 2005-03-24 | Ryal Kim Annon | Modifying content rating |
US20050097596A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Re-encrypted delivery of video-on-demand content |
US7853980B2 (en) | 2003-10-31 | 2010-12-14 | Sony Corporation | Bi-directional indices for trick mode video-on-demand |
US20050097598A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Batch mode session-based encryption of video on demand content |
US20050097614A1 (en) * | 2003-10-31 | 2005-05-05 | Pedlow Leo M.Jr. | Bi-directional indices for trick mode video-on-demand |
US20050094809A1 (en) * | 2003-11-03 | 2005-05-05 | Pedlow Leo M.Jr. | Preparation of content for multiple conditional access methods in video on demand |
US20050102702A1 (en) * | 2003-11-12 | 2005-05-12 | Candelore Brant L. | Cablecard with content manipulation |
US20050129233A1 (en) * | 2003-12-16 | 2005-06-16 | Pedlow Leo M.Jr. | Composite session-based encryption of Video On Demand content |
US20050169473A1 (en) * | 2004-02-03 | 2005-08-04 | Candelore Brant L. | Multiple selective encryption with DRM |
US7355622B2 (en) | 2004-04-30 | 2008-04-08 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using delta frames |
US7355623B2 (en) | 2004-04-30 | 2008-04-08 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using audio watermarking techniques |
US7362350B2 (en) | 2004-04-30 | 2008-04-22 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video |
US20050243167A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using delta frames |
US20050243168A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video using audio watermarking techniques |
US20050243166A1 (en) * | 2004-04-30 | 2005-11-03 | Microsoft Corporation | System and process for adding high frame-rate current speaker data to a low frame-rate video |
US20060018627A1 (en) * | 2004-07-20 | 2006-01-26 | Canon Kabushiki Kaisha | Image reproducing apparatus and image reproducing method |
US7593042B2 (en) | 2004-07-28 | 2009-09-22 | Microsoft Corporation | Maintenance of panoramic camera orientation |
US7495694B2 (en) | 2004-07-28 | 2009-02-24 | Microsoft Corp. | Omni-directional camera with calibration and up look angle improvements |
US20060023074A1 (en) * | 2004-07-28 | 2006-02-02 | Microsoft Corporation | Omni-directional camera with calibration and up look angle improvements |
US7593057B2 (en) | 2004-07-28 | 2009-09-22 | Microsoft Corp. | Multi-view integrated camera system with housing |
US20060023106A1 (en) * | 2004-07-28 | 2006-02-02 | Microsoft Corporation | Multi-view integrated camera system |
US20060023075A1 (en) * | 2004-07-28 | 2006-02-02 | Microsoft Corp. | Maintenance of panoramic camera orientation |
US7895617B2 (en) | 2004-12-15 | 2011-02-22 | Sony Corporation | Content substitution editor |
US8041190B2 (en) | 2004-12-15 | 2011-10-18 | Sony Corporation | System and method for the creation, synchronization and delivery of alternate content |
US20100322596A9 (en) * | 2004-12-15 | 2010-12-23 | Pedlow Leo M | Content substitution editor |
US20070189710A1 (en) * | 2004-12-15 | 2007-08-16 | Pedlow Leo M Jr | Content substitution editor |
US7812882B2 (en) | 2004-12-30 | 2010-10-12 | Microsoft Corporation | Camera lens shuttering mechanism |
US20060146177A1 (en) * | 2004-12-30 | 2006-07-06 | Microsoft Corp. | Camera lens shuttering mechanism |
US7768544B2 (en) | 2005-01-21 | 2010-08-03 | Cutler Ross G | Embedding a panoramic image in a video stream |
US20060164552A1 (en) * | 2005-01-21 | 2006-07-27 | Microsoft Corp. | Embedding a panoramic image in a video stream |
US7548977B2 (en) | 2005-02-11 | 2009-06-16 | International Business Machines Corporation | Client / server application task allocation based upon client resources |
US20060184626A1 (en) * | 2005-02-11 | 2006-08-17 | International Business Machines Corporation | Client / server application task allocation based upon client resources |
US20060291478A1 (en) * | 2005-06-24 | 2006-12-28 | Microsoft Corporation | Audio/video synchronization using audio hashing |
US7573868B2 (en) | 2005-06-24 | 2009-08-11 | Microsoft Corporation | Audio/video synchronization using audio hashing |
US20070058879A1 (en) * | 2005-09-15 | 2007-03-15 | Microsoft Corporation | Automatic detection of panoramic camera position and orientation table parameters |
US7630571B2 (en) | 2005-09-15 | 2009-12-08 | Microsoft Corporation | Automatic detection of panoramic camera position and orientation table parameters |
US8185921B2 (en) | 2006-02-28 | 2012-05-22 | Sony Corporation | Parental control of displayed content using closed captioning |
US8572183B2 (en) | 2006-06-26 | 2013-10-29 | Microsoft Corp. | Panoramic video in a live meeting client |
US7653705B2 (en) | 2006-06-26 | 2010-01-26 | Microsoft Corp. | Interactive recording and playback for network conferencing |
US20070299912A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation, Corporation In The State Of Washington | Panoramic video in a live meeting client |
US20080008458A1 (en) * | 2006-06-26 | 2008-01-10 | Microsoft Corporation | Interactive Recording and Playback for Network Conferencing |
US20070300165A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation, Corporation In The State Of Washington | User interface for sub-conferencing |
US20070299710A1 (en) * | 2006-06-26 | 2007-12-27 | Microsoft Corporation | Full collaboration breakout rooms for conferencing |
US20080118053A1 (en) * | 2006-11-21 | 2008-05-22 | Beam J Wade | Speak-up |
US20080255840A1 (en) * | 2007-04-16 | 2008-10-16 | Microsoft Corporation | Video Nametags |
US8526632B2 (en) | 2007-06-28 | 2013-09-03 | Microsoft Corporation | Microphone array for a camera speakerphone |
US20090002476A1 (en) * | 2007-06-28 | 2009-01-01 | Microsoft Corporation | Microphone array for a camera speakerphone |
US20090003678A1 (en) * | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Automatic gain and exposure control using region of interest detection |
US8330787B2 (en) | 2007-06-29 | 2012-12-11 | Microsoft Corporation | Capture device movement compensation for speaker indexing |
US8165416B2 (en) | 2007-06-29 | 2012-04-24 | Microsoft Corporation | Automatic gain and exposure control using region of interest detection |
US8749650B2 (en) | 2007-06-29 | 2014-06-10 | Microsoft Corporation | Capture device movement compensation for speaker indexing |
US20090002477A1 (en) * | 2007-06-29 | 2009-01-01 | Microsoft Corporation | Capture device movement compensation for speaker indexing |
US20090222729A1 (en) * | 2008-02-29 | 2009-09-03 | Deshpande Sachin G | Methods and Systems for Audio-Device Activation |
US20110123055A1 (en) * | 2009-11-24 | 2011-05-26 | Sharp Laboratories Of America, Inc. | Multi-channel on-display spatial audio system |
US20160286119A1 (en) * | 2011-04-18 | 2016-09-29 | 360fly, Inc. | Mobile Device-Mountable Panoramic Camera System and Method of Displaying Images Captured Therefrom |
US20190020748A1 (en) * | 2017-05-07 | 2019-01-17 | Compal Electronics, Inc. | Electronic device |
US10785360B2 (en) * | 2017-05-07 | 2020-09-22 | Compal Electronics, Inc. | Electronic device used for video conference |
US10951859B2 (en) | 2018-05-30 | 2021-03-16 | Microsoft Technology Licensing, Llc | Videoconferencing device and method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20040021764A1 (en) | Visual teleconferencing apparatus | |
US20040008423A1 (en) | Visual teleconferencing apparatus | |
JP4482330B2 (en) | System and method for providing recognition of a remote person in a room during a video conference | |
JP4885928B2 (en) | Video conference system | |
US8359619B2 (en) | Web cam with an interlude switch | |
EP2154885B1 (en) | A caption display method and a video communication control device | |
US7710450B2 (en) | System and method for dynamic control of image capture in a video conference system | |
US20110216153A1 (en) | Digital conferencing for mobile devices | |
US20040041902A1 (en) | Portable videoconferencing system | |
US20060001737A1 (en) | Video conference arrangement | |
US20040119814A1 (en) | Video conferencing system and method | |
US8749611B2 (en) | Video conference system | |
CN103888699A (en) | Projection device with video function and method for video conference by using same | |
CN103905668A (en) | Telephone with video function and method for video conference | |
US7277115B2 (en) | Communication terminal device capable of transmitting visage information | |
CN105959614A (en) | Method and system for processing video conference | |
WO2007035109A1 (en) | Method for gatekeeper streaming | |
KR101918676B1 (en) | Videoconferencing Server for Providing Multi-Screen Videoconferencing by Using Plural Videoconferencing Terminals and Camera Tracking Method therefor | |
WO2025010818A1 (en) | System for screen sharing among conference-participating devices | |
KR100725780B1 (en) | How to connect video call in mobile communication terminal | |
US20070120949A1 (en) | Video, sound, and voice over IP integration system | |
WO2002043360A2 (en) | Multimedia internet meeting interface phone | |
WO2004077829A1 (en) | Video conference system for mobile communication | |
CN101340547B (en) | three-party video conference system | |
TWI419563B (en) | Multimedia transferring system and method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |