US20120271461A1 - Capturing environmental information - Google Patents
Capturing environmental information Download PDFInfo
- Publication number
- US20120271461A1 US20120271461A1 US13/222,230 US201113222230A US2012271461A1 US 20120271461 A1 US20120271461 A1 US 20120271461A1 US 201113222230 A US201113222230 A US 201113222230A US 2012271461 A1 US2012271461 A1 US 2012271461A1
- Authority
- US
- United States
- Prior art keywords
- uav
- environmental information
- data center
- determining
- event
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000007613 environmental effect Effects 0.000 title claims abstract description 107
- 238000000034 method Methods 0.000 claims abstract description 48
- 230000009471 action Effects 0.000 claims abstract description 28
- 238000012545 processing Methods 0.000 claims description 62
- 238000001816 cooling Methods 0.000 claims description 51
- 230000003213 activating effect Effects 0.000 claims description 4
- 230000008859 change Effects 0.000 claims description 4
- 230000004044 response Effects 0.000 description 43
- 238000012544 monitoring process Methods 0.000 description 36
- 230000015654 memory Effects 0.000 description 35
- 238000004891 communication Methods 0.000 description 21
- 230000000007 visual effect Effects 0.000 description 14
- 230000005540 biological transmission Effects 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 238000004590 computer program Methods 0.000 description 5
- 238000001514 detection method Methods 0.000 description 5
- 208000037408 Device failure Diseases 0.000 description 4
- 238000013461 design Methods 0.000 description 4
- 230000001960 triggered effect Effects 0.000 description 4
- 230000002159 abnormal effect Effects 0.000 description 3
- 230000001413 cellular effect Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000005484 gravity Effects 0.000 description 3
- 230000000977 initiatory effect Effects 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000005019 pattern of movement Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 238000003032 molecular docking Methods 0.000 description 2
- 230000000737 periodic effect Effects 0.000 description 2
- 239000013589 supplement Substances 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 238000009529 body temperature measurement Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000013481 data capture Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000020169 heat generation Effects 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000000644 propagated effect Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000000087 stabilizing effect Effects 0.000 description 1
- 238000010998 test method Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 239000013598 vector Substances 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01W—METEOROLOGY
- G01W1/00—Meteorology
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C23/00—Combined instruments indicating more than one navigational value, e.g. for aircraft; Combined measuring devices for measuring two or more variables of movement, e.g. distance, speed or acceleration
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01W—METEOROLOGY
- G01W1/00—Meteorology
- G01W1/08—Adaptations of balloons, missiles, or aircraft for meteorological purposes; Radiosondes
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/10—Simultaneous control of position or course in three dimensions
- G05D1/101—Simultaneous control of position or course in three dimensions specially adapted for aircraft
- G05D1/102—Simultaneous control of position or course in three dimensions specially adapted for aircraft specially adapted for vertical take-off of aircraft
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D23/00—Control of temperature
- G05D23/19—Control of temperature characterised by the use of electric means
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/50—Navigation or guidance aids
- G08G5/55—Navigation or guidance aids for a single aircraft
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B64—AIRCRAFT; AVIATION; COSMONAUTICS
- B64U—UNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
- B64U2101/00—UAVs specially adapted for particular uses or applications
- B64U2101/35—UAVs specially adapted for particular uses or applications for science, e.g. meteorology
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01W—METEOROLOGY
- G01W1/00—Meteorology
- G01W2001/006—Main server receiving weather information from several sub-stations
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/50—Navigation or guidance aids
- G08G5/57—Navigation or guidance aids for unmanned aircraft
Definitions
- This specification relates to capturing environmental information.
- Data centers e.g., a room containing a number of machines, such as, for example, servers, network devices and storage devices
- Data centers are an example of a setting in which it is important to monitor environmental characteristics.
- data center equipment can generate a significant amount of heat, and increasing the power density of a system can result in an increase in the number of heat concentration pockets.
- Combinations of air and liquid cooling solutions and monitoring techniques can be used to manage the environmental characteristics of a setting such as a data center.
- a method includes receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center, storing the environmental information on one or more memory devices that are accessible by the one or more computing devices, determining, based on the environmental information, that an event has occurred, and performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
- UAV unmanned aerial vehicle
- the environmental information is associated with one or more of a temperature and a humidity level of at least a portion of the data center. Determining that the event has occurred includes determining that the temperature or the humidity level associated with the at least a portion of the data center falls outside a predetermined range. Performing the one or more actions includes activating a cooling system associated with the at least a portion of the data center. Performing the one or more actions includes adjusting an air vent associated with the at least a portion of the data center. Adjusting the air vent alters a flow of air through the at least a portion of the data center.
- the one or more actions include causing the UAV to travel to the at least a portion of the data center to collect additional environmental information. It is determined that the additional environmental information is associated with one or more of a second temperature and a second humidity level that falls outside the predetermined range, and a notification is provided. Providing the notification includes providing at least one of either: an on screen pop-up message, or an automated email, or a short message service (SMS) message, or any combination thereof. Determining that the event has occurred includes determining that a cooling system has been disabled. Performing the one or more actions includes altering an operation of one or more other cooling systems, and causing the UAV to travel to one or more portions of the data center to collect additional environmental information.
- SMS short message service
- a frequency is altered with which the UAV collects environmental information associated with a portion of the data center that is associated with the disabled cooling system.
- the environmental information specifies that the UAV has experienced a navigation deviation. Determining that the event has occurred includes determining, based on the navigation deviation, that a mechanical change has occurred in the data center. Determining that the mechanical change has occurred includes determining that a door located in the data center is ajar. Determining that the event has occurred includes determining that a predetermined amount of environmental information has been received. Performing the action includes analyzing the environmental information to determine the location of one or more portions of the data center that are associated with environmental information that falls outside of a predetermined range.
- Determining that the event has occurred includes determining that environmental information has not been collected for a portion of the data center. It is determined that the portion of the data center is inaccessible by the UAV. It is determined that a size of the portion of the data center exceeds a predetermined size. The UAV is caused to collect environmental information for the portion of the data center based on determining that the size of the portion of the data center exceeds the predetermined size.
- FIG. 1 is a diagram of a system for capturing environmental information.
- FIG. 2 is a flow chart of a process for capturing environmental information.
- FIG. 3 is a diagram of a base station.
- FIG. 4 is a diagram of data center.
- FIG. 5 is a diagram of user interface that displays captured environmental information.
- FIGS. 6-10 are diagrams of data centers.
- FIG. 11 is a diagram of a computing system.
- FIG. 1 shows a system 100 in which a vehicle patrols an area to monitor environmental conditions of the area.
- the vehicle is a UAV 102 and, specifically, is a self-charging quad-rotor UAV.
- the system 100 monitors environmental conditions such as temperature, noise, pressure, and humidity, within a setting such as a data center.
- the UAV 102 includes a number of sensors that are used to both navigate the UAV 102 and to record environmental information representative of environmental conditions of the environment in which the UAV 102 is operating.
- the environmental information recorded by the UAV 102 can be used to locate areas of concern in a data center, such as areas or pockets of abnormally high heat, noise or humidity.
- the environmental information can also be used to generate visual representations, such as histograms, of the environmental conditions within the monitored environment.
- the environmental information recorded by the UAV 102 can be used to generate data representations (e.g., graphs and spreadsheets) that reflect a time history of the monitored environmental conditions.
- Patterns can be detected in these data representations to automatically determine, for example, whether any corrective actions need to be taken. For example, if the environmental information shows that a particular location or locations within a monitored data center is abnormally hot during a particular time period each day (e.g., 9:30 AM on Mondays, when the machine might be under a heavy load), an administrator or an engine could choose to take extra temperature control measures during that time such as, for example, moving some equipment to another location to distribute the heat generation, or adjusting the airflow vents in the area to better cool the environment.
- Such patterns may be represented visually by a graph or chart.
- the UAV 102 includes a combination of hardware and software to both navigate the UAV 102 and to acquire and transmit environmental information to a central location (e.g., the base station 104 ).
- the UAV 102 includes one or more environmental sensors, such as a thermometer 124 and a humidity sensor 126 . These sensors are located in positions that will not be affected by the operation of the rotors 128 (e.g., above the rotors 128 or in front of a frame that supports the rotors, as shown).
- the UAV 102 also includes navigational features, such as an altimeter 112 , a radio-frequency identification (RFID) sensor 110 , a compass 116 (e.g., an electronic compass), and a proximity sensor 118 .
- the compass 116 provides a heading or bearing of the UAV 102 (e.g., by providing information that allows a relative bearing to be calculated) and can be an analog or digital compass.
- the altimeter 112 provides an altitude of the UAV 102 , and can be implemented as a downward-facing infrared altimeter or an ultrasonic altimeter.
- the proximity sensor 118 provides collision detection functionality using infrared or ultrasonic obstacle detection techniques.
- the UAV 102 may also include a horizon detection device 130 (e.g., a camera) for stabilizing and properly orienting the UAV 102 .
- the RFID sensor provides a position of the UAV 102 relative to one or more beacons (e.g., the first beacon 106 ), as described in further detail below.
- the navigational features communicate with a navigation engine 114 to navigate the UAV 102 .
- the navigation engine 114 is an application running on a processing device associated with the UAV 102 , and uses values provided by the navigational features to navigate the UAV 102 .
- the processing device may be on board UAV 102 and thus the navigation is performed locally.
- the processing device may be located remotely from UAV 102 .
- UAV 102 may send sensor data to the processing device wirelessly and may receive navigation information from the processing device also wirelessly.
- the UAV 102 also includes a report generation engine 120 .
- the report generation engine 120 generates reports (e.g., report 134 ) that provide the environmental conditions of a particular location.
- the report generation engine 120 uses data provided by the altimeter 112 , the RFID sensor 110 , the thermometer 124 , the humidity sensor 126 , and the compass 116 to generate reports that are transmitted to a central location using a transmission device 122 .
- the transmission device 122 transmits reports using one or more wireless transmission protocols, such as WiFi, Bluetooth, radio communication, and the like.
- An example of a protocol that can be used is XBee wireless communication protocol (IEEE 802.15.4) which uses low power radio frequency at 2.4 GH.
- the system 100 includes a first beacon 106 and a second beacon 108 .
- the number of beacons included a system can depend on the range of the beacon and the size of the environment being monitored.
- the first and second beacons 106 , 108 are configured to transmit respective pilot signals that can be detected by sensors (e.g., an RFID sensor 110 ) on the UAV 102 .
- the UAV 102 uses the pilot signals transmitted by the first and second beacons 106 , 108 to navigate to various locations within a monitored environment.
- the first and second beacons 106 , 108 can be placed at locations within a monitored environment to act as waypoints for the UAV 102 .
- Each of the first and second beacons 106 , 108 also transmits a Beacon ID that uniquely identifies its associated beacon.
- the UAV 102 remains docked at a base station 104 until a monitoring session is triggered, either by a user or as part of a predefined schedule.
- a monitoring session is triggered, either by a user or as part of a predefined schedule.
- the UAV is docked at the base station 104 .
- a battery 109 associated with the UAV 102 may be charged by a charging device ( FIG. 3 ) located in the base station 104 .
- Terminals of the battery 109 can be located in a base 132 of the UAV 102 .
- a monitoring session can be initiated which activates the UAV 102 (e.g., powers on the UAV 102 from an OFF or “sleep” state). Based on a triggering event, the UAV 102 powers up, executes one or more self-test procedures, and begins to navigate according to a flight plan that can be stored locally on the UAV 102 or at a remote control station (e.g., the base station 104 or a separate entity) and transmitted to the UAV 102 before or during the monitoring session.
- a remote control station e.g., the base station 104 or a separate entity
- the UAV 102 attempts to detect a pilot signal associated with the first beacon 106 using an onboard RFID sensor 110 .
- the UAV 102 detects a pilot signal associated with the first beacon 106 and determines the signal strength of that pilot signal.
- the signal strength can be measured in decibels (dB) or as dBm or dBmW (a power ratio in decibels of the measured power referenced to one milliwatt (mW)).
- the UAV determines that the signal strength of the pilot signal associated with the first beacon 106 is 10 dB.
- the UAV 102 may determine that it must navigate to a position at which pilot signal for the first beacon 106 is measured at 100 dB. Accordingly, the UAV 102 navigates while simultaneously monitoring for an increase in signal strength of the pilot signal associated with the first beacon 106 .
- the UAV has access to an initial heading or bearing when departing the base station 104 .
- the UAV 102 has navigated to a location adjacent to (e.g., directly above) the first beacon 106 .
- the UAV 102 measures the strength of the first beacon's pilot signal to be 100 dB.
- the UAV 102 can determine that it is located at the first beacon 106 by comparing the measured pilot signal strength to a list which compares distance and signal strength value, or by using one or more algorithms to convert the measured pilot signal strength into a distance.
- the UAV 102 can use the altimeter 112 to ensure that its altitude is greater than a known altitude of the first beacon 106 .
- the RFID beacon is mounted to a ceiling of the monitored environment, the UAV 102 may navigate to a position directly underneath the beacon.
- obstacle avoidance can be utilized during navigation without a need for pre-programmed signal altitudes.
- a monitoring pass refers to a predetermined pattern of movement undertaken by the UAV 102 upon reaching a location (e.g., upon navigating to a position above the first beacon 106 ). For example, initiating a monitoring pass could cause the UAV 102 to navigate in a pattern of movement that resembles a cyclone by causing the UAV 102 to navigate downward in a slowly descending spiral of movement, taking periodic measurements (e.g., every two seconds) along the way.
- the UAV 102 can initiate a flight path that resembles a horizontal spiral while maintaining a constant altitude. The UAV 102 can take periodic measurements while traveling along this flight path and, once the spiral has been completed, the UAV 102 can move to a different altitude and can begin a second horizontal spiral flight path.
- the UAV 102 After reaching the first monitoring pass position, the UAV 102 generates the report 132 , and transmits the report 134 to the base station 104 .
- the report 132 includes the beacon ID (e.g., a beacon ID associated with the first beacon 106 ); an altitude of the UAV 102 ; a distance from the beacon identified by the recorded beacon ID; a bearing of the UAV 102 relative to the beacon identified by the recorded beacon ID; and any recorded environmental information (e.g., one or more temperature and/or humidity measurements).
- the report 132 , 134 , and/or 136 can be a batch of reports.
- the UAV 102 may generate a batch of reports at each beacon. The batches of reports can be combined into fewer reports, such as a single report.
- the first and second monitoring pass positions do not need to be predefined positions, and can instead be defined by the location of the UAV 102 after navigating according to a predetermined patter of flight, or after the UAV 102 has navigated for a predefined amount of time.
- the UAV 102 can navigate to the second beacon 108 , can continue to a third monitoring pass position for the first beacon 106 , or can return to dock at the base station 104 for recharging.
- the UAV 102 can be configured to dock with the base station in such a way that terminals on its battery are aligned with charging terminals associated with the base station 104 .
- the UAV 102 can use a pilot signal transmitted by the base station 104 to hone in on the base station 104 in the same way that it locates and navigates to the beacons.
- FIG. 2 shows a process 200 for collecting environmental information using a vehicle.
- a vehicle determines that the vehicle is located within a predetermined range of a beacon ( 202 ).
- the UAV 102 can use a pilot signal strength transmitted by a beacon to determine its proximity to the beacon.
- the UAV 102 can be configured to navigate to a position within a threshold distance from the beacon, such as a position directly above the beacon, or a position several feet away from the beacon.
- the UAV 102 is configured to travel to a first beacon based on a predefined navigational plan that specifies an order in which the beacons should be visited.
- the beacons can be referenced by their unique beacon IDs in the navigational plan and may, for example, instruct the UAV 102 to “navigate to beacon 1 ” as a first waypoint.
- the pilot signal broadcast by each beacon indicates the beacon ID of its source beacon so that the UAV 102 is able to associate the pilot signal strength with a specific beacon.
- the UAV 102 may receive more than one pilot signal at a time, as the broadcast cells of multiple beacons may overlap.
- the vehicle is navigated to a first location based on determining that the vehicle is located with the predetermined range ( 204 ). For example, upon determining that the UAV 102 is located directly above the beacon, the UAV 102 begins a monitoring pass and navigates to a first monitoring pass position.
- the first monitoring pass position is defined by a distance from and a bearing relative to the beacon.
- the UAV 102 While navigating, the UAV 102 avoids obstacles using the proximity sensor 118 .
- the proximity sensor is an infrared or ultrasonic sensor that emits infrared or ultrasonic energy and then detects energy reflected from various surfaces proximal to the UAV 102 .
- the UAV 102 can use the proximity sensor 118 to evade obstacles such as, for example, machine racks in a data center, people working in the data center, and/or other UAVs.
- the proximity sensor 118 can further ensure that the UAV 102 is able to navigate safely.
- the proximity sensor 118 also ensures that the UAV 102 can detect dynamic obstacles that may have been introduced after the creation of the monitoring pass pattern of moment (e.g., if a new piece of equipment is added to a data center).
- a first report is generated based on determining that the vehicle is located at the first location ( 206 ). For example, upon reaching the first monitoring pass position, the UAV 102 generates a report using the report generation engine 120 .
- generating the report includes specifying the beacon identification, recording navigation data that includes an altitude, a distance from the beacon, and a bearing relative to the beacon, and recording environmental information.
- the navigation data can be determined using a combination of the altimeter 112 , the compass 116 , and the RFID sensor 110 .
- One or more processing devices and/or applications associated with the UAV 102 may process some of the raw data obtained from the altimeter 112 , the compass 116 , and the RFID sensor 110 in order to provide the desired values.
- a processing device may convert a pilot signal strength in decibels detected by the RFID sensor 110 into a distance in feet or meters.
- the UAV 102 While in some examples the UAV 102 generates a first report after determining that a first location has been reached, the UAV 102 may also generate and transmit a similar report before initiating a monitoring pass, such as upon determining that the UAV 102 is within a predetermined range of the first beacon 106 .
- the UAV 102 records environmental information using one or more of the thermometer 124 , the humidity sensor 126 , and any other sensors that can be used to detect environmental conditions.
- one or more processing devices and/or applications associated with the UAV 102 may process some of the raw data obtained from the environmental sensors (e.g., the humidity sensor 126 ) in order to provide the desired values.
- the UAV 102 can be configured to re-measure environmental conditions that are determined to be abnormal. For example, if a measured environmental condition appears to be in error (e.g., a value that is unattainable) or is determined to lie be outside of a normal range (e.g., a temperature that is higher than an average or a desirable data center temperature), the UAV 102 can automatically re-measure the environmental conditions.
- the report data can be uploaded to a computer that stores the data in a database or file for future use.
- the first report is transmitted from the vehicle to a base station ( 208 ).
- the UAV 102 can use the transmission device 122 to transmit the generated report to the base station 104 or to a second, different receiving station.
- the first report includes the beacon identification (e.g., “beacon 1 ”), the navigation data (e.g., an altitude of the UAV 102 , a distance of the UAV 102 from the beacon, and a bearing of the UAV 102 relative to the beacon), and the environmental information (e.g., a temperature and a level of humidity).
- the data can also be transmitted dynamically. For example, instead delaying transmission until the beacon ID, the navigation data, and the environmental information have been acquired, the UAV 102 can also transmit any or all of the data in a piecemeal fashion (e.g., the data can be transmitted as it is received).
- the UAV 102 can delay transferring the data to a central entity (e.g., the base station 104 ) until monitoring has been completed for a single beacon, for multiple beacons, or for the entire monitored environment.
- a central entity e.g., the base station 104
- the UAV 102 could store the data and/or the generated report onboard and could transfer the same to a central entity upon docking at the base station 104 .
- the UAV 102 could also be configured to only transmit reports if any environmental conditions are abnormal, as described above.
- the UAV 102 can take one or more subsequent actions. For example, the UAV 102 can move on to a second monitoring pass position to generate a second report for the same beacon. The UAV 102 could also navigate to a second beacon (e.g., second beacon 108 ) and begin monitoring environmental conditions in the vicinity of the second beacon. Furthermore, if all monitoring passes have been completed or if the UAV 102 is recalled, the UAV 102 can return to dock and recharge at the base station 104 . In some examples, the UAV 102 may monitor read its battery capacity (e.g., periodically, after the generation of one or more reports, after visiting one or more beacons, etc). The UAV 102 may choose to abort further activity and return to base to charge depending on the remaining capacity in its battery.
- the battery capacity e.g., periodically, after the generation of one or more reports, after visiting one or more beacons, etc. The UAV 102 may choose to abort further activity and return to base to charge depending on the remaining capacity in its battery.
- FIG. 3 shows a system 300 that includes the UAV 102 and the base station 104 .
- This example illustrates how the base station 104 and the UAV 102 can utilize a gravity-assisted design to dock the UAV 102 at the base station 104 .
- the UAV 102 includes a base portion 132 that is shaped to mate with a cavity 301 defined by a body of the base station 104 , even if the UAV 102 fails to land without precisely aligning itself with the cavity 301 .
- the base portion 132 of the UAV 102 is shaped to allow the base portion 132 of the UAV 102 to slide into alignment with the cavity 301 of the base station 104 .
- the edges of the base portion 132 and the housing that defines the cavity 301 are sloped, if the UAV 102 lands in a position that is slightly off-center (shown as position “ 1 ” in FIG. 3 ), the UAV 102 will slide into the cavity 301 until coming to rest in a centrally-aligned position (shown as position “ 2 ” in FIG. 3 ).
- the base portion 132 ultimately aligns with the cavity 301 allows the battery 109 of the UAV 102 to be charged by a charging unit 302 associated with the base station 104 . Because the UAV 102 is meant to operate without assistance from a human operator in some instances, the gravity-assisted landing techniques shown in FIG. 3 can reduce the instances in which the UAV 102 is called upon to perform environmental monitoring but is unable to perform due to a depleted battery.
- positive and negative terminals 304 , 306 of the battery 109 will be aligned with (and will be in electrical communication with) positive and negative terminals 308 , 310 of the charging unit 302 .
- the positive and negative terminals 308 , 310 of the base station 104 are shaped in a ring, such that the base portion 132 of the UAV 102 can dock at the base station 104 without regard to its rotation relative to the base station.
- the terminals 308 , 310 are rings that surround a substantial portion (or the entire portion) of the cavity 301 , the terminals 304 , 306 of the battery 109 may contact the terminals 308 , 310 of the base station 104 regardless of the degree of rotation of the base portion 312 relative to the base station 104 . Furthermore, this concentric ring terminal contact design may help prevent accidentally inverting the positive and negative terminals which could cause harm to the battery 109 , the UAV 102 and/or the base station 104 .
- the charging unit 302 can be configured to begin charging the battery 109 upon sensing that an electrical connection has been completed between the battery 109 and the charging unit 302 .
- the charging unit 302 may also reduce the amount of electrical current provided to the battery 109 as the battery 109 reaches various levels of charge, and may cease providing current to the battery 109 entirely when the battery reaches a fully-charged state.
- the charging unit 302 can also be configured to provide a “trickle charge” to the battery 109 if, for example, the battery 109 is to be maintained in a fully charged state for an extended period of time.
- the base station 104 can also be configured to instruct the UAV 102 to take off and re-attempt docking with the base station 104 if no electrical connection between the battery 109 and the charging unit 302 can be established (e.g., if the UAV 102 lands so poorly that even the gravity-assisted design cannot properly align the terminals 304 , 306 with the terminals 308 , 310 ). In the event that the UAV 102 is unable to return to the base station 104 on its own, the UAV 102 may transmit a request to an operator for manual intervention.
- the base station 104 also includes a broadcasting unit 312 for broadcasting the RFID pilot signal used by the UAV 102 to navigate throughout the monitored environment.
- the broadcasting unit 312 is an omnidirectional transmitter that is configured to broadcast the base station's pilot signal on a specified frequency and/or channel.
- the pilot signal includes an identity of the base station 104 or beacon (e.g., “beacon 0 ”).
- Other beacons may include similar broadcasting units.
- the vehicle has been described as a UAV such as a quad-rotor rotary wing aircraft, the techniques described above could be implemented on a variety of other vehicles.
- one or more track-guided vehicles could be used to measure environmental conditions using techniques similar to those described above.
- FIG. 4 shows a possible test environment 400 which, in this example, is a data center 401 .
- the data center includes a first row of machines (e.g., racks of equipment, free-standing equipment, etc.) 402 and a second row of machines 404 .
- the environmental conditions in the data center 401 are monitored using the techniques described above. For example, four beacons (beacons 1 , 2 , 3 , and 4 ) are located at various points within the data center 401 (e.g., at the ends of the first row of machines 402 and the ends of the second row of machines 404 ).
- Each beacon has a respective coverage cell in which its pilot signal is broadcasted.
- beacon 1 has a coverage cell 408 and beacon 2 has a coverage cell 410 .
- the coverage cells can be increased or decreased by increasing or decreasing the transmission power for the pilot signals, respectively.
- the coverage cells 408 , 410 also define shared coverage areas 406 .
- the first coverage cell 408 overlaps the second coverage cell 410 to define the shared coverage area 406 ; that is, if the UAV 102 is located in the shared coverage area 406 , its RFID sensor 110 could detect pilot signals broadcasted by both beacon 1 and beacon 2 .
- the size or number of shared coverage areas are a matter of design choice; however, having at least one shared coverage area within each coverage cell can enhance the UAV's ability to locate the next beacon in its flight plan.
- the UAV 102 may automatically navigate to other locations until it is able to detect the pilot signal from the beacon it should next visit.
- the UAV 102 can be configured to navigate in a predefined direction (e.g., the UAV 102 can rely on navigational hints that have been preloaded onto the UAV 102 ) or can fly at random until the appropriate pilot signal is detected.
- the navigational hints can include a predetermined direction that the UAV 102 should fly in if a signal is lost (e.g., toward the base station 104 , toward the last-visited beacon, etc.).
- the data center 401 also includes the base station 104 .
- the UAV 102 can return to the base station 104 to recharge its battery 109 .
- the UAV 102 can also be recalled to the base station 104 upon the occurrence of an event. For example, if the UAV 102 or the base station 104 detects that the battery 109 is running out of energy while the UAV 102 is deployed, the UAV 102 can return to the base station 104 , even if the UAV 102 has not finished monitoring the data center 401 .
- the UAV 102 may use additional techniques to navigate within the monitored environment. For example, in some situations, RFID-based navigation may not provide a high enough level of accuracy, and greater precision may be desired.
- the UAV 102 may further include one or more infrared (IR) receivers for receiving IR signals from one or more beacons and base stations within the monitored environment.
- IR infrared
- beacons or base stations may include respective IR transmitters that provide an IR signal encoded to a predetermined frequency (e.g., a frequency that will avoid interference with other light sources within the monitored environment, such as ceiling lights and lights on the monitored electronic equipment).
- IR transmitters and receivers can be used to guide the UAV 102 with a greater degree of precision than an RFID-based navigation system.
- the UAV 102 may use IR equipment in combination with the RFID-based navigation system discussed above to efficiently and accurately navigate between various points within a monitored environment. For example, the UAV 102 may acquire the RFID signal broadcast by a beacon from a relatively long distance, and may navigate toward that beacon using the RFID signal. In some examples, once the UAV 102 has arrived within a threshold distance of the beacon, has obtained a line-of-sight with the IR signal source on the beacon, or is otherwise able to detect an IR signal being transmitted by the beacon, the UAV 102 may use the IR signal generated by the beacon to fine-tune its navigation.
- the UAV may use an RFID signal to navigate to within five feet of a beacon, and may then acquire and use an IR signal generated by the beacon in order to move into a position directly above the beacon (e.g., a position that is within six inches of a desired position).
- the level of accuracy in the navigation of the UAV 102 can depend on adjustable tolerances for position sensing, and may also depend on the quality of the navigational equipment used by the UAV 102 , the beacons, the base station, and any other monitoring equipment.
- the UAV 102 can also use one or more cameras (e.g., the horizon camera 130 ( FIG. 1 )) to obtain further information about the surrounding environment.
- the UAV 102 can use the horizon camera 130 or one or more additional cameras to provide visual feedback of the monitored environment.
- the visual feedback can include the collection and storage of videos, images, or a combination thereof.
- the UAV 102 can be configured to use one or more cameras to capture video or images of monitored equipment that has been determined to be problematic.
- the UAV 102 can begin capturing visual data upon the detection of an abnormal environmental condition, upon receiving a control signal from a user (e.g., via a user interface that is in communication with the UAV 102 ), or may periodically capture visual data according to a predefined schedule (e.g., once per day).
- the additional visual data collected by the UAV 102 can be transmitted to the base station 104 to be analyzed by an operator, or can be automatically analyzed using one or more image recognition applications.
- an image recognition application may examine an image of a machine in a datacenter to determine whether one or more warning lights have been activated on the machine.
- the image recognition application may compare the received visual information to images or videos of the subject of the visual information operating in a normal state.
- a notification e.g., an alert or alarm
- the image recognition software can be implemented on the UAV 102 to facilitate the real-time monitoring of visual data.
- FIG. 5 shows an example of a user interface 500 for interacting with the data captured by the UAV 102 .
- the user interface 500 can be an environmental map of a data center that contains a first server rack row 502 and a second server rack row 504 ; that is, the user interface 500 can be a visual representation of monitored environment.
- the user interface 500 can include a number of tools, graphics, and other features that allow users to monitor the environmental conditions of a given location.
- the user interface 500 presents a number of environmental graphics, such as temperature graphic 508 , which graphically represent the temperature of a monitored point (or group of points) within the monitored environment.
- the user interface 500 includes controls, such as view control 512 and detail control 514 , that allow a user to specify which information the environmental graphics will be associated with.
- the environmental graphics represent the temperature recorded at the “low altitude” (e.g., the lowest of three measured altitudes) for one or more monitored locations represented by each environmental graphic, and the environmental graphics are both color-coded according to a legend 510 and display a numerical temperature.
- the temperature graphic 508 indicates that the temperature recorded at the low location was 103 degrees, which is demonstrated by the numerical value within the temperature graphic 508 . Furthermore, the temperature graphic 508 is colored red which, according to the legend 510 , indicates that the temperature of the location associated with the temperature graphic 508 is over 100 degrees.
- the user interface 500 may also include a graphical representation of a base station 506 , as well as graphical representations that show the locations of one or more beacons within the monitored environment (e.g., beacons 1 , 2 , 3 , 4 , 5 , and 6 ).
- the view control 512 can be manipulated to cause the user interface 500 to display various types of information. For example, activating the view control 512 can cause the user interface 500 to display a high temperature (e.g., the temperature measured at the “high” location), a middle temperature (e.g., the temperature measured at the “middle” location) to show thermal layers based on altitude within the environment, and/or the measured humidity.
- the user interface 500 can also be configured to display alerts that are related to the monitored environment. For example, if the temperature for a given location is measured to be above a threshold value, an alarm can be triggered (e.g., an audible alarm, and/or an email alert or other notification).
- certain conditions may trigger responsive actions, such as the shutdown of a machine or group of machines that are operating in a region associated with an unfavorable environmental condition (e.g., high heat and/or high humidity).
- FIG. 6 shows an example data center 600 that includes machine racks 602 - 604 , the UAV 102 , the base station 104 , and a backend processing engine 606 . While in this example (and in some examples that follow) the backend processing engine 606 is shown as being located within the data center 600 , the backend processing engine 606 can also be located outside the data center 600 . For example, the backend processing engine 606 (or one or more additional backend processing engines) may be located remotely relative to the data center 600 , and may communicate with, for example, the base station 104 over a network (e.g., the Internet). In some examples, some or all of the functions of the backend processing engine 606 can be provided by the base station 104 , the UAV 102 , or some combination thereof. In some examples, the UAV 102 can communicate directly with the backend processing engine 606 .
- the backend processing engine 606 may be located remotely relative to the data center 600 , and may communicate with, for example, the base station 104 over
- the backend processing engine 606 uses data collected by the UAV 102 to monitor conditions of, and effect changes in, the data center 606 .
- the backend processing engine 606 may also use data collected by the UAV 102 to generate reports, or visualizations of the conditions within the data center 600 .
- the backend processing engine 606 includes an event engine 608 and a response engine 610 .
- the event engine 608 can receive data (e.g., reports provided by the UAV 102 ) and can analyze the received data to determine whether one or more events have occurred.
- the event engine 608 can detect and identify a variety of events, examples of which will be described if greater detail below.
- the event engine 608 can provide an indication that an event has occurred to the response engine 610 .
- the response engine 610 may take one or more predefined actions based on the manner of event identified by the event engine 608 . As a simplified example, if the event engine 608 determines that the data center 600 is too warm (e.g., if the temperature of the data center 600 detected by the UAV 102 exceeds a threshold temperature) and sends an event notification to the response engine 610 , the response engine 610 could take an action to, for example, provide additional cooling to one or more portions of the data center 600 .
- the response engine 10 comprises one or more response rules that can be applied to event notifications received from the event engine.
- FIG. 7 shows a data center 700 that includes machine racks 702 - 705 , the UAV 102 , the base station 104 , and the backend processing engine 606 .
- the UAV 102 collects environmental information (e.g., temperature data) within the data center 700 .
- the UAV 102 may collect the environmental information using, for example, the collection and monitoring techniques described above.
- FIG. 7 shows an example of how the backend processing engine 606 may use data collected by the UAV 102 to monitor conditions within the data center 700 , and further shows a possible corrective response that can be executed based on the collected data.
- location information e.g., location information
- other environmental information e.g., other temperature measurements of the machine rack 705 or other machine racks, humidity data, etc.
- the base station 104 collects the report that includes the first temperature 708 from the UAV 102 and provides the first temperature 708 to the backend processing engine 606 .
- the base station 104 may provide information to the backend processing engine 606 as the information is received at the base station 104 , or may provide information to the backend processing engine 606 in batches (e.g., after the base station 104 receives a predetermined amount of data or reports).
- the event engine 608 of the backend processing engine 606 can analyze the information provided by the base station 104 (including the first temperature 708 ) to determine whether an event has occurred. For example, the backend processing engine 606 may analyze the first temperature 708 and, as a result, may determine that the first temperature 708 exceeds a threshold temperature 712 .
- the threshold temperature 712 may be associated with the data center 700 as a whole, or may be associated with a specific area of the data center 700 (e.g., an area in the vicinity of the machine rack 705 ). If the first temperature 708 exceeds the threshold temperature 712 , the event engine 608 may determine that an event has occurred, and may provide an event notification 715 to the response engine 610 .
- the response engine 710 may execute one or more predetermined responses 716 .
- the response engine 610 may determine that because the first temperature 708 exceeds the threshold temperature 712 (e.g., by a predetermined margin), the predetermine response 716 should be executed to lower the first temperature 708 .
- the one or more predetermined responses 716 include a response that causes the backend processing engine 606 to activate a cooling system 706 associated with the data center 600 .
- activating the cooling system 706 includes altering one or more data center chillers (e.g., by altering an air temperature level, a fan speed, or other setting of the chiller). While some data center chillers include temperature sensors disposed on or near the chiller itself, allowing the UAV 102 to act as a remote temperature sensor can increase the efficiency of the cooling system 706 by allowing customized temperature readings to replace or supplement the chiller's on-board sensor. In this way, the cooling system 706 can begin reducing a temperature of the data center 700 even if the area of high temperature is located relatively far from the chiller's onboard temperature sensor, which might otherwise delay the activation of the cooling system 706 .
- altering one or more data center chillers e.g., by altering an air temperature level, a fan speed, or other setting of the chiller. While some data center chillers include temperature sensors disposed on or near the chiller itself, allowing the UAV 102 to act as a remote temperature sensor can increase the efficiency of the cooling system 706 by allowing customized
- the one or more predetermined responses 716 may include a response that causes the backend processing engine 606 to adjust adjustment an air duct 707 within the data center 700 provide additional cooling to the area associated with the first temperature 708 .
- the response engine 610 may adjust one or more vents to direct cooling air toward the machine rack 705 .
- the air duct 707 may be associated with the cooling system 706 , or may be a standalone aspect of a cooling solution of the data center 700 .
- the UAV 102 may report the second temperature 710 to the backend processing engine 606 , and similar evaluations can be performed to determine whether any events have occurred (e.g., whether the second temperature exceeds the threshold temperature 714 ) and/or whether additional corrective actions should be executed.
- FIG. 8 shows an example predetermined response 716 that can be executed by the response engine 710 .
- the response engine 610 can execute one of the predetermined responses 716 to provide a notification 802 .
- providing the notification 802 can include providing a visual alert 804 on a display 806 associated with the backend processing engine 606 .
- the notification 802 can also be provided as an audible alarm.
- alerts such as the visual alert 804 can be configured to notify an operator of the backend processing engine 606 of the occurrence of an event.
- the visual alert 804 includes a warning related to the elevated temperature near machine rack 705 .
- the notification may also provide one or more suggested corrective actions that can be selected by an operator to address the event identified by the event notification 715 .
- the suggested corrective actions may allow an operator to activate the cooling system 706 or adjust the vents of the air duct 707 .
- FIG. 9 shows a data center 900 that includes the machine racks 702 - 705 , the UAV 102 , the base station 104 , and the backend processing engine 606 .
- the UAV has collected environmental information within the data center 900 and has generated a report 902 that indicates the temperature detected near each of the machine racks 702 - 704 .
- the report 902 further indicates that the temperature near the machine rack 705 is not available.
- the backend processing engine 606 can determine whether an event has occurred using the event engine 608 .
- the event engine 608 determines that an event has occurred based on the absence of temperature data for machine rack 705 in the report 902 .
- the event engine 608 may determine whether the area associated with the missing environmental information is large enough to warrant a corrective action. For example, the event engine 608 may compare a size of the area associated with the missing environmental information to a threshold size (e.g., 25 square feet) and, if the area exceeds the threshold size, an event notification can be generated.
- a threshold size e.g. 25 square feet
- the event engine 608 provides an event notification 915 to the response engine 610 that indicates the type of event that has occurred. Based on receiving the event notification 915 , the response engine 610 executes one of the predetermined responses 716 . In this example, the response engine 610 may execute a predetermined response that causes updated monitoring instructions 904 to be provided to the UAV 102 which, in turn, may cause the UAV to return to the machine rack 905 to attempt to collect the missing environmental information. In some examples, the updated monitoring instructions 904 include instructions to direct the UAV 102 to the nearest RFID beacon, and may also include relative vectors from the beacon to the missed area.
- the backend processing engine 606 may mark (or may allow an operator to mark) the area as unreachable so that additional monitoring passes are not triggered.
- FIG. 10 shows a data center 1000 that includes machine racks 1002 and 1004 , the UAV 102 , the base station 104 , and the backend processing engine 606 .
- the backend processing engine 606 includes a display 1006 .
- the UAV 102 has encountered a navigation deviation near the machine rack 1004 due to the open cabinet door 1008 of the machine rack 1004 .
- the UAV 102 provides a report 1001 to the backend processing engine 606 (e.g., via the base station 104 ) that identifies the navigation deviation and the location of the UAV 102 when the navigation deviation was encountered.
- the event engine 606 analyzes the report 1001 to determine that an event has occurred (e.g., to determine that the cabinet door 1008 is ajar). Based on determining that an event has occurred, the event engine 608 provides an event notification 1015 to the response engine 610 .
- the response engine 610 uses the event notification 1015 to select one or more of the predetermined responses 716 to execute to correct the problem that triggered the event notification 1015 .
- the response engine 610 causes a notification 1004 to be displayed on the display 1006 associated with the backend processing engine 606 .
- the notification 1004 warns an operator that the cabinet door 1008 has been left ajar.
- the notification 1004 could also be in the form of a pop-up window, an automated email, an SMS text message, or an audible alert.
- Cabinet doors such as the cabinet door 1008 , can provide a level of physical security within the data center 1000 . This is especially important in multi-tenant and high-security data center environments. As a result, keeping the cabinet door 1008 closed can also help promote effective security, and may also improve an airflow within the data center 100 .
- Other events may also trigger one or more predetermined responses.
- the techniques described above can be used to address both real and simulated cooling system failures within a data center.
- other cooling devices can operate at a higher capacity to provide the necessary coverage (e.g., to maintain appropriate temperature/humidity levels).
- the UAV 102 and the back end processing provided by the backend processing engine 606 can be used to efficiently calibrate the cooling systems within a data center to maintain desired temperature levels.
- a response engine may automatically adjust other cooling devices within the data center to operate at a higher capacity. After the other cooling devices have been adjusted, the UAV 102 may continue to monitor the environmental conditions within the data center to assess both the effects of the failed cooling device and the results of the adjusted devices.
- the backend processing engine 606 may cause the UAV 102 to perform additional or modified monitoring passes to increase the frequency of data capture associated with the floor space near the failed cooling device. After the failed cooling device is restored, the backend processing engine 606 can automatically return the settings of all the adjusted cooling devices back to normal levels, and may confirm that the temperatures within the data center stay within expected ranges.
- cooling devices may be associated with application program interfaces (APIs) that support input from external sources using various protocols.
- APIs application program interfaces
- older cooling devices that do not have a published interface or network connectivity can be retrofitted with analog interfaces in the form of, for example, a microcontroller platform with a network interface which provides these APIs (e.g., an chicken-based device).
- APIs application program interfaces
- other cooling devices' thermostats can be set to a lower temperature to compensate for the unavailable cooling device.
- Cooling system failures such as that described above can also be simulated using the UAV 102 and its associated back end processing.
- simulated cooling system failures can be created by disabling one or more cooling devices at a time and measuring the impact of the device's failure on the data center. After the results of the simulated failures are captured, a failure response plan can be generated that specifies adjustments for other cooling devices that should be implemented if a given cooling device fails. As a result, in the event of an actual cooling system failure, the response engine could simply execute the stored failure response plan to adjust the remaining active cooling devices.
- the failure response plan may include various options.
- the failure response plan may allow the backend processing engine 606 or an operator to select from varying response levels in the event of a device failure (e.g., maximum cooling, minimum cooling, or minimum power use).
- a thermostat can be adjusted to a high temperature (e.g., to prevent its fan from providing cooling air). A thermostat can then be adjusted to a lower temperature to activate its fans. After a predetermined amount of time (e.g., a few minutes_, the UAV 102 can be launched to observe the temperature pattern and to determine the impact of the simulated cooling device failure and response.
- the backend processing engine 606 can be configured to perform historical analysis of the environmental information collected by the UAV 102 .
- the backend processing engine 606 can use a historical record of the collected environmental information to optimize conditions within a data center. For example, the backend processing engine 606 may accumulate a batch of environmental information for a predetermined length of time (e.g., thirty days), and may then evaluate the batch to provide optimization recommendations.
- the backend processing engine 606 may use historical environmental information to identify areas within a data center in which temperatures or humidity levels are consistently above or below desired levels. As a result, the backend processing engine 606 may recommend options for relocating equipment (e.g., machine racks) within the data center to correct the areas of concern.
- the backend processing engine 606 may also provide an indication that a mechanical failure has possibly occurred (e.g., that a fan in a device is disabled and/or not configured properly).
- the backend processing engine 606 may determine, based on the low, medium, and high altitude passes discussed above, a particular region of a machine rack that stores to determine where within the machine rack the offending equipment, or areas of machine racks that have available capacity (e.g., empty machine slots within a machine rack). Additionally, the results of the historical analysis can be used to identify opportunities to adjust air ducts or cooling system vents direct to more efficiently direct cooling air and to more effectively exhaust warm air from the data center.
- FIG. 11 shows an example of a computing device 1100 and a mobile computing device 1150 that can be used to implement the techniques described in this disclosure.
- the computing device 1100 is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers.
- the mobile computing device 1150 is intended to represent various forms of mobile devices, such as personal digital assistants, cellular telephones, smart-phones, and other similar computing devices.
- the components shown here, their connections and relationships, and their functions, are meant to be examples only, and are not meant to be limiting.
- the computing device 1100 includes a processor 1102 , a memory 1104 , a storage device 1106 , a high-speed interface 1108 connecting to the memory 1104 and multiple high-speed expansion ports 1110 , and a low-speed interface 1112 connecting to a low-speed expansion port 1114 and the storage device 1106 .
- Each of the processor 1102 , the memory 1104 , the storage device 1106 , the high-speed interface 1108 , the high-speed expansion ports 1110 , and the low-speed interface 1112 are interconnected using various busses, and may be mounted on a common motherboard or in other manners as appropriate.
- the processor 1102 can process instructions for execution within the computing device 1100 , including instructions stored in the memory 1104 or on the storage device 1106 to display graphical information for a GUI on an external input/output device, such as a display 1116 coupled to the high-speed interface 1108 .
- an external input/output device such as a display 1116 coupled to the high-speed interface 1108 .
- multiple processors and/or multiple buses may be used, as appropriate, along with multiple memories and types of memory.
- multiple computing devices may be connected, with each device providing portions of the necessary operations (e.g., as a server bank, a group of blade servers, or a multi-processor system).
- the memory 1104 stores information within the computing device 1100 .
- the memory 1104 is a volatile memory unit or units.
- the memory 1104 is a non-volatile memory unit or units.
- the memory 1104 may also be another form of computer-readable medium, such as a magnetic or optical disk.
- the storage device 1106 is capable of providing mass storage for the computing device 1100 .
- the storage device 1106 may be or contain a computer-readable medium, such as a floppy disk device, a hard disk device, an optical disk device, or a tape device, a flash memory or other similar solid state memory device, or an array of devices, including devices in a storage area network or other configurations.
- Instructions can be stored in an information carrier.
- the instructions when executed by one or more processing devices (for example, processor 1102 ), perform one or more methods, such as those described above.
- the instructions can also be stored by one or more storage devices such as computer- or machine-readable mediums (for example, the memory 1104 , the storage device 1106 , or memory on the processor 1102 ).
- the high-speed interface 1108 manages bandwidth-intensive operations for the computing device 1100 , while the low-speed interface 1112 manages lower bandwidth-intensive operations. Such allocation of functions is an example only.
- the high-speed interface 1108 is coupled to the memory 1104 , the display 1116 (e.g., through a graphics processor or accelerator), and to the high-speed expansion ports 1110 , which may accept various expansion cards (not shown).
- the low-speed interface 1112 is coupled to the storage device 1106 and the low-speed expansion port 1114 .
- the low-speed expansion port 1114 which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices, such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter.
- input/output devices such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter.
- the computing device 1100 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a standard server 1120 , or multiple times in a group of such servers. In addition, it may be implemented in a personal computer such as a laptop computer 1122 . It may also be implemented as part of a rack server system 1124 . Alternatively, components from the computing device 1100 may be combined with other components in a mobile device (not shown), such as a mobile computing device 1150 . Each of such devices may contain one or more of the computing device 1100 and the mobile computing device 1150 , and an entire system may be made up of multiple computing devices communicating with each other.
- the mobile computing device 1150 includes a processor 1152 , a memory 1164 , an input/output device such as a display 1154 , a communication interface 1166 , and a transceiver 1168 , among other components.
- the mobile computing device 1150 may also be provided with a storage device, such as a micro-drive or other device, to provide additional storage.
- a storage device such as a micro-drive or other device, to provide additional storage.
- Each of the processor 1152 , the memory 1164 , the display 1154 , the communication interface 1166 , and the transceiver 1168 are interconnected using various buses, and several of the components may be mounted on a common motherboard or in other manners as appropriate.
- the processor 1152 can execute instructions within the mobile computing device 1150 , including instructions stored in the memory 1164 .
- the processor 1152 may be implemented as a chipset of chips that include separate and multiple analog and digital processors.
- the processor 1152 may provide, for example, for coordination of the other components of the mobile computing device 1150 , such as control of user interfaces, applications run by the mobile computing device 1150 , and wireless communication by the mobile computing device 1150 .
- the processor 1152 may communicate with a user through a control interface 1158 and a display interface 1156 coupled to the display 1154 .
- the display 1154 may be, for example, a TFT (Thin-Film-Transistor Liquid Crystal Display) display or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology.
- the display interface 1156 may comprise appropriate circuitry for driving the display 1154 to present graphical and other information to a user.
- the control interface 1158 may receive commands from a user and convert them for submission to the processor 1152 .
- an external interface 1162 may provide communication with the processor 1152 , so as to enable near area communication of the mobile computing device 1150 with other devices.
- the external interface 1162 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used.
- the memory 1164 stores information within the mobile computing device 1150 .
- the memory 1164 can be implemented as one or more of a computer-readable medium or media, a volatile memory unit or units, or a non-volatile memory unit or units.
- An expansion memory 1174 may also be provided and connected to the mobile computing device 1150 through an expansion interface 1172 , which may include, for example, a SIMM (Single In Line Memory Module) card interface.
- SIMM Single In Line Memory Module
- the expansion memory 1174 may provide extra storage space for the mobile computing device 1150 , or may also store applications or other information for the mobile computing device 1150 .
- the expansion memory 1174 may include instructions to carry out or supplement the processes described above, and may include secure information also.
- the expansion memory 1174 may be provide as a security module for the mobile computing device 1150 , and may be programmed with instructions that permit secure use of the mobile computing device 1150 .
- secure applications may be provided via the SIMM cards, along with additional information, such as placing identifying information on the SIMM card in a non-hackable manner.
- the memory may include, for example, flash memory and/or NVRAM memory (non-volatile random access memory), as discussed below.
- instructions are stored in an information carrier. that the instructions, when executed by one or more processing devices (for example, processor 1152 ), perform one or more methods, such as those described above.
- the instructions can also be stored by one or more storage devices, such as one or more computer- or machine-readable mediums (for example, the memory 1164 , the expansion memory 1174 , or memory on the processor 1152 ).
- the instructions can be received in a propagated signal, for example, over the transceiver 1168 or the external interface 1162 .
- the mobile computing device 1150 may communicate wirelessly through the communication interface 1166 , which may include digital signal processing circuitry where necessary.
- the communication interface 1166 may provide for communications under various modes or protocols, such as GSM voice calls (Global System for Mobile communications), SMS (Short Message Service), EMS (Enhanced Messaging Service), or MMS messaging (Multimedia Messaging Service), CDMA (code division multiple access), TDMA (time division multiple access), PDC (Personal Digital Cellular), WCDMA (Wideband Code Division Multiple Access), CDMA2000, or GPRS (General Packet Radio Service), among others.
- GSM voice calls Global System for Mobile communications
- SMS Short Message Service
- EMS Enhanced Messaging Service
- MMS messaging Multimedia Messaging Service
- CDMA code division multiple access
- TDMA time division multiple access
- PDC Personal Digital Cellular
- WCDMA Wideband Code Division Multiple Access
- CDMA2000 Code Division Multiple Access
- GPRS General Packet Radio Service
- a GPS (Global Positioning System) receiver module 1170 may provide additional navigation- and location-related wireless data to the mobile computing device 1150 , which may be used as appropriate by applications running on the mobile computing device 1150 .
- the mobile computing device 1150 may also communicate audibly using an audio codec 1160 , which may receive spoken information from a user and convert it to usable digital information.
- the audio codec 1160 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of the mobile computing device 1150 .
- Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on the mobile computing device 1150 .
- the mobile computing device 1150 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as a cellular telephone 1180 . It may also be implemented as part of a smart-phone 1182 , personal digital assistant, or other similar mobile device.
- implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, specially designed ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof.
- ASICs application specific integrated circuits
- These various implementations can include implementation in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, coupled to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
- machine-readable medium and computer-readable medium refer to any computer program product, apparatus and/or device (e.g., magnetic discs, optical disks, memory, Programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal.
- machine-readable signal refers to any signal used to provide machine instructions and/or data to a programmable processor.
- the systems and techniques described here can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user and a keyboard and a pointing device (e.g., a mouse or a trackball) by which the user can provide input to the computer.
- a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
- a keyboard and a pointing device e.g., a mouse or a trackball
- Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user can be received in any form, including acoustic, speech, or tactile input.
- the systems and techniques described here can be implemented in a computing system that includes a back end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front end component (e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back end, middleware, or front end components.
- the components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network (LAN), a wide area network (WAN), and the Internet.
- LAN local area network
- WAN wide area network
- the Internet the global information network
- the computing system can include clients and servers.
- a client and server are generally remote from each other and typically interact through a communication network.
- the relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
- the delegate(s) may be employed by other applications implemented by one or more processors, such as an application executing on one or more servers.
- the logic flows depicted in the figures do not require the particular order shown, or sequential order, to achieve desirable results.
- other actions may be provided, or actions may be eliminated, from the described flows, and other components may be added to, or removed from, the described systems. Accordingly, other implementations are within the scope of the following claims.
Landscapes
- Engineering & Computer Science (AREA)
- Environmental & Geological Engineering (AREA)
- General Physics & Mathematics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Atmospheric Sciences (AREA)
- Biodiversity & Conservation Biology (AREA)
- Ecology (AREA)
- Environmental Sciences (AREA)
- Automation & Control Theory (AREA)
- Traffic Control Systems (AREA)
Abstract
In general, a method includes receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center, storing the environmental information on one or more memory devices that are accessible by the one or more computing devices, determining, based on the environmental information, that an event has occurred, and performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
Description
- This patent application claims the benefit of priority of U.S. Provisional Patent Application No. 61/477,377, filed Apr. 20, 2011. The content of U.S. Provisional Patent Application No. 61/477,377 is hereby incorporated by reference into this application as if set forth herein in full.
- This specification relates to capturing environmental information.
- It is often useful to monitor the environmental characteristics of certain settings. Data centers (e.g., a room containing a number of machines, such as, for example, servers, network devices and storage devices) are an example of a setting in which it is important to monitor environmental characteristics. For example, while under load, data center equipment can generate a significant amount of heat, and increasing the power density of a system can result in an increase in the number of heat concentration pockets. Combinations of air and liquid cooling solutions and monitoring techniques can be used to manage the environmental characteristics of a setting such as a data center.
- In general, in one aspect, a method includes receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center, storing the environmental information on one or more memory devices that are accessible by the one or more computing devices, determining, based on the environmental information, that an event has occurred, and performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
- Aspects may include none, one, or more of the following features. The environmental information is associated with one or more of a temperature and a humidity level of at least a portion of the data center. Determining that the event has occurred includes determining that the temperature or the humidity level associated with the at least a portion of the data center falls outside a predetermined range. Performing the one or more actions includes activating a cooling system associated with the at least a portion of the data center. Performing the one or more actions includes adjusting an air vent associated with the at least a portion of the data center. Adjusting the air vent alters a flow of air through the at least a portion of the data center.
- The one or more actions include causing the UAV to travel to the at least a portion of the data center to collect additional environmental information. It is determined that the additional environmental information is associated with one or more of a second temperature and a second humidity level that falls outside the predetermined range, and a notification is provided. Providing the notification includes providing at least one of either: an on screen pop-up message, or an automated email, or a short message service (SMS) message, or any combination thereof. Determining that the event has occurred includes determining that a cooling system has been disabled. Performing the one or more actions includes altering an operation of one or more other cooling systems, and causing the UAV to travel to one or more portions of the data center to collect additional environmental information. A frequency is altered with which the UAV collects environmental information associated with a portion of the data center that is associated with the disabled cooling system. The environmental information specifies that the UAV has experienced a navigation deviation. Determining that the event has occurred includes determining, based on the navigation deviation, that a mechanical change has occurred in the data center. Determining that the mechanical change has occurred includes determining that a door located in the data center is ajar. Determining that the event has occurred includes determining that a predetermined amount of environmental information has been received. Performing the action includes analyzing the environmental information to determine the location of one or more portions of the data center that are associated with environmental information that falls outside of a predetermined range. Determining that the event has occurred includes determining that environmental information has not been collected for a portion of the data center. It is determined that the portion of the data center is inaccessible by the UAV. It is determined that a size of the portion of the data center exceeds a predetermined size. The UAV is caused to collect environmental information for the portion of the data center based on determining that the size of the portion of the data center exceeds the predetermined size.
- Other embodiments of these aspects include corresponding systems, apparatus, and computer programs, configured to perform the actions of the methods, encoded on computer storage devices.
- The details of one or more embodiments of the subject matter described in this specification are set forth in the accompanying drawings and the description below. Other potential features, aspects, and advantages of the subject matter will become apparent from the description, the drawings, and the claims.
-
FIG. 1 is a diagram of a system for capturing environmental information. -
FIG. 2 is a flow chart of a process for capturing environmental information. -
FIG. 3 is a diagram of a base station. -
FIG. 4 is a diagram of data center. -
FIG. 5 is a diagram of user interface that displays captured environmental information. -
FIGS. 6-10 are diagrams of data centers. -
FIG. 11 is a diagram of a computing system. - Like reference symbols in the various drawings indicate like elements.
-
FIG. 1 shows asystem 100 in which a vehicle patrols an area to monitor environmental conditions of the area. In the example ofFIG. 1 , the vehicle is aUAV 102 and, specifically, is a self-charging quad-rotor UAV. In general, thesystem 100 monitors environmental conditions such as temperature, noise, pressure, and humidity, within a setting such as a data center. In some examples, the UAV 102 includes a number of sensors that are used to both navigate theUAV 102 and to record environmental information representative of environmental conditions of the environment in which the UAV 102 is operating. - The environmental information recorded by the UAV 102 can be used to locate areas of concern in a data center, such as areas or pockets of abnormally high heat, noise or humidity. The environmental information can also be used to generate visual representations, such as histograms, of the environmental conditions within the monitored environment. For example, the environmental information recorded by the
UAV 102 can be used to generate data representations (e.g., graphs and spreadsheets) that reflect a time history of the monitored environmental conditions. - Patterns can be detected in these data representations to automatically determine, for example, whether any corrective actions need to be taken. For example, if the environmental information shows that a particular location or locations within a monitored data center is abnormally hot during a particular time period each day (e.g., 9:30 AM on Mondays, when the machine might be under a heavy load), an administrator or an engine could choose to take extra temperature control measures during that time such as, for example, moving some equipment to another location to distribute the heat generation, or adjusting the airflow vents in the area to better cool the environment. Such patterns may be represented visually by a graph or chart.
- The UAV 102 includes a combination of hardware and software to both navigate the UAV 102 and to acquire and transmit environmental information to a central location (e.g., the base station 104). For example, the UAV 102 includes one or more environmental sensors, such as a
thermometer 124 and ahumidity sensor 126. These sensors are located in positions that will not be affected by the operation of the rotors 128 (e.g., above therotors 128 or in front of a frame that supports the rotors, as shown). - The UAV 102 also includes navigational features, such as an
altimeter 112, a radio-frequency identification (RFID) sensor 110, a compass 116 (e.g., an electronic compass), and aproximity sensor 118. Thecompass 116 provides a heading or bearing of the UAV 102 (e.g., by providing information that allows a relative bearing to be calculated) and can be an analog or digital compass. Thealtimeter 112 provides an altitude of theUAV 102, and can be implemented as a downward-facing infrared altimeter or an ultrasonic altimeter. Theproximity sensor 118 provides collision detection functionality using infrared or ultrasonic obstacle detection techniques. - Additional proximity sensors can be located on the UAV 102 to provide an increased range of coverage for detecting collisions and obstacles. The UAV 102 may also include a horizon detection device 130 (e.g., a camera) for stabilizing and properly orienting the
UAV 102. The RFID sensor provides a position of theUAV 102 relative to one or more beacons (e.g., the first beacon 106), as described in further detail below. The navigational features communicate with anavigation engine 114 to navigate theUAV 102. In some examples, thenavigation engine 114 is an application running on a processing device associated with theUAV 102, and uses values provided by the navigational features to navigate theUAV 102. In some embodiments, the processing device may be onboard UAV 102 and thus the navigation is performed locally. In alternative embodiments, the processing device may be located remotely fromUAV 102. In such cases,UAV 102 may send sensor data to the processing device wirelessly and may receive navigation information from the processing device also wirelessly. - The
UAV 102 also includes areport generation engine 120. In some examples, thereport generation engine 120 generates reports (e.g., report 134) that provide the environmental conditions of a particular location. Thereport generation engine 120 uses data provided by thealtimeter 112, the RFID sensor 110, thethermometer 124, thehumidity sensor 126, and thecompass 116 to generate reports that are transmitted to a central location using atransmission device 122. In some examples, thetransmission device 122 transmits reports using one or more wireless transmission protocols, such as WiFi, Bluetooth, radio communication, and the like. An example of a protocol that can be used is XBee wireless communication protocol (IEEE 802.15.4) which uses low power radio frequency at 2.4 GH. - In some examples, the
system 100 includes afirst beacon 106 and asecond beacon 108. The number of beacons included a system can depend on the range of the beacon and the size of the environment being monitored. The first and 106, 108 are configured to transmit respective pilot signals that can be detected by sensors (e.g., an RFID sensor 110) on thesecond beacons UAV 102. TheUAV 102 uses the pilot signals transmitted by the first and 106, 108 to navigate to various locations within a monitored environment. The first andsecond beacons 106, 108 can be placed at locations within a monitored environment to act as waypoints for thesecond beacons UAV 102. Each of the first and 106, 108 also transmits a Beacon ID that uniquely identifies its associated beacon.second beacons - In some examples, the
UAV 102 remains docked at abase station 104 until a monitoring session is triggered, either by a user or as part of a predefined schedule. Thus, in an initial state (referred to as “time zero,” or “t=0”), the UAV is docked at thebase station 104. While theUAV 102 is docked at thebase station 104, abattery 109 associated with theUAV 102 may be charged by a charging device (FIG. 3 ) located in thebase station 104. Terminals of thebattery 109 can be located in abase 132 of theUAV 102. - While the
UAV 102 is docked at thebase station 104 at t=0, a monitoring session can be initiated which activates the UAV 102 (e.g., powers on theUAV 102 from an OFF or “sleep” state). Based on a triggering event, theUAV 102 powers up, executes one or more self-test procedures, and begins to navigate according to a flight plan that can be stored locally on theUAV 102 or at a remote control station (e.g., thebase station 104 or a separate entity) and transmitted to theUAV 102 before or during the monitoring session. - If the flight plan specifies that the
UAV 102 will first travel toward thefirst beacon 106, theUAV 102 attempts to detect a pilot signal associated with thefirst beacon 106 using an onboard RFID sensor 110. In this example, at t=0 theUAV 102 detects a pilot signal associated with thefirst beacon 106 and determines the signal strength of that pilot signal. The signal strength can be measured in decibels (dB) or as dBm or dBmW (a power ratio in decibels of the measured power referenced to one milliwatt (mW)). - At t=0, the UAV determines that the signal strength of the pilot signal associated with the
first beacon 106 is 10 dB. In order to navigate to thefirst beacon 106, theUAV 102 may determine that it must navigate to a position at which pilot signal for thefirst beacon 106 is measured at 100 dB. Accordingly, theUAV 102 navigates while simultaneously monitoring for an increase in signal strength of the pilot signal associated with thefirst beacon 106. In some examples, the UAV has access to an initial heading or bearing when departing thebase station 104. - At t=1, the
UAV 102 has navigated to a location adjacent to (e.g., directly above) thefirst beacon 106. At this position, theUAV 102 measures the strength of the first beacon's pilot signal to be 100 dB. TheUAV 102 can determine that it is located at thefirst beacon 106 by comparing the measured pilot signal strength to a list which compares distance and signal strength value, or by using one or more algorithms to convert the measured pilot signal strength into a distance. In order to navigate to a position directly above thefirst beacon 106, theUAV 102 can use thealtimeter 112 to ensure that its altitude is greater than a known altitude of thefirst beacon 106. In some examples, if the RFID beacon is mounted to a ceiling of the monitored environment, theUAV 102 may navigate to a position directly underneath the beacon. In some examples, obstacle avoidance can be utilized during navigation without a need for pre-programmed signal altitudes. - Once the
UAV 102 determines that it is located in an appropriate position, it may initiate a monitoring pass. In some examples, a monitoring pass refers to a predetermined pattern of movement undertaken by theUAV 102 upon reaching a location (e.g., upon navigating to a position above the first beacon 106). For example, initiating a monitoring pass could cause theUAV 102 to navigate in a pattern of movement that resembles a cyclone by causing theUAV 102 to navigate downward in a slowly descending spiral of movement, taking periodic measurements (e.g., every two seconds) along the way. In some examples, theUAV 102 can initiate a flight path that resembles a horizontal spiral while maintaining a constant altitude. TheUAV 102 can take periodic measurements while traveling along this flight path and, once the spiral has been completed, theUAV 102 can move to a different altitude and can begin a second horizontal spiral flight path. - During the monitoring pass, the
UAV 102 can periodically measure environmental information using the thermometer 125, thehumidity sensor 126, and or any other sensor. For example, at t=2, theUAV 102 has initiated a monitoring pass and has reached a first monitoring pass position. - After reaching the first monitoring pass position, the
UAV 102 generates thereport 132, and transmits thereport 134 to thebase station 104. In some examples, thereport 132 includes the beacon ID (e.g., a beacon ID associated with the first beacon 106); an altitude of theUAV 102; a distance from the beacon identified by the recorded beacon ID; a bearing of theUAV 102 relative to the beacon identified by the recorded beacon ID; and any recorded environmental information (e.g., one or more temperature and/or humidity measurements). The 132, 134, and/or 136 can be a batch of reports. For example, thereport UAV 102 may generate a batch of reports at each beacon. The batches of reports can be combined into fewer reports, such as a single report. - The
UAV 102 may also generate and transmit a similar report before initiating a monitoring pass, such as upon determining that theUAV 102 is within a predetermined range of thefirst beacon 106. After generating thereport 134, theUAV 102 continues to a second monitoring pass position at t=3, where it generates asecond report 136. - The first and second monitoring pass positions do not need to be predefined positions, and can instead be defined by the location of the
UAV 102 after navigating according to a predetermined patter of flight, or after theUAV 102 has navigated for a predefined amount of time. After generating and transmitting thesecond report 136, at t=4 theUAV 102 can navigate to thesecond beacon 108, can continue to a third monitoring pass position for thefirst beacon 106, or can return to dock at thebase station 104 for recharging. - As shown in
FIG. 3 , theUAV 102 can be configured to dock with the base station in such a way that terminals on its battery are aligned with charging terminals associated with thebase station 104. In some examples, theUAV 102 can use a pilot signal transmitted by thebase station 104 to hone in on thebase station 104 in the same way that it locates and navigates to the beacons. -
FIG. 2 shows aprocess 200 for collecting environmental information using a vehicle. A vehicle determines that the vehicle is located within a predetermined range of a beacon (202). For example, theUAV 102 can use a pilot signal strength transmitted by a beacon to determine its proximity to the beacon. TheUAV 102 can be configured to navigate to a position within a threshold distance from the beacon, such as a position directly above the beacon, or a position several feet away from the beacon. - The
UAV 102 is configured to travel to a first beacon based on a predefined navigational plan that specifies an order in which the beacons should be visited. The beacons can be referenced by their unique beacon IDs in the navigational plan and may, for example, instruct theUAV 102 to “navigate tobeacon 1” as a first waypoint. The pilot signal broadcast by each beacon indicates the beacon ID of its source beacon so that theUAV 102 is able to associate the pilot signal strength with a specific beacon. theUAV 102 may receive more than one pilot signal at a time, as the broadcast cells of multiple beacons may overlap. - The vehicle is navigated to a first location based on determining that the vehicle is located with the predetermined range (204). For example, upon determining that the
UAV 102 is located directly above the beacon, theUAV 102 begins a monitoring pass and navigates to a first monitoring pass position. The first monitoring pass position is defined by a distance from and a bearing relative to the beacon. - While navigating, the
UAV 102 avoids obstacles using theproximity sensor 118. In some examples, the proximity sensor is an infrared or ultrasonic sensor that emits infrared or ultrasonic energy and then detects energy reflected from various surfaces proximal to theUAV 102. TheUAV 102 can use theproximity sensor 118 to evade obstacles such as, for example, machine racks in a data center, people working in the data center, and/or other UAVs. - While in some examples the monitoring pass is defined by a predefined pattern of movement that endeavors to avoid known obstacles in the monitored environment, the
proximity sensor 118 can further ensure that theUAV 102 is able to navigate safely. Theproximity sensor 118 also ensures that theUAV 102 can detect dynamic obstacles that may have been introduced after the creation of the monitoring pass pattern of moment (e.g., if a new piece of equipment is added to a data center). - A first report is generated based on determining that the vehicle is located at the first location (206). For example, upon reaching the first monitoring pass position, the
UAV 102 generates a report using thereport generation engine 120. In some examples, generating the report includes specifying the beacon identification, recording navigation data that includes an altitude, a distance from the beacon, and a bearing relative to the beacon, and recording environmental information. - The navigation data can be determined using a combination of the
altimeter 112, thecompass 116, and the RFID sensor 110. One or more processing devices and/or applications associated with theUAV 102 may process some of the raw data obtained from thealtimeter 112, thecompass 116, and the RFID sensor 110 in order to provide the desired values. For example, a processing device may convert a pilot signal strength in decibels detected by the RFID sensor 110 into a distance in feet or meters. While in some examples theUAV 102 generates a first report after determining that a first location has been reached, theUAV 102 may also generate and transmit a similar report before initiating a monitoring pass, such as upon determining that theUAV 102 is within a predetermined range of thefirst beacon 106. - In some examples, the
UAV 102 records environmental information using one or more of thethermometer 124, thehumidity sensor 126, and any other sensors that can be used to detect environmental conditions. Again, one or more processing devices and/or applications associated with theUAV 102 may process some of the raw data obtained from the environmental sensors (e.g., the humidity sensor 126) in order to provide the desired values. - The
UAV 102 can be configured to re-measure environmental conditions that are determined to be abnormal. For example, if a measured environmental condition appears to be in error (e.g., a value that is unattainable) or is determined to lie be outside of a normal range (e.g., a temperature that is higher than an average or a desirable data center temperature), theUAV 102 can automatically re-measure the environmental conditions. The report data can be uploaded to a computer that stores the data in a database or file for future use. - The first report is transmitted from the vehicle to a base station (208). For example, the
UAV 102 can use thetransmission device 122 to transmit the generated report to thebase station 104 or to a second, different receiving station. In some examples, the first report includes the beacon identification (e.g., “beacon 1”), the navigation data (e.g., an altitude of theUAV 102, a distance of theUAV 102 from the beacon, and a bearing of theUAV 102 relative to the beacon), and the environmental information (e.g., a temperature and a level of humidity). - While in some examples a report is transmitted as a single packet (or as a single file transmitted as multiple packets) after all the data is gathered and organized into the report, the data can also be transmitted dynamically. For example, instead delaying transmission until the beacon ID, the navigation data, and the environmental information have been acquired, the
UAV 102 can also transmit any or all of the data in a piecemeal fashion (e.g., the data can be transmitted as it is received). - Furthermore, in some examples, the
UAV 102 can delay transferring the data to a central entity (e.g., the base station 104) until monitoring has been completed for a single beacon, for multiple beacons, or for the entire monitored environment. For example, theUAV 102 could store the data and/or the generated report onboard and could transfer the same to a central entity upon docking at thebase station 104. TheUAV 102 could also be configured to only transmit reports if any environmental conditions are abnormal, as described above. - After transmitting the first report, the
UAV 102 can take one or more subsequent actions. For example, theUAV 102 can move on to a second monitoring pass position to generate a second report for the same beacon. TheUAV 102 could also navigate to a second beacon (e.g., second beacon 108) and begin monitoring environmental conditions in the vicinity of the second beacon. Furthermore, if all monitoring passes have been completed or if theUAV 102 is recalled, theUAV 102 can return to dock and recharge at thebase station 104. In some examples, theUAV 102 may monitor read its battery capacity (e.g., periodically, after the generation of one or more reports, after visiting one or more beacons, etc). TheUAV 102 may choose to abort further activity and return to base to charge depending on the remaining capacity in its battery. -
FIG. 3 shows asystem 300 that includes theUAV 102 and thebase station 104. This example illustrates how thebase station 104 and theUAV 102 can utilize a gravity-assisted design to dock theUAV 102 at thebase station 104. In some examples, theUAV 102 includes abase portion 132 that is shaped to mate with acavity 301 defined by a body of thebase station 104, even if theUAV 102 fails to land without precisely aligning itself with thecavity 301. - The
base portion 132 of theUAV 102 is shaped to allow thebase portion 132 of theUAV 102 to slide into alignment with thecavity 301 of thebase station 104. For example, because the edges of thebase portion 132 and the housing that defines thecavity 301 are sloped, if theUAV 102 lands in a position that is slightly off-center (shown as position “1” inFIG. 3 ), theUAV 102 will slide into thecavity 301 until coming to rest in a centrally-aligned position (shown as position “2” inFIG. 3 ). - Ensuring that the
base portion 132 ultimately aligns with thecavity 301 allows thebattery 109 of theUAV 102 to be charged by a chargingunit 302 associated with thebase station 104. Because theUAV 102 is meant to operate without assistance from a human operator in some instances, the gravity-assisted landing techniques shown inFIG. 3 can reduce the instances in which theUAV 102 is called upon to perform environmental monitoring but is unable to perform due to a depleted battery. - If the
UAV 102 is properly docked at thebase station 104, positive and 304, 306 of thenegative terminals battery 109 will be aligned with (and will be in electrical communication with) positive and 308, 310 of the chargingnegative terminals unit 302. In some examples, the positive and 308, 310 of thenegative terminals base station 104 are shaped in a ring, such that thebase portion 132 of theUAV 102 can dock at thebase station 104 without regard to its rotation relative to the base station. That is, because the 308, 310 are rings that surround a substantial portion (or the entire portion) of theterminals cavity 301, the 304, 306 of theterminals battery 109 may contact the 308, 310 of theterminals base station 104 regardless of the degree of rotation of thebase portion 312 relative to thebase station 104. Furthermore, this concentric ring terminal contact design may help prevent accidentally inverting the positive and negative terminals which could cause harm to thebattery 109, theUAV 102 and/or thebase station 104. - The charging
unit 302 can be configured to begin charging thebattery 109 upon sensing that an electrical connection has been completed between thebattery 109 and thecharging unit 302. The chargingunit 302 may also reduce the amount of electrical current provided to thebattery 109 as thebattery 109 reaches various levels of charge, and may cease providing current to thebattery 109 entirely when the battery reaches a fully-charged state. The chargingunit 302 can also be configured to provide a “trickle charge” to thebattery 109 if, for example, thebattery 109 is to be maintained in a fully charged state for an extended period of time. - The
base station 104 can also be configured to instruct theUAV 102 to take off and re-attempt docking with thebase station 104 if no electrical connection between thebattery 109 and thecharging unit 302 can be established (e.g., if theUAV 102 lands so poorly that even the gravity-assisted design cannot properly align the 304, 306 with theterminals terminals 308, 310). In the event that theUAV 102 is unable to return to thebase station 104 on its own, theUAV 102 may transmit a request to an operator for manual intervention. - The
base station 104 also includes abroadcasting unit 312 for broadcasting the RFID pilot signal used by theUAV 102 to navigate throughout the monitored environment. In some examples, thebroadcasting unit 312 is an omnidirectional transmitter that is configured to broadcast the base station's pilot signal on a specified frequency and/or channel. The pilot signal includes an identity of thebase station 104 or beacon (e.g., “beacon 0”). Other beacons may include similar broadcasting units. - While the vehicle has been described as a UAV such as a quad-rotor rotary wing aircraft, the techniques described above could be implemented on a variety of other vehicles. For example, one or more track-guided vehicles could be used to measure environmental conditions using techniques similar to those described above.
-
FIG. 4 shows apossible test environment 400 which, in this example, is adata center 401. The data center includes a first row of machines (e.g., racks of equipment, free-standing equipment, etc.) 402 and a second row ofmachines 404. The environmental conditions in thedata center 401 are monitored using the techniques described above. For example, four beacons ( 1, 2, 3, and 4) are located at various points within the data center 401 (e.g., at the ends of the first row ofbeacons machines 402 and the ends of the second row of machines 404). - Each beacon has a respective coverage cell in which its pilot signal is broadcasted. For example,
beacon 1 has acoverage cell 408 andbeacon 2 has acoverage cell 410. The coverage cells can be increased or decreased by increasing or decreasing the transmission power for the pilot signals, respectively. - The
408, 410 also define sharedcoverage cells coverage areas 406. For example, thefirst coverage cell 408 overlaps thesecond coverage cell 410 to define the sharedcoverage area 406; that is, if theUAV 102 is located in the sharedcoverage area 406, its RFID sensor 110 could detect pilot signals broadcasted by bothbeacon 1 andbeacon 2. The size or number of shared coverage areas are a matter of design choice; however, having at least one shared coverage area within each coverage cell can enhance the UAV's ability to locate the next beacon in its flight plan. - If the
UAV 102 is located in an area of a coverage cell that is not a shared coverage area, theUAV 102 may automatically navigate to other locations until it is able to detect the pilot signal from the beacon it should next visit. In these situations, theUAV 102 can be configured to navigate in a predefined direction (e.g., theUAV 102 can rely on navigational hints that have been preloaded onto the UAV 102) or can fly at random until the appropriate pilot signal is detected. In some examples, the navigational hints can include a predetermined direction that theUAV 102 should fly in if a signal is lost (e.g., toward thebase station 104, toward the last-visited beacon, etc.). - The
data center 401 also includes thebase station 104. After theUAV 102 has surveyed the environmental conditions of thedata center 401, theUAV 102 can return to thebase station 104 to recharge itsbattery 109. TheUAV 102 can also be recalled to thebase station 104 upon the occurrence of an event. For example, if theUAV 102 or thebase station 104 detects that thebattery 109 is running out of energy while theUAV 102 is deployed, theUAV 102 can return to thebase station 104, even if theUAV 102 has not finished monitoring thedata center 401. - In some examples, the
UAV 102 may use additional techniques to navigate within the monitored environment. For example, in some situations, RFID-based navigation may not provide a high enough level of accuracy, and greater precision may be desired. To increase the level of precision with which theUAV 102 can navigate (e.g., to beacons and to its base station), theUAV 102 may further include one or more infrared (IR) receivers for receiving IR signals from one or more beacons and base stations within the monitored environment. For example, beacons or base stations may include respective IR transmitters that provide an IR signal encoded to a predetermined frequency (e.g., a frequency that will avoid interference with other light sources within the monitored environment, such as ceiling lights and lights on the monitored electronic equipment). In some examples, depending on the equipment, IR transmitters and receivers can be used to guide theUAV 102 with a greater degree of precision than an RFID-based navigation system. - The
UAV 102 may use IR equipment in combination with the RFID-based navigation system discussed above to efficiently and accurately navigate between various points within a monitored environment. For example, theUAV 102 may acquire the RFID signal broadcast by a beacon from a relatively long distance, and may navigate toward that beacon using the RFID signal. In some examples, once theUAV 102 has arrived within a threshold distance of the beacon, has obtained a line-of-sight with the IR signal source on the beacon, or is otherwise able to detect an IR signal being transmitted by the beacon, theUAV 102 may use the IR signal generated by the beacon to fine-tune its navigation. For example, the UAV may use an RFID signal to navigate to within five feet of a beacon, and may then acquire and use an IR signal generated by the beacon in order to move into a position directly above the beacon (e.g., a position that is within six inches of a desired position). The level of accuracy in the navigation of theUAV 102 can depend on adjustable tolerances for position sensing, and may also depend on the quality of the navigational equipment used by theUAV 102, the beacons, the base station, and any other monitoring equipment. - In some examples, the
UAV 102 can also use one or more cameras (e.g., the horizon camera 130 (FIG. 1 )) to obtain further information about the surrounding environment. For example, theUAV 102 can use thehorizon camera 130 or one or more additional cameras to provide visual feedback of the monitored environment. The visual feedback can include the collection and storage of videos, images, or a combination thereof. For example, theUAV 102 can be configured to use one or more cameras to capture video or images of monitored equipment that has been determined to be problematic. TheUAV 102 can begin capturing visual data upon the detection of an abnormal environmental condition, upon receiving a control signal from a user (e.g., via a user interface that is in communication with the UAV 102), or may periodically capture visual data according to a predefined schedule (e.g., once per day). The additional visual data collected by theUAV 102 can be transmitted to thebase station 104 to be analyzed by an operator, or can be automatically analyzed using one or more image recognition applications. For example, an image recognition application may examine an image of a machine in a datacenter to determine whether one or more warning lights have been activated on the machine. In some examples, the image recognition application may compare the received visual information to images or videos of the subject of the visual information operating in a normal state. Upon the detection of an anomaly by the image recognition software, a notification (e.g., an alert or alarm) can be generated. In some examples, the image recognition software can be implemented on theUAV 102 to facilitate the real-time monitoring of visual data. -
FIG. 5 shows an example of auser interface 500 for interacting with the data captured by theUAV 102. In some examples, theuser interface 500 can be an environmental map of a data center that contains a firstserver rack row 502 and a secondserver rack row 504; that is, theuser interface 500 can be a visual representation of monitored environment. Theuser interface 500 can include a number of tools, graphics, and other features that allow users to monitor the environmental conditions of a given location. - In this example, the
user interface 500 presents a number of environmental graphics, such as temperature graphic 508, which graphically represent the temperature of a monitored point (or group of points) within the monitored environment. Theuser interface 500 includes controls, such asview control 512 anddetail control 514, that allow a user to specify which information the environmental graphics will be associated with. For example, as theview control 512 specifies that a “low temp” will be used anddetail control 514 specifies that both colors and numbers will be used, the environmental graphics represent the temperature recorded at the “low altitude” (e.g., the lowest of three measured altitudes) for one or more monitored locations represented by each environmental graphic, and the environmental graphics are both color-coded according to alegend 510 and display a numerical temperature. For example, the temperature graphic 508 indicates that the temperature recorded at the low location was 103 degrees, which is demonstrated by the numerical value within thetemperature graphic 508. Furthermore, the temperature graphic 508 is colored red which, according to thelegend 510, indicates that the temperature of the location associated with the temperature graphic 508 is over 100 degrees. Theuser interface 500 may also include a graphical representation of abase station 506, as well as graphical representations that show the locations of one or more beacons within the monitored environment (e.g., 1, 2, 3, 4, 5, and 6).beacons - The
view control 512 can be manipulated to cause theuser interface 500 to display various types of information. For example, activating theview control 512 can cause theuser interface 500 to display a high temperature (e.g., the temperature measured at the “high” location), a middle temperature (e.g., the temperature measured at the “middle” location) to show thermal layers based on altitude within the environment, and/or the measured humidity. Theuser interface 500 can also be configured to display alerts that are related to the monitored environment. For example, if the temperature for a given location is measured to be above a threshold value, an alarm can be triggered (e.g., an audible alarm, and/or an email alert or other notification). Furthermore, certain conditions may trigger responsive actions, such as the shutdown of a machine or group of machines that are operating in a region associated with an unfavorable environmental condition (e.g., high heat and/or high humidity). -
FIG. 6 shows anexample data center 600 that includes machine racks 602-604, theUAV 102, thebase station 104, and abackend processing engine 606. While in this example (and in some examples that follow) thebackend processing engine 606 is shown as being located within thedata center 600, thebackend processing engine 606 can also be located outside thedata center 600. For example, the backend processing engine 606 (or one or more additional backend processing engines) may be located remotely relative to thedata center 600, and may communicate with, for example, thebase station 104 over a network (e.g., the Internet). In some examples, some or all of the functions of thebackend processing engine 606 can be provided by thebase station 104, theUAV 102, or some combination thereof. In some examples, theUAV 102 can communicate directly with thebackend processing engine 606. - In general, the
backend processing engine 606 uses data collected by theUAV 102 to monitor conditions of, and effect changes in, thedata center 606. Thebackend processing engine 606 may also use data collected by theUAV 102 to generate reports, or visualizations of the conditions within thedata center 600. In some examples, thebackend processing engine 606 includes anevent engine 608 and aresponse engine 610. Theevent engine 608 can receive data (e.g., reports provided by the UAV 102) and can analyze the received data to determine whether one or more events have occurred. Theevent engine 608 can detect and identify a variety of events, examples of which will be described if greater detail below. Based on detecting one or more events, theevent engine 608 can provide an indication that an event has occurred to theresponse engine 610. Theresponse engine 610 may take one or more predefined actions based on the manner of event identified by theevent engine 608. As a simplified example, if theevent engine 608 determines that thedata center 600 is too warm (e.g., if the temperature of thedata center 600 detected by theUAV 102 exceeds a threshold temperature) and sends an event notification to theresponse engine 610, theresponse engine 610 could take an action to, for example, provide additional cooling to one or more portions of thedata center 600. In some examples, theresponse engine 10 comprises one or more response rules that can be applied to event notifications received from the event engine. -
FIG. 7 shows adata center 700 that includes machine racks 702-705, theUAV 102, thebase station 104, and thebackend processing engine 606. In this example, theUAV 102 collects environmental information (e.g., temperature data) within thedata center 700. TheUAV 102 may collect the environmental information using, for example, the collection and monitoring techniques described above.FIG. 7 shows an example of how thebackend processing engine 606 may use data collected by theUAV 102 to monitor conditions within thedata center 700, and further shows a possible corrective response that can be executed based on the collected data. - As the UAV collects environmental information within the
data center 700, theUAV 102 may detect specific temperatures near the machine racks 702-705. For example, at t=0, theUAV 102 may collect environmental information that includes afirst temperature 708 associated with themachine rack 705. After recording thefirst temperature 708 of themachine rack 705, theUAV 102 may provide a report to thebase station 104 that includes thefirst temperature 708. The report may include information other than thefirst temperature 708. For example, the report may include data that associates thefirst temperature 708 with the machine rack 705 (e.g., location information) as well as other environmental information (e.g., other temperature measurements of themachine rack 705 or other machine racks, humidity data, etc). - The
base station 104 collects the report that includes thefirst temperature 708 from theUAV 102 and provides thefirst temperature 708 to thebackend processing engine 606. Thebase station 104 may provide information to thebackend processing engine 606 as the information is received at thebase station 104, or may provide information to thebackend processing engine 606 in batches (e.g., after thebase station 104 receives a predetermined amount of data or reports). Theevent engine 608 of thebackend processing engine 606 can analyze the information provided by the base station 104 (including the first temperature 708) to determine whether an event has occurred. For example, thebackend processing engine 606 may analyze thefirst temperature 708 and, as a result, may determine that thefirst temperature 708 exceeds athreshold temperature 712. Thethreshold temperature 712 may be associated with thedata center 700 as a whole, or may be associated with a specific area of the data center 700 (e.g., an area in the vicinity of the machine rack 705). If thefirst temperature 708 exceeds thethreshold temperature 712, theevent engine 608 may determine that an event has occurred, and may provide anevent notification 715 to theresponse engine 610. - After receiving the
event notification 715, theresponse engine 710 may execute one or morepredetermined responses 716. For example, theresponse engine 610 may determine that because thefirst temperature 708 exceeds the threshold temperature 712 (e.g., by a predetermined margin), thepredetermine response 716 should be executed to lower thefirst temperature 708. In some examples, the one or morepredetermined responses 716 include a response that causes thebackend processing engine 606 to activate acooling system 706 associated with thedata center 600. In the example ofFIG. 7 , thebackend processing engine 606 has activated thecooling system 706 at t=1 in order to lower the temperature of the data center 700 (or a portion thereof, such as an area associated with the machine rack 705). In some examples, activating thecooling system 706 includes altering one or more data center chillers (e.g., by altering an air temperature level, a fan speed, or other setting of the chiller). While some data center chillers include temperature sensors disposed on or near the chiller itself, allowing theUAV 102 to act as a remote temperature sensor can increase the efficiency of thecooling system 706 by allowing customized temperature readings to replace or supplement the chiller's on-board sensor. In this way, thecooling system 706 can begin reducing a temperature of thedata center 700 even if the area of high temperature is located relatively far from the chiller's onboard temperature sensor, which might otherwise delay the activation of thecooling system 706. - Still referring to
FIG. 7 , the one or morepredetermined responses 716 may include a response that causes thebackend processing engine 606 to adjust adjustment anair duct 707 within thedata center 700 provide additional cooling to the area associated with thefirst temperature 708. For example, at t=1, theresponse engine 610 may adjust one or more vents to direct cooling air toward themachine rack 705. Theair duct 707 may be associated with thecooling system 706, or may be a standalone aspect of a cooling solution of thedata center 700. - After the
response engine 610 has executed one or more of the predetermined responses 716 (also referred to as corrective actions), theUAV 102 can be dispatched to the location associated with thefirst temperature 708 to gather additional environmental information (e.g., to determine the effectiveness of any of thepredetermined responses 716 that were executed). For example, at t=2, thebackend processing engine 606 can dispatch theUAV 102 to themachine rack 705 to take additional temperature readings. In this example, theUAV 102 has determined that the area near themachine rack 705 has dropped to asecond temperature 710. TheUAV 102 may report thesecond temperature 710 to thebackend processing engine 606, and similar evaluations can be performed to determine whether any events have occurred (e.g., whether the second temperature exceeds the threshold temperature 714) and/or whether additional corrective actions should be executed. -
FIG. 8 shows an examplepredetermined response 716 that can be executed by theresponse engine 710. In this example, after receiving theevent notification 715, theresponse engine 610 can execute one of thepredetermined responses 716 to provide anotification 802. In some examples, providing thenotification 802 can include providing avisual alert 804 on adisplay 806 associated with thebackend processing engine 606. Thenotification 802 can also be provided as an audible alarm. In some examples, alerts such as thevisual alert 804 can be configured to notify an operator of thebackend processing engine 606 of the occurrence of an event. In this example, thevisual alert 804 includes a warning related to the elevated temperature nearmachine rack 705. The notification may also provide one or more suggested corrective actions that can be selected by an operator to address the event identified by theevent notification 715. For example, the suggested corrective actions may allow an operator to activate thecooling system 706 or adjust the vents of theair duct 707. -
FIG. 9 shows adata center 900 that includes the machine racks 702-705, theUAV 102, thebase station 104, and thebackend processing engine 606. In this example, the UAV has collected environmental information within thedata center 900 and has generated areport 902 that indicates the temperature detected near each of the machine racks 702-704. Thereport 902 further indicates that the temperature near themachine rack 705 is not available. - After receiving the report 902 (e.g., via the base station 104), the
backend processing engine 606 can determine whether an event has occurred using theevent engine 608. In this example, theevent engine 608 determines that an event has occurred based on the absence of temperature data formachine rack 705 in thereport 902. In some examples, theevent engine 608 may determine whether the area associated with the missing environmental information is large enough to warrant a corrective action. For example, theevent engine 608 may compare a size of the area associated with the missing environmental information to a threshold size (e.g., 25 square feet) and, if the area exceeds the threshold size, an event notification can be generated. - The
event engine 608 provides anevent notification 915 to theresponse engine 610 that indicates the type of event that has occurred. Based on receiving theevent notification 915, theresponse engine 610 executes one of thepredetermined responses 716. In this example, theresponse engine 610 may execute a predetermined response that causes updatedmonitoring instructions 904 to be provided to theUAV 102 which, in turn, may cause the UAV to return to the machine rack 905 to attempt to collect the missing environmental information. In some examples, the updatedmonitoring instructions 904 include instructions to direct theUAV 102 to the nearest RFID beacon, and may also include relative vectors from the beacon to the missed area. In the event theUAV 102 is unable to reach the missed area (e.g., due to the presence of equipment or an obstacle), thebackend processing engine 606 may mark (or may allow an operator to mark) the area as unreachable so that additional monitoring passes are not triggered. -
FIG. 10 shows adata center 1000 that includes 1002 and 1004, themachine racks UAV 102, thebase station 104, and thebackend processing engine 606. Thebackend processing engine 606 includes adisplay 1006. In this example, theUAV 102 has encountered a navigation deviation near themachine rack 1004 due to theopen cabinet door 1008 of themachine rack 1004. For example, theUAV 102 may have used its onboard sensors to detect an obstacle (the cabinet door 1008) in its flight path (e.g., a default flight path) while traveling during the interval identified by t=0 and t=1. As a result, theUAV 102 provides areport 1001 to the backend processing engine 606 (e.g., via the base station 104) that identifies the navigation deviation and the location of theUAV 102 when the navigation deviation was encountered. - In some examples, the
event engine 606 analyzes thereport 1001 to determine that an event has occurred (e.g., to determine that thecabinet door 1008 is ajar). Based on determining that an event has occurred, theevent engine 608 provides anevent notification 1015 to theresponse engine 610. Theresponse engine 610 uses theevent notification 1015 to select one or more of thepredetermined responses 716 to execute to correct the problem that triggered theevent notification 1015. In this example, theresponse engine 610 causes anotification 1004 to be displayed on thedisplay 1006 associated with thebackend processing engine 606. Thenotification 1004 warns an operator that thecabinet door 1008 has been left ajar. Thenotification 1004 could also be in the form of a pop-up window, an automated email, an SMS text message, or an audible alert. Cabinet doors, such as thecabinet door 1008, can provide a level of physical security within thedata center 1000. This is especially important in multi-tenant and high-security data center environments. As a result, keeping thecabinet door 1008 closed can also help promote effective security, and may also improve an airflow within thedata center 100. - Other events may also trigger one or more predetermined responses. For example, the techniques described above can be used to address both real and simulated cooling system failures within a data center. In some examples, in the event of a single cooling device outage, other cooling devices can operate at a higher capacity to provide the necessary coverage (e.g., to maintain appropriate temperature/humidity levels). The
UAV 102 and the back end processing provided by thebackend processing engine 606 can be used to efficiently calibrate the cooling systems within a data center to maintain desired temperature levels. For example, upon detecting that a cooling device has failed (e.g., as a result of theUAV 102 determining that areas near chilling vents or ducts are exceed a threshold temperature), a response engine may automatically adjust other cooling devices within the data center to operate at a higher capacity. After the other cooling devices have been adjusted, theUAV 102 may continue to monitor the environmental conditions within the data center to assess both the effects of the failed cooling device and the results of the adjusted devices. Thebackend processing engine 606 may cause theUAV 102 to perform additional or modified monitoring passes to increase the frequency of data capture associated with the floor space near the failed cooling device. After the failed cooling device is restored, thebackend processing engine 606 can automatically return the settings of all the adjusted cooling devices back to normal levels, and may confirm that the temperatures within the data center stay within expected ranges. - In some examples, cooling devices may be associated with application program interfaces (APIs) that support input from external sources using various protocols. Similarly, older cooling devices that do not have a published interface or network connectivity can be retrofitted with analog interfaces in the form of, for example, a microcontroller platform with a network interface which provides these APIs (e.g., an Arduino-based device). In the event a particular cooling device is unavailable, other cooling devices' thermostats can be set to a lower temperature to compensate for the unavailable cooling device.
- Cooling system failures such as that described above can also be simulated using the
UAV 102 and its associated back end processing. For example, simulated cooling system failures can be created by disabling one or more cooling devices at a time and measuring the impact of the device's failure on the data center. After the results of the simulated failures are captured, a failure response plan can be generated that specifies adjustments for other cooling devices that should be implemented if a given cooling device fails. As a result, in the event of an actual cooling system failure, the response engine could simply execute the stored failure response plan to adjust the remaining active cooling devices. The failure response plan may include various options. For example, the failure response plan may allow thebackend processing engine 606 or an operator to select from varying response levels in the event of a device failure (e.g., maximum cooling, minimum cooling, or minimum power use). In some examples, to simulate a cooling device failure, a thermostat can be adjusted to a high temperature (e.g., to prevent its fan from providing cooling air). A thermostat can then be adjusted to a lower temperature to activate its fans. After a predetermined amount of time (e.g., a few minutes_, theUAV 102 can be launched to observe the temperature pattern and to determine the impact of the simulated cooling device failure and response. - In some examples, the
backend processing engine 606 can be configured to perform historical analysis of the environmental information collected by theUAV 102. Thebackend processing engine 606 can use a historical record of the collected environmental information to optimize conditions within a data center. For example, thebackend processing engine 606 may accumulate a batch of environmental information for a predetermined length of time (e.g., thirty days), and may then evaluate the batch to provide optimization recommendations. In some examples, thebackend processing engine 606 may use historical environmental information to identify areas within a data center in which temperatures or humidity levels are consistently above or below desired levels. As a result, thebackend processing engine 606 may recommend options for relocating equipment (e.g., machine racks) within the data center to correct the areas of concern. For example, if a temperature is consistently high in one area in a rack over an extended period of time, a recommendation can be provided to an administrator that some of the equipment should be moved to another rack, turned off, or adjusted to potentially lower the temperature of the identified area. Thebackend processing engine 606 may also provide an indication that a mechanical failure has possibly occurred (e.g., that a fan in a device is disabled and/or not configured properly). - In some examples, the
backend processing engine 606 may determine, based on the low, medium, and high altitude passes discussed above, a particular region of a machine rack that stores to determine where within the machine rack the offending equipment, or areas of machine racks that have available capacity (e.g., empty machine slots within a machine rack). Additionally, the results of the historical analysis can be used to identify opportunities to adjust air ducts or cooling system vents direct to more efficiently direct cooling air and to more effectively exhaust warm air from the data center. - While some of the examples above refer to the
UAV 102 as the environmental information collection device, other collection devices could be used in combination with the collection and back end processing techniques described above. For example, instead of, or in addition to, theUAV 102, other land, air, or aquatic vehicles could be used to collect the environmental information. Similar collection and back end processing techniques can also be used in combination with data collected by sensors that are not part of a vehicle (e.g., stationary sensors, such as sensors that are integral to a data center or a machine rack). -
FIG. 11 shows an example of acomputing device 1100 and amobile computing device 1150 that can be used to implement the techniques described in this disclosure. Thecomputing device 1100 is intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. Themobile computing device 1150 is intended to represent various forms of mobile devices, such as personal digital assistants, cellular telephones, smart-phones, and other similar computing devices. The components shown here, their connections and relationships, and their functions, are meant to be examples only, and are not meant to be limiting. - The
computing device 1100 includes aprocessor 1102, amemory 1104, astorage device 1106, a high-speed interface 1108 connecting to thememory 1104 and multiple high-speed expansion ports 1110, and a low-speed interface 1112 connecting to a low-speed expansion port 1114 and thestorage device 1106. Each of theprocessor 1102, thememory 1104, thestorage device 1106, the high-speed interface 1108, the high-speed expansion ports 1110, and the low-speed interface 1112, are interconnected using various busses, and may be mounted on a common motherboard or in other manners as appropriate. Theprocessor 1102 can process instructions for execution within thecomputing device 1100, including instructions stored in thememory 1104 or on thestorage device 1106 to display graphical information for a GUI on an external input/output device, such as adisplay 1116 coupled to the high-speed interface 1108. In other implementations, multiple processors and/or multiple buses may be used, as appropriate, along with multiple memories and types of memory. Also, multiple computing devices may be connected, with each device providing portions of the necessary operations (e.g., as a server bank, a group of blade servers, or a multi-processor system). - The
memory 1104 stores information within thecomputing device 1100. In some implementations, thememory 1104 is a volatile memory unit or units. In some implementations, thememory 1104 is a non-volatile memory unit or units. Thememory 1104 may also be another form of computer-readable medium, such as a magnetic or optical disk. - The
storage device 1106 is capable of providing mass storage for thecomputing device 1100. In some implementations, thestorage device 1106 may be or contain a computer-readable medium, such as a floppy disk device, a hard disk device, an optical disk device, or a tape device, a flash memory or other similar solid state memory device, or an array of devices, including devices in a storage area network or other configurations. Instructions can be stored in an information carrier. The instructions, when executed by one or more processing devices (for example, processor 1102), perform one or more methods, such as those described above. The instructions can also be stored by one or more storage devices such as computer- or machine-readable mediums (for example, thememory 1104, thestorage device 1106, or memory on the processor 1102). - The high-
speed interface 1108 manages bandwidth-intensive operations for thecomputing device 1100, while the low-speed interface 1112 manages lower bandwidth-intensive operations. Such allocation of functions is an example only. In some implementations, the high-speed interface 1108 is coupled to thememory 1104, the display 1116 (e.g., through a graphics processor or accelerator), and to the high-speed expansion ports 1110, which may accept various expansion cards (not shown). In the implementation, the low-speed interface 1112 is coupled to thestorage device 1106 and the low-speed expansion port 1114. The low-speed expansion port 1114, which may include various communication ports (e.g., USB, Bluetooth, Ethernet, wireless Ethernet) may be coupled to one or more input/output devices, such as a keyboard, a pointing device, a scanner, or a networking device such as a switch or router, e.g., through a network adapter. - The
computing device 1100 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as astandard server 1120, or multiple times in a group of such servers. In addition, it may be implemented in a personal computer such as a laptop computer 1122. It may also be implemented as part of arack server system 1124. Alternatively, components from thecomputing device 1100 may be combined with other components in a mobile device (not shown), such as amobile computing device 1150. Each of such devices may contain one or more of thecomputing device 1100 and themobile computing device 1150, and an entire system may be made up of multiple computing devices communicating with each other. - The
mobile computing device 1150 includes aprocessor 1152, amemory 1164, an input/output device such as adisplay 1154, acommunication interface 1166, and atransceiver 1168, among other components. Themobile computing device 1150 may also be provided with a storage device, such as a micro-drive or other device, to provide additional storage. Each of theprocessor 1152, thememory 1164, thedisplay 1154, thecommunication interface 1166, and thetransceiver 1168, are interconnected using various buses, and several of the components may be mounted on a common motherboard or in other manners as appropriate. - The
processor 1152 can execute instructions within themobile computing device 1150, including instructions stored in thememory 1164. Theprocessor 1152 may be implemented as a chipset of chips that include separate and multiple analog and digital processors. Theprocessor 1152 may provide, for example, for coordination of the other components of themobile computing device 1150, such as control of user interfaces, applications run by themobile computing device 1150, and wireless communication by themobile computing device 1150. - The
processor 1152 may communicate with a user through acontrol interface 1158 and adisplay interface 1156 coupled to thedisplay 1154. Thedisplay 1154 may be, for example, a TFT (Thin-Film-Transistor Liquid Crystal Display) display or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology. Thedisplay interface 1156 may comprise appropriate circuitry for driving thedisplay 1154 to present graphical and other information to a user. Thecontrol interface 1158 may receive commands from a user and convert them for submission to theprocessor 1152. In addition, anexternal interface 1162 may provide communication with theprocessor 1152, so as to enable near area communication of themobile computing device 1150 with other devices. Theexternal interface 1162 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used. - The
memory 1164 stores information within themobile computing device 1150. Thememory 1164 can be implemented as one or more of a computer-readable medium or media, a volatile memory unit or units, or a non-volatile memory unit or units. Anexpansion memory 1174 may also be provided and connected to themobile computing device 1150 through anexpansion interface 1172, which may include, for example, a SIMM (Single In Line Memory Module) card interface. Theexpansion memory 1174 may provide extra storage space for themobile computing device 1150, or may also store applications or other information for themobile computing device 1150. Specifically, theexpansion memory 1174 may include instructions to carry out or supplement the processes described above, and may include secure information also. Thus, for example, theexpansion memory 1174 may be provide as a security module for themobile computing device 1150, and may be programmed with instructions that permit secure use of themobile computing device 1150. In addition, secure applications may be provided via the SIMM cards, along with additional information, such as placing identifying information on the SIMM card in a non-hackable manner. - The memory may include, for example, flash memory and/or NVRAM memory (non-volatile random access memory), as discussed below. In some implementations, instructions are stored in an information carrier. that the instructions, when executed by one or more processing devices (for example, processor 1152), perform one or more methods, such as those described above. The instructions can also be stored by one or more storage devices, such as one or more computer- or machine-readable mediums (for example, the
memory 1164, theexpansion memory 1174, or memory on the processor 1152). In some implementations, the instructions can be received in a propagated signal, for example, over thetransceiver 1168 or theexternal interface 1162. - The
mobile computing device 1150 may communicate wirelessly through thecommunication interface 1166, which may include digital signal processing circuitry where necessary. Thecommunication interface 1166 may provide for communications under various modes or protocols, such as GSM voice calls (Global System for Mobile communications), SMS (Short Message Service), EMS (Enhanced Messaging Service), or MMS messaging (Multimedia Messaging Service), CDMA (code division multiple access), TDMA (time division multiple access), PDC (Personal Digital Cellular), WCDMA (Wideband Code Division Multiple Access), CDMA2000, or GPRS (General Packet Radio Service), among others. Such communication may occur, for example, through thetransceiver 1168 using a radio-frequency. In addition, short-range communication may occur, such as using a Bluetooth, WiFi, or other such transceiver (not shown). In addition, a GPS (Global Positioning System)receiver module 1170 may provide additional navigation- and location-related wireless data to themobile computing device 1150, which may be used as appropriate by applications running on themobile computing device 1150. - The
mobile computing device 1150 may also communicate audibly using anaudio codec 1160, which may receive spoken information from a user and convert it to usable digital information. Theaudio codec 1160 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of themobile computing device 1150. Such sound may include sound from voice telephone calls, may include recorded sound (e.g., voice messages, music files, etc.) and may also include sound generated by applications operating on themobile computing device 1150. - The
mobile computing device 1150 may be implemented in a number of different forms, as shown in the figure. For example, it may be implemented as acellular telephone 1180. It may also be implemented as part of a smart-phone 1182, personal digital assistant, or other similar mobile device. - Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, specially designed ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various implementations can include implementation in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, coupled to receive data and instructions from, and to transmit data and instructions to, a storage system, at least one input device, and at least one output device.
- These computer programs (also known as programs, software, software applications or code) include machine instructions for a programmable processor, and can be implemented in a high-level procedural and/or object-oriented programming language, and/or in assembly/machine language. As used herein, the terms machine-readable medium and computer-readable medium refer to any computer program product, apparatus and/or device (e.g., magnetic discs, optical disks, memory, Programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term machine-readable signal refers to any signal used to provide machine instructions and/or data to a programmable processor.
- To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user and a keyboard and a pointing device (e.g., a mouse or a trackball) by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user can be received in any form, including acoustic, speech, or tactile input.
- The systems and techniques described here can be implemented in a computing system that includes a back end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front end component (e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back end, middleware, or front end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include a local area network (LAN), a wide area network (WAN), and the Internet.
- The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
- Although a few implementations have been described in detail above, other modifications are possible. For example, while a client application is described as accessing the delegate(s), in other implementations the delegate(s) may be employed by other applications implemented by one or more processors, such as an application executing on one or more servers. In addition, the logic flows depicted in the figures do not require the particular order shown, or sequential order, to achieve desirable results. In addition, other actions may be provided, or actions may be eliminated, from the described flows, and other components may be added to, or removed from, the described systems. Accordingly, other implementations are within the scope of the following claims.
Claims (23)
1. A method performed on one or more computing devices, the method comprising:
receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center;
storing the environmental information on one or more memory devices that are accessible by the one or more computing devices;
determining, based on the environmental information, that an event has occurred; and
performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
2. The method of claim 1 , wherein the environmental information is associated with one or more of a temperature and a humidity level of at least a portion of the data center.
3. The method of claim 2 , wherein determining that the event has occurred comprises determining that the temperature or the humidity level associated with the at least a portion of the data center falls outside a predetermined range.
4. The method of claim 3 , wherein performing the one or more actions comprises activating a cooling system associated with the at least a portion of the data center.
5. The method of claim 3 , wherein performing the one or more actions comprises adjusting an air vent associated with the at least a portion of the data center.
6. The method of claim 5 , wherein adjusting the air vent alters a flow of air through the at least a portion of the data center.
7. The method of claim 3 , wherein the one or more actions comprise causing the UAV to travel to the at least a portion of the data center to collect additional environmental information.
8. The method of claim 6 , further comprising:
determining that the additional environmental information is associated with one or more of a second temperature and a second humidity level that falls outside the predetermined range; and
providing a notification.
9. The method of claim 8 , wherein providing the notification comprises providing at least one of either: an on screen pop-up message, or an automated email, or a short message service (SMS) message, or any combination thereof.
10. The method of claim 1 , wherein determining that the event has occurred comprises determining that a cooling system has been disabled.
11. The method of claim 10 , wherein performing the one or more actions comprises:
altering an operation of one or more other cooling systems; and
causing the UAV to travel to one or more portions of the data center to collect additional environmental information.
12. The method of claim 10 , further comprising altering a frequency with which the UAV collects environmental information associated with a portion of the data center that is associated with the disabled cooling system.
13. The method of claim 1 , wherein the environmental information specifies that the UAV has experienced a navigation deviation.
14. The method of claim 13 , wherein determining that the event has occurred comprises determining, based on the navigation deviation, that a mechanical change has occurred in the data center.
15. The method of claim 14 , wherein determining that the mechanical change has occurred comprises determining that a door located in the data center is ajar.
16. The method of claim 1 , wherein determining that the event has occurred comprises determining that a predetermined amount of environmental information has been received.
17. The method of claim 16 , wherein performing the action comprises analyzing the environmental information to determine the location of one or more portions of the data center that are associated with environmental information that falls outside of a predetermined range.
18. The method of claim 1 , wherein determining that the event has occurred comprises determining that environmental information has not been collected for a portion of the data center.
19. The method of claim 18 , further comprising determining that the portion of the data center is inaccessible by the UAV.
20. The method of claim 18 , further comprising determining that a size of the portion of the data center exceeds a predetermined size.
21. The method of claim 19 , further comprising causing the UAV to collect environmental information for the portion of the data center based on determining that the size of the portion of the data center exceeds the predetermined size.
22. One or more computer storage devices comprising instructions that, when executed by one or more processing devices, cause the one or more processing devices to perform operations comprising:
receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center;
storing the environmental information on one or more memory devices that are accessible by the one or more computing devices;
determining, based on the environmental information, that an event has occurred; and
performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
23. A system comprising:
one or more processing devices; and
one or more memory devices comprising instructions that, when executed by one or more processing devices, cause the one or more processing devices to perform operations comprising:
receiving environmental information collected by an unmanned aerial vehicle (UAV), the environmental information being associated with one or more conditions of a data center;
storing the environmental information on one or more memory devices that are accessible by the one or more computing devices;
determining, based on the environmental information, that an event has occurred; and
performing, by the one or more computing devices and based on determining that the event has occurred, one or more actions.
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US13/222,230 US20120271461A1 (en) | 2011-04-20 | 2011-08-31 | Capturing environmental information |
| EP12002777.6A EP2515147B1 (en) | 2011-04-20 | 2012-04-20 | Capturing environmental information |
| US13/451,922 US8818705B2 (en) | 2011-04-20 | 2012-04-20 | Capturing environmental information |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201161477377P | 2011-04-20 | 2011-04-20 | |
| US13/222,230 US20120271461A1 (en) | 2011-04-20 | 2011-08-31 | Capturing environmental information |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/451,922 Continuation US8818705B2 (en) | 2011-04-20 | 2012-04-20 | Capturing environmental information |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20120271461A1 true US20120271461A1 (en) | 2012-10-25 |
Family
ID=46085318
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/222,230 Abandoned US20120271461A1 (en) | 2011-04-20 | 2011-08-31 | Capturing environmental information |
| US13/451,922 Active 2032-01-01 US8818705B2 (en) | 2011-04-20 | 2012-04-20 | Capturing environmental information |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US13/451,922 Active 2032-01-01 US8818705B2 (en) | 2011-04-20 | 2012-04-20 | Capturing environmental information |
Country Status (2)
| Country | Link |
|---|---|
| US (2) | US20120271461A1 (en) |
| EP (1) | EP2515147B1 (en) |
Cited By (66)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20120296497A1 (en) * | 2011-05-18 | 2012-11-22 | Hon Hai Precision Industry Co., Ltd. | Unmanned aerial vehicle and method for controlling the unmanned aerial vehicle |
| US20140277854A1 (en) * | 2013-03-15 | 2014-09-18 | Azure Sky Group Llc | Modular drone and methods for use |
| US20150120094A1 (en) * | 2013-10-26 | 2015-04-30 | Amazon Technologies, Inc. | Unmanned aerial vehicle delivery system |
| US20150161712A1 (en) * | 2013-12-10 | 2015-06-11 | 12 Retail (HK) Limited | Unifying shopping experience system |
| US20150338855A1 (en) * | 2013-02-15 | 2015-11-26 | Disney Enterprises, Inc. | Controlling unmanned aerial vehicles as a flock to synchronize flight in aerial displays |
| US20160116915A1 (en) * | 2014-10-22 | 2016-04-28 | Honeywell International Inc. | Surveying areas using a radar system and an unmanned aerial vehicle |
| WO2016100601A3 (en) * | 2014-12-17 | 2016-08-11 | Picpocket, Inc. | Drone based systems and methodologies for capturing images |
| US9439092B1 (en) * | 2015-07-27 | 2016-09-06 | Sprint Communications Company L.P. | Detection of component fault at cell towers |
| JP2016197404A (en) * | 2015-03-25 | 2016-11-24 | 横河電機株式会社 | System and method for monitoring industrial plant |
| US9513635B1 (en) | 2015-12-30 | 2016-12-06 | Unmanned Innovation, Inc. | Unmanned aerial vehicle inspection system |
| US9594372B1 (en) * | 2016-01-21 | 2017-03-14 | X Development Llc | Methods and systems for providing feedback based on information received from an aerial vehicle |
| US9609288B1 (en) | 2015-12-31 | 2017-03-28 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| JP2017071233A (en) * | 2015-10-05 | 2017-04-13 | 廣田 祐次 | Drone shelter |
| WO2017066647A1 (en) * | 2015-10-15 | 2017-04-20 | T-Mobile Usa, Inc. | Dynamic wireless communications network with a plurality of aerial drones |
| US20170122909A1 (en) * | 2012-10-27 | 2017-05-04 | Valerian Goroshevskiy | Non-destructive system and method for detecting structural defects |
| US9740200B2 (en) | 2015-12-30 | 2017-08-22 | Unmanned Innovation, Inc. | Unmanned aerial vehicle inspection system |
| US9747480B2 (en) | 2011-12-05 | 2017-08-29 | Adasa Inc. | RFID and robots for multichannel shopping |
| WO2017151156A1 (en) * | 2016-03-02 | 2017-09-08 | Lazo-Antunez Raynel | Automated hands-free umbrella |
| US9780435B2 (en) | 2011-12-05 | 2017-10-03 | Adasa Inc. | Aerial inventory antenna |
| US9802701B1 (en) | 2014-10-21 | 2017-10-31 | Joshua Hawes | Variable elevation signal acquisition and data collection system and method |
| JP2017534513A (en) * | 2014-10-02 | 2017-11-24 | スウィフト・エンジニアリング・インコーポレーテッド | Portable ground station for unmanned aerial vehicles |
| CN107430402A (en) * | 2015-03-31 | 2017-12-01 | 深圳市大疆创新科技有限公司 | Systems and methods for identifying and authenticating geofencing devices |
| US10050330B2 (en) | 2011-12-05 | 2018-08-14 | Adasa Inc. | Aerial inventory antenna |
| WO2018148636A1 (en) * | 2017-02-13 | 2018-08-16 | Top Flight Technologies, Inc. | Weather sensing |
| US10097293B2 (en) | 2014-12-04 | 2018-10-09 | At&T Intellectual Property I, L.P. | Communicate an alert via an unmanned vehicle |
| US10126745B2 (en) * | 2015-01-04 | 2018-11-13 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
| US10220954B2 (en) | 2015-01-04 | 2019-03-05 | Zero Zero Robotics Inc | Aerial system thermal control system and method |
| US10222800B2 (en) | 2015-01-04 | 2019-03-05 | Hangzhou Zero Zero Technology Co., Ltd | System and method for automated aerial system operation |
| JP2019055774A (en) * | 2018-10-18 | 2019-04-11 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | System for operating unmanned aircraft |
| US10353052B2 (en) * | 2016-09-15 | 2019-07-16 | Lawrence Livermore National Security, Llc | Object discrimination based on a swarm of agents |
| US10358214B2 (en) | 2015-01-04 | 2019-07-23 | Hangzhou Zero Zro Technology Co., Ltd. | Aerial vehicle and method of operation |
| CN110130879A (en) * | 2019-04-28 | 2019-08-16 | 中国石油天然气股份有限公司 | Method and system for collecting oil well parameters |
| US10405223B1 (en) | 2017-02-14 | 2019-09-03 | Sprint Communications Company L.P. | System and methods for intelligent reset delay for cell sites in a network |
| US10434885B2 (en) | 2014-08-05 | 2019-10-08 | Telecom Italia S.P.A. | Landing platform for an unmanned aerial vehicle |
| US10476130B2 (en) | 2011-12-05 | 2019-11-12 | Adasa Inc. | Aerial inventory antenna |
| US10534068B2 (en) * | 2018-12-27 | 2020-01-14 | Intel Corporation | Localization system, vehicle control system, and methods thereof |
| US10574614B2 (en) | 2009-08-03 | 2020-02-25 | Picpocket Labs, Inc. | Geofencing of obvious geographic locations and events |
| US10659144B1 (en) | 2019-01-31 | 2020-05-19 | At&T Intellectual Property I, L.P. | Management of massively distributed internet of things (IOT) gateways based on software-defined networking (SDN) via fly-by master drones |
| US10719080B2 (en) | 2015-01-04 | 2020-07-21 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system and detachable housing |
| US10785323B2 (en) | 2015-01-05 | 2020-09-22 | Picpocket Labs, Inc. | Use of a dynamic geofence to control media sharing and aggregation associated with a mobile target |
| US10846497B2 (en) | 2011-12-05 | 2020-11-24 | Adasa Inc. | Holonomic RFID reader |
| CN111986394A (en) * | 2019-05-21 | 2020-11-24 | 顺丰科技有限公司 | A security system and method for delivery of goods |
| US10856115B2 (en) | 2009-08-03 | 2020-12-01 | Picpocket Labs, Inc. | Systems and methods for aggregating media related to an event |
| US20210163134A1 (en) * | 2018-06-14 | 2021-06-03 | Beijing Xiaomi Mobile Software Co., Ltd. | Information sending and receiving method and apparatus, device, and storage medium |
| CN112912312A (en) * | 2018-10-19 | 2021-06-04 | 安炳烈 | Unmanned aerial vehicle who possesses conflict prevention and retrieve function |
| US11029352B2 (en) | 2016-05-18 | 2021-06-08 | Skydio, Inc. | Unmanned aerial vehicle electromagnetic avoidance and utilization system |
| US11027833B2 (en) | 2016-04-24 | 2021-06-08 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system propulsion assembly and method of use |
| US11086337B2 (en) | 2017-06-20 | 2021-08-10 | Planck Aerosystems Inc. | Systems and methods for charging unmanned aerial vehicles on a moving platform |
| US11094202B2 (en) | 2015-03-31 | 2021-08-17 | SZ DJI Technology Co., Ltd. | Systems and methods for geo-fencing device communications |
| US11093722B2 (en) | 2011-12-05 | 2021-08-17 | Adasa Inc. | Holonomic RFID reader |
| US11108849B2 (en) | 2018-12-03 | 2021-08-31 | At&T Intellectual Property I, L.P. | Global internet of things (IOT) quality of service (QOS) realization through collaborative edge gateways |
| US11120456B2 (en) | 2015-03-31 | 2021-09-14 | SZ DJI Technology Co., Ltd. | Authentication systems and methods for generating flight regulations |
| CN113589222A (en) * | 2020-04-07 | 2021-11-02 | 英西图公司 | Aircraft guidance with beacon transmission |
| US20220073204A1 (en) * | 2015-11-10 | 2022-03-10 | Matternet, Inc. | Methods and systems for transportation using unmanned aerial vehicles |
| US20220148445A1 (en) * | 2015-12-31 | 2022-05-12 | Skydio, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US11573575B2 (en) | 2017-04-12 | 2023-02-07 | Lawrence Livermore National Security, Llc | Attract-repel path planner system for collision avoidance |
| WO2023025200A1 (en) * | 2021-08-25 | 2023-03-02 | 深圳市道通智能航空技术股份有限公司 | Control method and apparatus for forced landing of unmanned aerial vehicle, remote control apparatus, and storage medium |
| US11710191B2 (en) | 2014-09-22 | 2023-07-25 | State Farm Mutual Automobile Insurance Company | Insurance underwriting and re-underwriting implementing unmanned aerial vehicles (UAVs) |
| US11796673B2 (en) | 2016-07-06 | 2023-10-24 | Lawrence Livermore National Security, Llc | Object sense and avoid system for autonomous vehicles |
| US11927972B2 (en) | 2020-11-24 | 2024-03-12 | Lawrence Livermore National Security, Llc | Collision avoidance based on traffic management data |
| US20240239531A1 (en) * | 2022-08-09 | 2024-07-18 | Pete Bitar | Compact and Lightweight Drone Delivery Device called an ArcSpear Electric Jet Drone System Having an Electric Ducted Air Propulsion System and Being Relatively Difficult to Track in Flight |
| US12084179B2 (en) | 2018-05-23 | 2024-09-10 | Aerovironment, Inc. | System and method for drone tethering |
| US12131656B2 (en) | 2012-05-09 | 2024-10-29 | Singularity University | Transportation using network of unmanned aerial vehicles |
| CN119247513A (en) * | 2024-12-05 | 2025-01-03 | 南京大桥机器有限公司 | A mobile delivery control system and method for intelligent weather detection device |
| US12228407B2 (en) | 2019-01-15 | 2025-02-18 | Aerovironment, Inc. | Systems and methods for delivery using unmanned aerial vehicles |
| US12473108B1 (en) * | 2023-12-15 | 2025-11-18 | Amazon Technologies, Inc. | Stations for guided docking evolutions by aerial vehicles |
Families Citing this family (83)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP5240132B2 (en) * | 2009-09-04 | 2013-07-17 | 富士通株式会社 | Thermal fluid simulation analyzer |
| US10019000B2 (en) * | 2012-07-17 | 2018-07-10 | Elwha Llc | Unmanned device utilization methods and systems |
| US9254363B2 (en) * | 2012-07-17 | 2016-02-09 | Elwha Llc | Unmanned device interaction methods and systems |
| US9198363B2 (en) * | 2012-12-12 | 2015-12-01 | The Boeing Company | Tree metrology system |
| US8843303B1 (en) * | 2012-12-17 | 2014-09-23 | Rockwell Collins, Inc. | Risk-aware contingency flight re-planner system and related method |
| US20140316614A1 (en) * | 2012-12-17 | 2014-10-23 | David L. Newman | Drone for collecting images and system for categorizing image data |
| CN103135550B (en) * | 2013-01-31 | 2015-05-20 | 南京航空航天大学 | Multiple obstacle-avoidance control method of unmanned plane used for electric wire inspection |
| KR20200105530A (en) | 2013-05-03 | 2020-09-07 | 에어로바이론먼트 인크 | Vertical takeoff and landing (vtol) air vehicle |
| US9577445B2 (en) * | 2013-09-09 | 2017-02-21 | Olaeris, Inc. | Vehicle replenishment |
| US9936114B2 (en) * | 2013-10-25 | 2018-04-03 | Elwha Llc | Mobile device for requesting the capture of an image |
| US9158304B2 (en) * | 2013-11-10 | 2015-10-13 | Google Inc. | Methods and systems for alerting and aiding an emergency situation |
| EP2895819B1 (en) | 2013-12-10 | 2020-05-20 | SZ DJI Technology Co., Ltd. | Sensor fusion |
| CN103914075B (en) * | 2013-12-13 | 2017-05-03 | 深圳市大疆创新科技有限公司 | Control method and device for unmanned aerial vehicle |
| US9426674B2 (en) * | 2014-02-03 | 2016-08-23 | Honeywell International Inc. | Planning a wireless network |
| US9621258B2 (en) * | 2014-02-26 | 2017-04-11 | Kutta Technologies, Inc. | Bi-directional communication for control of unmanned systems |
| US9562773B2 (en) | 2014-03-15 | 2017-02-07 | Aurora Flight Sciences Corporation | Autonomous vehicle navigation system and method |
| US9862504B1 (en) | 2014-04-04 | 2018-01-09 | Olaeris, Inc. | Positioning hovering objects for docking |
| US9875661B2 (en) | 2014-05-10 | 2018-01-23 | Aurora Flight Sciences Corporation | Dynamic collision-avoidance system and method |
| CN104029825B (en) * | 2014-06-13 | 2016-01-20 | 中国人民解放军装甲兵工程学院 | On-the-spot actual situation coupling Unmanned Aircraft Systems (UAS) |
| US11126193B2 (en) | 2014-06-19 | 2021-09-21 | Husqvarna Ab | Automatic beacon position determination |
| ES2601136T3 (en) | 2014-07-01 | 2017-02-14 | The Boeing Company | Infrastructure and mobile management system for unmanned aerial vehicles and related methods |
| US20160003954A1 (en) * | 2014-07-07 | 2016-01-07 | Schlumberger Technology Corporation | Aerial vehicle acquisition of seismic data |
| US9087451B1 (en) * | 2014-07-14 | 2015-07-21 | John A. Jarrell | Unmanned aerial vehicle communication, monitoring, and traffic management |
| WO2016008125A1 (en) * | 2014-07-16 | 2016-01-21 | 深圳市大疆创新科技有限公司 | Electric unmanned aerial vehicle and intelligent electric quantity protection method therefor |
| CN105980950B (en) | 2014-09-05 | 2019-05-28 | 深圳市大疆创新科技有限公司 | Speed control of unmanned aerial vehicles |
| DK3428766T3 (en) | 2014-09-05 | 2021-06-07 | Sz Dji Technology Co Ltd | MULTI-SENSOR FOR IMAGING THE ENVIRONMENT |
| CN105517666B (en) | 2014-09-05 | 2019-08-27 | 深圳市大疆创新科技有限公司 | Scenario-based flight mode selection |
| US9062948B1 (en) * | 2014-10-03 | 2015-06-23 | ASGS Associates, Trustee for Aerial Smoke Generator System CRT Trust | Aerial smoke generator system |
| US10093432B2 (en) * | 2014-10-09 | 2018-10-09 | James McKinnon | Drone receiving systems and methods |
| US20160155097A1 (en) * | 2014-12-02 | 2016-06-02 | Ebay Inc. | Reports of repairable objects and events |
| US9977435B2 (en) | 2015-02-11 | 2018-05-22 | Aeroviroment, Inc. | Survey migration system for vertical take-off and landing (VTOL) unmanned aerial vehicles (UAVS) |
| US9880563B2 (en) | 2015-02-11 | 2018-01-30 | Aerovironment, Inc. | Geographic survey system for vertical take-off and landing (VTOL) unmanned aerial vehicles (UAVs) |
| WO2016130797A1 (en) | 2015-02-11 | 2016-08-18 | Aerovironment, Inc. | Pod cover system for a vertical take-off and landing (vtol) unmanned aerial vehicle (uav) |
| WO2016130847A1 (en) | 2015-02-11 | 2016-08-18 | Aerovironment, Inc. | Pod launch and landing system for vertical take-off and landing (vtol) unmanned aerial vehicles (uavs) |
| US11021266B2 (en) * | 2015-02-11 | 2021-06-01 | Aerovironment, Inc. | Pod operating system for a vertical take-off and landing (VTOL) unmanned aerial vehicle (UAV) |
| US20160307447A1 (en) | 2015-02-13 | 2016-10-20 | Unmanned Innovation, Inc. | Unmanned aerial vehicle remote flight planning system |
| SG11201707306YA (en) | 2015-03-12 | 2017-10-30 | Nightingale Intelligent Systems | Automated drone systems |
| US9739615B2 (en) | 2015-03-16 | 2017-08-22 | International Business Machines Corporation | Autonomous data evacuation |
| ES1140161Y (en) * | 2015-05-18 | 2015-09-10 | Creadores De Estrategia Para Proyectos De Ingeniería S L | REMOTE AIR DEVICE FOR MEASUREMENT OF ENVIRONMENTAL VARIABLES IN CLOSED SPACES |
| US9828093B2 (en) * | 2015-05-27 | 2017-11-28 | First Principles, Inc. | System for recharging remotely controlled aerial vehicle, charging station and rechargeable remotely controlled aerial vehicle, and method of use thereof |
| CN105318888B (en) * | 2015-12-07 | 2018-08-14 | 北京航空航天大学 | Automatic driving vehicle paths planning method based on unmanned plane perception |
| CN105425208A (en) * | 2015-12-21 | 2016-03-23 | 深圳思科尼亚科技有限公司 | Positioning system and method used for accurate navigation of unmanned aerial vehicle |
| US10580301B2 (en) * | 2016-03-16 | 2020-03-03 | Kiomars Anvari | Moving vehicles in a smart environment |
| US10700767B2 (en) | 2016-03-16 | 2020-06-30 | Honeywell International Inc. | Requesting weather data based on pre-selected events |
| US10188580B2 (en) * | 2016-05-09 | 2019-01-29 | Toyota Motor Engineering & Manufacturing North America, Inc. | Systems and methods for providing environment information using an unmanned vehicle |
| DE102016109242A1 (en) * | 2016-05-19 | 2017-11-23 | Keil Group GmbH | monitoring system |
| US10097952B2 (en) * | 2016-05-20 | 2018-10-09 | Bi Incorporated | Systems and methods for monitoring altitude sensing beacons |
| CN105955303A (en) * | 2016-07-05 | 2016-09-21 | 北京奇虎科技有限公司 | UAV autonomous obstacle avoidance method and device |
| AU2017296346A1 (en) * | 2016-07-12 | 2019-02-28 | Daniel Aljadeff | Methods and systems for the location and monitoring of mobile units |
| CN107783548B (en) * | 2016-08-25 | 2021-02-26 | 大连楼兰科技股份有限公司 | Data processing method based on multi-sensor information fusion technology |
| CN111580507B (en) * | 2016-08-31 | 2022-01-14 | 深圳市大疆创新科技有限公司 | Unmanned aerial vehicle, electronic speed regulator, control method and control system thereof |
| CN109716816B (en) * | 2016-09-27 | 2022-10-14 | 索尼公司 | Circuit, base station, method, and recording medium |
| KR20180056068A (en) * | 2016-11-18 | 2018-05-28 | 삼성전자주식회사 | Electronic device and method for controlling unmanned aerial vehicle |
| AT519417B1 (en) * | 2016-11-22 | 2019-04-15 | Riegl Laser Measurement Systems Gmbh | Method of measuring a condition of a metallurgical vessel in a steel mill and steelwork therefor |
| CN106774408A (en) * | 2016-12-30 | 2017-05-31 | 易瓦特科技股份公司 | It is applied to the method and system of unmanned plane |
| CN107065854A (en) * | 2016-12-30 | 2017-08-18 | 易瓦特科技股份公司 | The method and apparatus for being charged and being can return to for unmanned plane |
| CN106843267A (en) * | 2016-12-30 | 2017-06-13 | 易瓦特科技股份公司 | Unmanned plane charges and returns the method and device taken off |
| IL267810B2 (en) | 2017-01-06 | 2025-07-01 | Aurora Flight Sciences Corp | Collision-avoidance system and method for unmanned aircraft |
| JP6938969B2 (en) * | 2017-03-07 | 2021-09-22 | 富士フイルムビジネスイノベーション株式会社 | Environmental measurement systems and programs |
| US10690781B2 (en) | 2017-04-05 | 2020-06-23 | At&T Intellectual Property I, L.P. | Unmanned aerial vehicle drive testing and mapping of carrier signals |
| US10642264B2 (en) * | 2017-07-19 | 2020-05-05 | Superior Communications, Inc. | Security drone system |
| WO2019079394A2 (en) * | 2017-10-17 | 2019-04-25 | Top Flight Technologies, Inc. | Portable launch system |
| CN110015097B (en) * | 2017-10-23 | 2021-11-02 | 中国移动通信有限公司研究院 | A UAV charging method and related equipment |
| US11148805B2 (en) * | 2018-04-10 | 2021-10-19 | Government Of The United States, As Represented By The Secretary Of The Army | Enclosure for an unmanned aerial system |
| CN108693372B (en) * | 2018-04-13 | 2020-07-07 | 南京航空航天大学 | Course axis angular velocity estimation method of four-rotor aircraft |
| CN108759814B (en) * | 2018-04-13 | 2020-07-07 | 南京航空航天大学 | A method for estimating roll axis angular velocity and pitch axis angular velocity of quadrotor aircraft |
| US11119212B2 (en) | 2018-08-10 | 2021-09-14 | Aurora Flight Sciences Corporation | System and method to reduce DVE effect on lidar return |
| CN109062261A (en) * | 2018-08-14 | 2018-12-21 | 深圳市烽焌信息科技有限公司 | A method of guidance unmanned plane enters charging platform |
| US11037453B2 (en) | 2018-10-12 | 2021-06-15 | Aurora Flight Sciences Corporation | Adaptive sense and avoid system |
| US12153188B2 (en) | 2018-10-29 | 2024-11-26 | International Business Machines Corporation | Micro-weather report capture by unmanned aerial system (UAS) deployment |
| CN109542055A (en) * | 2018-10-31 | 2019-03-29 | 南宁学院 | A kind of automatic control aquatic farm master control platform control method |
| IT201800010936A1 (en) * | 2018-12-10 | 2020-06-10 | Cap Holding Spa | System and method of data acquisition from a plurality of smart meters through autonomous flight of one or more drones |
| CN109885088B (en) * | 2019-03-12 | 2021-01-19 | 西安交通大学 | Unmanned aerial vehicle flight trajectory optimization method based on machine learning in edge computing network |
| CA3134042A1 (en) * | 2019-03-18 | 2020-12-17 | Daniel Baumgartner | Drone-assisted systems and methods of calculating a ballistic solution for a projectile |
| DE102019106946A1 (en) * | 2019-03-19 | 2020-09-24 | Ebm-Papst Mulfingen Gmbh & Co. Kg | Position determination system and method for position determination of fans |
| CN110087204B (en) * | 2019-04-09 | 2022-05-24 | 深圳大学 | A UAV-Internet of Things data collection method and system based on parameter adjustment |
| CN110062345B (en) * | 2019-04-09 | 2022-03-22 | 深圳大学 | A UAV-Internet of Things data collection method and system |
| CN110119155A (en) * | 2019-05-20 | 2019-08-13 | 南京邮电大学 | A kind of automatic obstacle avoiding system and its implementation based on quadrotor drone |
| CN110705524B (en) * | 2019-10-24 | 2023-12-29 | 佛山科学技术学院 | Visual-based monitoring method and device for unmanned aerial vehicle in specific area |
| US11052851B1 (en) * | 2020-03-24 | 2021-07-06 | Kiomars Anvari | Use of IoT network and IoT ranging device for a navigation and protection system |
| US12030668B2 (en) * | 2020-04-06 | 2024-07-09 | Workhorse Group Inc. | Flying vehicle systems and methods |
| US11440679B2 (en) * | 2020-10-27 | 2022-09-13 | Cowden Technologies, Inc. | Drone docking station and docking module |
| CN114051249B (en) | 2022-01-06 | 2022-04-15 | 杭州筋斗腾云科技有限公司 | Digital office system and method based on mobile equipment |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20030221821A1 (en) * | 2002-05-31 | 2003-12-04 | Patel Chandrakant D. | Controlled cooling of a data center |
| US20040160897A1 (en) * | 1999-10-27 | 2004-08-19 | Netbotz, Inc. | Method and system for monitoring computer networks and equipment |
| US20040243280A1 (en) * | 2003-05-29 | 2004-12-02 | Bash Cullen E. | Data center robotic device |
| US20050038562A1 (en) * | 2003-08-13 | 2005-02-17 | Bash Cullen E. | Semi-autonomous operation of a robotic device |
Family Cites Families (13)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5455409A (en) | 1993-08-16 | 1995-10-03 | Texas Digital Systems, Inc. | Apparatus and method for monitoring a plurality of coded articles and for identifying the location of selected articles |
| JP2000147085A (en) | 1998-11-11 | 2000-05-26 | Kansai Electric Power Co Inc:The | Three-dimensional automatic moving body, three- dimensional automatic navigating flying vehicle, automatic weather observation airship, automatic atmospheric air survey airship and raptors automatic observing airship |
| WO2005081012A1 (en) * | 2004-02-17 | 2005-09-01 | Jadi, Inc. | Ultra wide band navigation system with mobile base stations |
| US9182480B2 (en) | 2005-01-28 | 2015-11-10 | Hewlett-Packard Development Company, L.P. | Information technology (IT) equipment positioning system |
| US7725212B2 (en) | 2005-10-21 | 2010-05-25 | Hewlett-Packard Development Company, L.P. | Datacenter with automated robotic maintenance |
| US7532119B2 (en) | 2005-11-08 | 2009-05-12 | Hewlett-Packard Development Company, L.P. | Multi-tiered network for gathering detected condition information |
| US7581702B2 (en) | 2006-06-09 | 2009-09-01 | Insitu, Inc. | Wirelessly controlling unmanned aircraft and accessing associated surveillance data |
| US20080144884A1 (en) * | 2006-07-20 | 2008-06-19 | Babak Habibi | System and method of aerial surveillance |
| US8554478B2 (en) * | 2007-02-23 | 2013-10-08 | Honeywell International Inc. | Correlation position determination |
| US20080299939A1 (en) * | 2007-05-30 | 2008-12-04 | Vincent Apodaca | Emergency beacon for cell phone or the like |
| US7642914B2 (en) | 2007-09-14 | 2010-01-05 | International Business Machines Corporation | Auto-locating system and method for data center mapping and monitoring |
| FR2940607B1 (en) * | 2008-12-29 | 2012-04-06 | Didier Capon | GLENOIDAL IMPLANT COMPRISING A CUP FOR COOPERATING WITH A PROTHETIC HUMERAL HEAD |
| US8521418B2 (en) * | 2011-09-26 | 2013-08-27 | Honeywell International Inc. | Generic surface feature extraction from a set of range data |
-
2011
- 2011-08-31 US US13/222,230 patent/US20120271461A1/en not_active Abandoned
-
2012
- 2012-04-20 US US13/451,922 patent/US8818705B2/en active Active
- 2012-04-20 EP EP12002777.6A patent/EP2515147B1/en active Active
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20040160897A1 (en) * | 1999-10-27 | 2004-08-19 | Netbotz, Inc. | Method and system for monitoring computer networks and equipment |
| US20030221821A1 (en) * | 2002-05-31 | 2003-12-04 | Patel Chandrakant D. | Controlled cooling of a data center |
| US20040243280A1 (en) * | 2003-05-29 | 2004-12-02 | Bash Cullen E. | Data center robotic device |
| US20050038562A1 (en) * | 2003-08-13 | 2005-02-17 | Bash Cullen E. | Semi-autonomous operation of a robotic device |
Cited By (114)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10574614B2 (en) | 2009-08-03 | 2020-02-25 | Picpocket Labs, Inc. | Geofencing of obvious geographic locations and events |
| US10856115B2 (en) | 2009-08-03 | 2020-12-01 | Picpocket Labs, Inc. | Systems and methods for aggregating media related to an event |
| US20120296497A1 (en) * | 2011-05-18 | 2012-11-22 | Hon Hai Precision Industry Co., Ltd. | Unmanned aerial vehicle and method for controlling the unmanned aerial vehicle |
| US8554462B2 (en) * | 2011-05-18 | 2013-10-08 | Hon Hai Precision Industry Co., Ltd. | Unmanned aerial vehicle and method for controlling the unmanned aerial vehicle |
| US10846497B2 (en) | 2011-12-05 | 2020-11-24 | Adasa Inc. | Holonomic RFID reader |
| US9780435B2 (en) | 2011-12-05 | 2017-10-03 | Adasa Inc. | Aerial inventory antenna |
| US11093722B2 (en) | 2011-12-05 | 2021-08-17 | Adasa Inc. | Holonomic RFID reader |
| US10476130B2 (en) | 2011-12-05 | 2019-11-12 | Adasa Inc. | Aerial inventory antenna |
| US10050330B2 (en) | 2011-12-05 | 2018-08-14 | Adasa Inc. | Aerial inventory antenna |
| US9747480B2 (en) | 2011-12-05 | 2017-08-29 | Adasa Inc. | RFID and robots for multichannel shopping |
| US12131656B2 (en) | 2012-05-09 | 2024-10-29 | Singularity University | Transportation using network of unmanned aerial vehicles |
| US20170122909A1 (en) * | 2012-10-27 | 2017-05-04 | Valerian Goroshevskiy | Non-destructive system and method for detecting structural defects |
| US9964519B2 (en) * | 2012-10-27 | 2018-05-08 | Valerian Goroshevskiy | Non-destructive system and method for detecting structural defects |
| US20150338855A1 (en) * | 2013-02-15 | 2015-11-26 | Disney Enterprises, Inc. | Controlling unmanned aerial vehicles as a flock to synchronize flight in aerial displays |
| US9809306B2 (en) * | 2013-02-15 | 2017-11-07 | Disney Enterprises, Inc. | Controlling unmanned aerial vehicles as a flock to synchronize flight in aerial displays |
| US9533759B2 (en) | 2013-03-15 | 2017-01-03 | Azure Sky Group LLC. | Modular drone and methods for use |
| US8989922B2 (en) * | 2013-03-15 | 2015-03-24 | Azure Sky Group, LLC. | Modular drone and methods for use |
| US20140277854A1 (en) * | 2013-03-15 | 2014-09-18 | Azure Sky Group Llc | Modular drone and methods for use |
| US9573684B2 (en) * | 2013-10-26 | 2017-02-21 | Amazon Technologies, Inc. | Unmanned aerial vehicle delivery system |
| US12394324B2 (en) | 2013-10-26 | 2025-08-19 | Amazon Technologies, Inc. | Aerial vehicle delivery location |
| US11749125B2 (en) | 2013-10-26 | 2023-09-05 | Amazon Technologies, Inc. | Aerial vehicle delivery location |
| US10403155B2 (en) | 2013-10-26 | 2019-09-03 | Amazon Technologies, Inc. | Aerial vehicle delivery of items available through an E-commerce shopping site |
| US11195422B2 (en) | 2013-10-26 | 2021-12-07 | Amazon Technologies, Inc. | Aerial vehicle delivery location |
| US20150120094A1 (en) * | 2013-10-26 | 2015-04-30 | Amazon Technologies, Inc. | Unmanned aerial vehicle delivery system |
| US20150161712A1 (en) * | 2013-12-10 | 2015-06-11 | 12 Retail (HK) Limited | Unifying shopping experience system |
| US10434885B2 (en) | 2014-08-05 | 2019-10-08 | Telecom Italia S.P.A. | Landing platform for an unmanned aerial vehicle |
| US11710191B2 (en) | 2014-09-22 | 2023-07-25 | State Farm Mutual Automobile Insurance Company | Insurance underwriting and re-underwriting implementing unmanned aerial vehicles (UAVs) |
| US11816736B2 (en) | 2014-09-22 | 2023-11-14 | State Farm Mutual Automobile Insurance Company | Insurance underwriting and re-underwriting implementing unmanned aerial vehicles (UAVs) |
| US12020330B2 (en) | 2014-09-22 | 2024-06-25 | State Farm Mutual Automobile Insurance Company | Accident reconstruction implementing unmanned aerial vehicles (UAVs) |
| US12033221B2 (en) | 2014-09-22 | 2024-07-09 | State Farm Mutual Automobile Insurance Company | Insurance underwriting and re-underwriting implementing unmanned aerial vehicles (UAVs) |
| US12062097B1 (en) * | 2014-09-22 | 2024-08-13 | State Farm Mutual Automobile Insurance Company | Disaster damage analysis and loss mitigation implementing unmanned aerial vehicles (UAVs) |
| JP2017534513A (en) * | 2014-10-02 | 2017-11-24 | スウィフト・エンジニアリング・インコーポレーテッド | Portable ground station for unmanned aerial vehicles |
| US9802701B1 (en) | 2014-10-21 | 2017-10-31 | Joshua Hawes | Variable elevation signal acquisition and data collection system and method |
| US9429945B2 (en) * | 2014-10-22 | 2016-08-30 | Honeywell International Inc. | Surveying areas using a radar system and an unmanned aerial vehicle |
| US20160116915A1 (en) * | 2014-10-22 | 2016-04-28 | Honeywell International Inc. | Surveying areas using a radar system and an unmanned aerial vehicle |
| US10530506B2 (en) | 2014-12-04 | 2020-01-07 | At&T Intellectual Property I, L.P. | Communicate an alert via an unmanned vehicle |
| US10097293B2 (en) | 2014-12-04 | 2018-10-09 | At&T Intellectual Property I, L.P. | Communicate an alert via an unmanned vehicle |
| US10640208B2 (en) | 2014-12-17 | 2020-05-05 | Picpocket Labs, Inc. | Drone based systems and methodologies for capturing images |
| WO2016100601A3 (en) * | 2014-12-17 | 2016-08-11 | Picpocket, Inc. | Drone based systems and methodologies for capturing images |
| US10528049B2 (en) | 2015-01-04 | 2020-01-07 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
| US10824149B2 (en) | 2015-01-04 | 2020-11-03 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
| US10719080B2 (en) | 2015-01-04 | 2020-07-21 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system and detachable housing |
| US10220954B2 (en) | 2015-01-04 | 2019-03-05 | Zero Zero Robotics Inc | Aerial system thermal control system and method |
| US10222800B2 (en) | 2015-01-04 | 2019-03-05 | Hangzhou Zero Zero Technology Co., Ltd | System and method for automated aerial system operation |
| US10824167B2 (en) | 2015-01-04 | 2020-11-03 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
| US10126745B2 (en) * | 2015-01-04 | 2018-11-13 | Hangzhou Zero Zero Technology Co., Ltd. | System and method for automated aerial system operation |
| US10358214B2 (en) | 2015-01-04 | 2019-07-23 | Hangzhou Zero Zro Technology Co., Ltd. | Aerial vehicle and method of operation |
| US10785323B2 (en) | 2015-01-05 | 2020-09-22 | Picpocket Labs, Inc. | Use of a dynamic geofence to control media sharing and aggregation associated with a mobile target |
| JP2016197404A (en) * | 2015-03-25 | 2016-11-24 | 横河電機株式会社 | System and method for monitoring industrial plant |
| US12067885B2 (en) | 2015-03-31 | 2024-08-20 | SZ DJI Technology Co., Ltd. | Systems and methods for geo-fencing device communications |
| US11367081B2 (en) | 2015-03-31 | 2022-06-21 | SZ DJI Technology Co., Ltd. | Authentication systems and methods for generating flight regulations |
| US11961093B2 (en) | 2015-03-31 | 2024-04-16 | SZ DJI Technology Co., Ltd. | Authentication systems and methods for generating flight regulations |
| CN107430402A (en) * | 2015-03-31 | 2017-12-01 | 深圳市大疆创新科技有限公司 | Systems and methods for identifying and authenticating geofencing devices |
| US11120456B2 (en) | 2015-03-31 | 2021-09-14 | SZ DJI Technology Co., Ltd. | Authentication systems and methods for generating flight regulations |
| US11094202B2 (en) | 2015-03-31 | 2021-08-17 | SZ DJI Technology Co., Ltd. | Systems and methods for geo-fencing device communications |
| US9439092B1 (en) * | 2015-07-27 | 2016-09-06 | Sprint Communications Company L.P. | Detection of component fault at cell towers |
| JP2017071233A (en) * | 2015-10-05 | 2017-04-13 | 廣田 祐次 | Drone shelter |
| CN108353290A (en) * | 2015-10-15 | 2018-07-31 | T移动美国公司 | Dynamic radio communications network with multiple aerial unmanned planes |
| WO2017066647A1 (en) * | 2015-10-15 | 2017-04-20 | T-Mobile Usa, Inc. | Dynamic wireless communications network with a plurality of aerial drones |
| US10988038B2 (en) | 2015-10-15 | 2021-04-27 | T-Mobile Usa, Inc. | Dynamic wireless communications network with a plurality of aerial drones |
| US10207590B2 (en) | 2015-10-15 | 2019-02-19 | T-Mobile Usa, Inc. | Dynamic wireless communications network with a plurality of aerial drones |
| US20220073204A1 (en) * | 2015-11-10 | 2022-03-10 | Matternet, Inc. | Methods and systems for transportation using unmanned aerial vehicles |
| US11820507B2 (en) * | 2015-11-10 | 2023-11-21 | Matternet, Inc. | Methods and systems for transportation using unmanned aerial vehicles |
| US10761525B2 (en) | 2015-12-30 | 2020-09-01 | Skydio, Inc. | Unmanned aerial vehicle inspection system |
| US11550315B2 (en) | 2015-12-30 | 2023-01-10 | Skydio, Inc. | Unmanned aerial vehicle inspection system |
| US12007761B2 (en) | 2015-12-30 | 2024-06-11 | Skydio, Inc. | Unmanned aerial vehicle inspection system |
| US9513635B1 (en) | 2015-12-30 | 2016-12-06 | Unmanned Innovation, Inc. | Unmanned aerial vehicle inspection system |
| US9740200B2 (en) | 2015-12-30 | 2017-08-22 | Unmanned Innovation, Inc. | Unmanned aerial vehicle inspection system |
| US20170193829A1 (en) * | 2015-12-31 | 2017-07-06 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US9618940B1 (en) | 2015-12-31 | 2017-04-11 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US20250006063A1 (en) * | 2015-12-31 | 2025-01-02 | Skydio, Inc. | Unmanned Aerial Vehicle Rooftop Inspection System |
| US9613538B1 (en) | 2015-12-31 | 2017-04-04 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US10083616B2 (en) * | 2015-12-31 | 2018-09-25 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US10061470B2 (en) | 2015-12-31 | 2018-08-28 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US9915946B2 (en) | 2015-12-31 | 2018-03-13 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US20220148445A1 (en) * | 2015-12-31 | 2022-05-12 | Skydio, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US9609288B1 (en) | 2015-12-31 | 2017-03-28 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US9881213B2 (en) | 2015-12-31 | 2018-01-30 | Unmanned Innovation, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US12039875B2 (en) * | 2015-12-31 | 2024-07-16 | Skydio, Inc. | Unmanned aerial vehicle rooftop inspection system |
| US9594372B1 (en) * | 2016-01-21 | 2017-03-14 | X Development Llc | Methods and systems for providing feedback based on information received from an aerial vehicle |
| US10258534B1 (en) * | 2016-01-21 | 2019-04-16 | Wing Aviation Llc | Methods and systems for providing feedback based on information received from an aerial vehicle |
| WO2017151156A1 (en) * | 2016-03-02 | 2017-09-08 | Lazo-Antunez Raynel | Automated hands-free umbrella |
| US11027833B2 (en) | 2016-04-24 | 2021-06-08 | Hangzhou Zero Zero Technology Co., Ltd. | Aerial system propulsion assembly and method of use |
| US12504454B2 (en) | 2016-05-18 | 2025-12-23 | Skydio, Inc. | Unmanned aerial vehicle electromagnetic avoidance and utilization system |
| US11835561B2 (en) | 2016-05-18 | 2023-12-05 | Skydio, Inc. | Unmanned aerial vehicle electromagnetic avoidance and utilization system |
| US11029352B2 (en) | 2016-05-18 | 2021-06-08 | Skydio, Inc. | Unmanned aerial vehicle electromagnetic avoidance and utilization system |
| US11796673B2 (en) | 2016-07-06 | 2023-10-24 | Lawrence Livermore National Security, Llc | Object sense and avoid system for autonomous vehicles |
| US10353052B2 (en) * | 2016-09-15 | 2019-07-16 | Lawrence Livermore National Security, Llc | Object discrimination based on a swarm of agents |
| US20190285730A1 (en) * | 2016-09-15 | 2019-09-19 | Lawerence Livermore National Security, Llc | Object discrimination based on a swarm of agents |
| US10935635B2 (en) * | 2016-09-15 | 2021-03-02 | Lawrence Livermore National Security, Llc | Object discrimination based on a swarm of agents |
| WO2018148636A1 (en) * | 2017-02-13 | 2018-08-16 | Top Flight Technologies, Inc. | Weather sensing |
| US10405223B1 (en) | 2017-02-14 | 2019-09-03 | Sprint Communications Company L.P. | System and methods for intelligent reset delay for cell sites in a network |
| US11573575B2 (en) | 2017-04-12 | 2023-02-07 | Lawrence Livermore National Security, Llc | Attract-repel path planner system for collision avoidance |
| US11086337B2 (en) | 2017-06-20 | 2021-08-10 | Planck Aerosystems Inc. | Systems and methods for charging unmanned aerial vehicles on a moving platform |
| US12084179B2 (en) | 2018-05-23 | 2024-09-10 | Aerovironment, Inc. | System and method for drone tethering |
| US20210163134A1 (en) * | 2018-06-14 | 2021-06-03 | Beijing Xiaomi Mobile Software Co., Ltd. | Information sending and receiving method and apparatus, device, and storage medium |
| US11760480B2 (en) * | 2018-06-14 | 2023-09-19 | Beijing Xiaomi Mobile Software Co., Ltd. | Information sending and receiving method and apparatus, device, and storage medium |
| JP2019055774A (en) * | 2018-10-18 | 2019-04-11 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | System for operating unmanned aircraft |
| CN112912312A (en) * | 2018-10-19 | 2021-06-04 | 安炳烈 | Unmanned aerial vehicle who possesses conflict prevention and retrieve function |
| US11503111B2 (en) | 2018-12-03 | 2022-11-15 | At&T Intellectual Property I, L.P. | Global internet of things (IoT) quality of service (QoS) realization through collaborative edge gateways |
| US11108849B2 (en) | 2018-12-03 | 2021-08-31 | At&T Intellectual Property I, L.P. | Global internet of things (IOT) quality of service (QOS) realization through collaborative edge gateways |
| US10534068B2 (en) * | 2018-12-27 | 2020-01-14 | Intel Corporation | Localization system, vehicle control system, and methods thereof |
| US12228407B2 (en) | 2019-01-15 | 2025-02-18 | Aerovironment, Inc. | Systems and methods for delivery using unmanned aerial vehicles |
| US10659144B1 (en) | 2019-01-31 | 2020-05-19 | At&T Intellectual Property I, L.P. | Management of massively distributed internet of things (IOT) gateways based on software-defined networking (SDN) via fly-by master drones |
| US10887001B2 (en) | 2019-01-31 | 2021-01-05 | At&T Intellectual Property I, L.P. | Management of massively distributed internet of things (IoT) gateways based on software-defined networking (SDN) via fly-by master drones |
| CN110130879A (en) * | 2019-04-28 | 2019-08-16 | 中国石油天然气股份有限公司 | Method and system for collecting oil well parameters |
| CN111986394A (en) * | 2019-05-21 | 2020-11-24 | 顺丰科技有限公司 | A security system and method for delivery of goods |
| CN113589222A (en) * | 2020-04-07 | 2021-11-02 | 英西图公司 | Aircraft guidance with beacon transmission |
| US11927972B2 (en) | 2020-11-24 | 2024-03-12 | Lawrence Livermore National Security, Llc | Collision avoidance based on traffic management data |
| WO2023025200A1 (en) * | 2021-08-25 | 2023-03-02 | 深圳市道通智能航空技术股份有限公司 | Control method and apparatus for forced landing of unmanned aerial vehicle, remote control apparatus, and storage medium |
| US12145753B2 (en) * | 2022-08-09 | 2024-11-19 | Pete Bitar | Compact and lightweight drone delivery device called an ArcSpear electric jet drone system having an electric ducted air propulsion system and being relatively difficult to track in flight |
| US20240239531A1 (en) * | 2022-08-09 | 2024-07-18 | Pete Bitar | Compact and Lightweight Drone Delivery Device called an ArcSpear Electric Jet Drone System Having an Electric Ducted Air Propulsion System and Being Relatively Difficult to Track in Flight |
| US12473108B1 (en) * | 2023-12-15 | 2025-11-18 | Amazon Technologies, Inc. | Stations for guided docking evolutions by aerial vehicles |
| CN119247513A (en) * | 2024-12-05 | 2025-01-03 | 南京大桥机器有限公司 | A mobile delivery control system and method for intelligent weather detection device |
Also Published As
| Publication number | Publication date |
|---|---|
| US20120271491A1 (en) | 2012-10-25 |
| EP2515147B1 (en) | 2016-10-26 |
| EP2515147A3 (en) | 2013-04-17 |
| US8818705B2 (en) | 2014-08-26 |
| EP2515147A2 (en) | 2012-10-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US8818705B2 (en) | Capturing environmental information | |
| US20230316930A1 (en) | Systems and methods for autonomous hazardous area data collection | |
| TWI817961B (en) | Aerial robotic vehicle with adjustable object avoidance proximity threshold and method and processing device for the same | |
| US11897630B2 (en) | Drone landing ground station with magnetic fields | |
| CN111566009B (en) | Method and apparatus for adjusting flight parameters of an aerial robotic vehicle | |
| US9794942B1 (en) | System and method for saving energy in a locator apparatus | |
| US20180088098A1 (en) | System, Method, and Apparatus for Condition Monitoring of Food and Other Perishable Products As Well As Environmentally Sensitive Industrial Supply Chains | |
| CN108513640B (en) | Control method of movable platform and movable platform | |
| US11479357B1 (en) | Perspective angle acquisition and adjustment of security camera drone | |
| US20190049275A1 (en) | Method, a circuit and a system for environmental sensing | |
| CN110487119A (en) | A kind of unmanned plane being integrated in wisdom radio car shoots down system and shoots down method | |
| CN109477891B (en) | Unmanned aerial vehicle detection method, detection equipment, control method of server and server | |
| US20220053292A1 (en) | Low Cost, High Performance Asset Tracking Systems and Methods | |
| KR20180129240A (en) | Intrusion object detection system and method for controlling the system thereof | |
| US11619747B2 (en) | Location monitoring apparatuses configured for low-power operation | |
| US9949086B2 (en) | Automatic system control based on mobile device location relative to a physical space | |
| JP2024135386A (en) | Flying robot control system and flying robot control method | |
| KR20210016678A (en) | Apparatus and method for controlling a unmanned aerial vehicle | |
| JP7287690B2 (en) | Unmanned aircraft, fault diagnosis device, fault diagnosis method, and program | |
| CN207867349U (en) | A kind of substation inspection unmanned plane during flying security system | |
| US20250040477A1 (en) | Mobile robot and control method thereof | |
| CN111279399A (en) | Control method, control system, computer-readable storage medium, and flightable device | |
| CN110134144A (en) | Rotary-wing unmanned aerial vehicle multiple fire extinguishing water bomb device, control system, method and unmanned aerial vehicle | |
| JP2024135384A (en) | Flying robot control system and flying robot control method | |
| JP2024135388A (en) | Flying robot control system and flying robot control method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: ACCENTURE GLOBAL SERVICE LIMITED, IRELAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SPATA, GREGORY P.;REEL/FRAME:026842/0728 Effective date: 20110831 |
|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION |