WO2019041516A1 - 屏幕状态自动检测机器人、方法及计算机可读存储介质 - Google Patents
屏幕状态自动检测机器人、方法及计算机可读存储介质 Download PDFInfo
- Publication number
- WO2019041516A1 WO2019041516A1 PCT/CN2017/108740 CN2017108740W WO2019041516A1 WO 2019041516 A1 WO2019041516 A1 WO 2019041516A1 CN 2017108740 W CN2017108740 W CN 2017108740W WO 2019041516 A1 WO2019041516 A1 WO 2019041516A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- preset
- display
- service device
- detected
- image
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 81
- 230000005856 abnormality Effects 0.000 claims abstract description 22
- 239000003086 colorant Substances 0.000 claims description 39
- 238000012545 processing Methods 0.000 claims description 17
- 239000007787 solid Substances 0.000 claims description 16
- 238000001914 filtration Methods 0.000 claims description 12
- 238000007781 pre-processing Methods 0.000 claims description 11
- 238000004891 communication Methods 0.000 claims description 10
- 230000000873 masking effect Effects 0.000 claims description 7
- 238000012806 monitoring device Methods 0.000 claims description 6
- 238000001514 detection method Methods 0.000 description 32
- 230000000694 effects Effects 0.000 description 15
- 238000007796 conventional method Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 6
- 230000011218 segmentation Effects 0.000 description 6
- 230000002159 abnormal effect Effects 0.000 description 4
- 238000000605 extraction Methods 0.000 description 4
- 238000002594 fluoroscopy Methods 0.000 description 3
- 238000012937 correction Methods 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000007689 inspection Methods 0.000 description 2
- 230000001788 irregular Effects 0.000 description 2
- 238000012804 iterative process Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 238000004590 computer program Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0246—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means
- G05D1/0253—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using a video camera in combination with image processing means extracting relative motion information from a plurality of images taken successively, e.g. visual odometry, optical flow
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0214—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory in accordance with safety or protection criteria, e.g. avoiding hazardous areas
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0259—Control of position or course in two dimensions specially adapted to land vehicles using magnetic or electromagnetic means
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0276—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle
- G05D1/028—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle using a RF signal
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0276—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle
- G05D1/0285—Control of position or course in two dimensions specially adapted to land vehicles using signals provided by a source external to the vehicle using signals transmitted via a public communication network, e.g. GSM network
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06K—GRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K7/00—Methods or arrangements for sensing record carriers, e.g. for reading patterns
- G06K7/10—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
- G06K7/14—Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation using light without selection of wavelength, e.g. sensing reflected white light
- G06K7/1404—Methods for optical code recognition
- G06K7/1408—Methods for optical code recognition the method being specifically adapted for the type of code
- G06K7/1417—2D bar codes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/187—Segmentation; Edge detection involving region growing; involving region merging; involving connected component labelling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30121—CRT, LCD or plasma display
Definitions
- the present application relates to the field of terminal technologies, and in particular, to a screen state automatic detecting robot, a method, and a computer readable storage medium.
- unmanned network areas for example, bank self-service outlets
- the store has business equipment for different businesses, such as deposit and withdrawal machines, ATMs, self-service payment machines, and PCs for customers (for example, PCs designed for large customers, you can check financial management, funds , stock information, etc.), self-checking equipment, numbering machine, promotional equipment, etc., these business equipment for different services usually need to configure the terminal screen.
- the terminal screen is a delicate electronic device, and display problems often occur. Once the display problem affects the customer's use, the industry usually needs to periodically check and maintain the status of the screen of each self-service device in the unmanned network area.
- the scheme for detecting the screen state of the self-service device usually includes: scheme 1, feedback through the manual inspection, and scheme 2, and feedback through the customer complaint.
- the drawback of the above scheme 1 is that the manual inspection method is labor-intensive, the timeliness is relatively poor, and some details may not be carefully checked. For example, the small crack artificial is not necessarily carefully observed, and the error is likely to occur.
- the shortcoming of the above solution 2 is that it will affect the user experience, and many users may not give feedback. Even if the user gives feedback, the problem usually is not comprehensive and objective, not only the timeliness is bad, but also the error will occur. Therefore, how to automatically and accurately detect the screen state of each self-service device in the unmanned network area has become a technical problem to be solved urgently.
- the present application provides a screen state automatic detecting robot, a method, and a computer readable storage medium, the main purpose of which is to automatically detect the screen state of each self-service device in the unmanned network point area without human power.
- the present application provides a screen state automatic detecting robot, which includes a memory and a processor, and the memory stores a screen state automatic detecting program, when the screen state automatic detecting program is executed by the processor Implement the following operations:
- the service device corresponding to the preset area is used as a service device to be detected, and a preset graphic code is sent to the service device to be detected, according to the preset graphic code. Determining, by the content displayed on the display screen, whether the circuit to be detected has a circuit failure;
- the device to be detected is controlled to perform image display according to the preset display parameter, and the image displayed on the display screen is analyzed to analyze the service to be detected. Whether the preset type of exception has occurred on the display screen of the device.
- the present application further provides a method for automatically detecting a screen state, the method comprising:
- the service device corresponding to the preset area is regarded as a to-be-tested industry. And sending a preset graphic code to the service device to be detected, to determine whether a circuit fault occurs in the service device to be detected according to the content displayed on the display screen by the preset graphic code;
- the device to be detected is controlled to perform image display according to the preset display parameter, and the image displayed on the display screen is analyzed to analyze the service to be detected. Whether the preset type of exception has occurred on the display screen of the device.
- the present application further provides a computer readable storage medium having a screen state automatic detection program stored thereon, the screen state automatic detection program being configurable by one or more processors Execute to implement the following steps:
- the service device corresponding to the preset area is used as a service device to be detected, and a preset graphic code is sent to the service device to be detected, according to the preset graphic code. Determining, by the content displayed on the display screen, whether the circuit to be detected has a circuit failure;
- the device to be detected is controlled to perform image display according to the preset display parameter, and the image displayed on the display screen is analyzed to analyze the service to be detected. Whether the preset type of exception has occurred on the display screen of the device.
- the screen state automatic detecting robot, the method and the computer readable storage medium proposed by the application respectively moves to a preset area of each service device of the unmanned network point area, and if the robot moves to a preset area of a special service device, And sending a preset graphic code to the service device to be detected, to determine whether a circuit fault occurs in the service device to be detected according to the content displayed on the display screen by the preset graphic code, and if the circuit fault does not occur, controlling the service device to be detected
- the graphic display is performed according to the preset display parameter, and the image displayed on the display screen is analyzed according to the preset analysis rule to analyze whether the display screen of the service device to be detected has a preset type abnormality, and the solution does not need manual Participate in the automatic detection of the circuit fault of the device and the screen display state by moving the robot to the corresponding area.
- FIG. 1 is a schematic diagram of a preferred embodiment of a screen state automatic detecting robot of the present application
- 2a to 2d are two-dimensional codes in which partial information is occluded
- 3a is a two-dimensional code in which perspective is generated;
- FIG. 3b is an outer contour model of the two-dimensional code after seeing;
- FIG. 3c is an effect of reversely seeing the two-dimensional code after see-through;
- FIG. 4 is a schematic view of display areas of different sizes set on a display screen
- Figure 5 is a boundary of a small-sized rectangular area and a medium-sized rectangular display area
- Figure 6 is an outer contour curve of a large-sized rectangular area
- Figure 7 is a schematic illustration of a plurality of points taken from a rectangular boundary
- FIG. 8 is a schematic diagram of a straight line obtained by a least squares straight line fitting algorithm for acquired points
- Figure 9 is a screen state before performing noise filtering enhancement preprocessing
- Figure 10 is a screen state after performing noise filtering enhancement preprocessing
- Figure 11 is a schematic view showing a state in which a bad line appears on the screen
- Figure 12 is a schematic view showing a state in which a bad spot appears on the screen
- Figure 13 is a schematic view showing a state in which a crack occurs on the screen
- Figure 14 is a schematic view showing a state of a lateral crack appearing on a screen
- 15 is a schematic diagram of a program module of an automatic screen state detecting program in an embodiment of a screen state automatic detecting robot of the present application;
- 16 is a flowchart of a preferred embodiment of a screen state automatic detecting method according to the present application.
- FIG. 17 is a flowchart of a second embodiment of a method for automatically detecting a screen state according to the present application.
- the application provides a screen state automatic detecting robot.
- FIG. 1 a schematic diagram of a preferred embodiment of a screen state automatic detecting robot of the present application is shown.
- the screen state automatic detecting robot includes a memory 11, a processor 12, a communication bus 13, a network interface 14, and a camera 15.
- the memory 11 includes at least one type of readable storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (for example, an SD or DX memory, etc.), a magnetic memory, a magnetic disk, an optical disk, and the like.
- the memory 11 may be an internal storage unit of the screen state automatic detection robot in some embodiments, for example, the screen state automatically detects the hard disk of the robot.
- the memory 11 may also be an external storage device of the screen state automatic detection robot in other embodiments, such as a plug-in hard disk equipped on the screen state automatic detection robot, a smart memory card (SMC), and a secure digital (Secure) Digital, SD) cards, flash cards, etc.
- the memory 11 may also include both an internal storage unit of the screen state automatic detection robot and an external storage device.
- the memory 11 can be used not only for storing application software and various types of data installed in the screen state automatic detection robot, such as code of the screen state automatic detection program, but also for temporarily storing data that has been output or is to be output.
- the processor 12 may be a Central Processing Unit (CPU), controller, microcontroller, microprocessor or other data processing chip for running program code or processing stored in the memory 11. Data, such as performing a screen state automatic detection program, and the like.
- CPU Central Processing Unit
- controller microcontroller
- microprocessor or other data processing chip for running program code or processing stored in the memory 11.
- Data such as performing a screen state automatic detection program, and the like.
- Communication bus 13 is used to implement connection communication between these components.
- the network interface 14 can optionally include a standard wired interface, a wireless interface (such as a WI-FI interface), and is typically used to establish a communication connection between the robot and other electronic devices.
- the robot can connect to the user terminal through the network interface 14, receive the detection command sent by the user terminal, or connect to the service device through the network interface 14, and control the service device to display the image according to the preset display parameter.
- the camera 15 is used to collect the display content of the display screen of the business device.
- Figure 1 shows only the screen state automatic detection robot with components 11-15 and the screen state automatic detection program, but it should be understood that not all illustrated components are required to be implemented, and alternative implementations may be implemented more or less. Component.
- the robot may further include a user interface
- the user interface may include a display, an input unit such as a keyboard, a physical button for triggering the detection instruction, and the like
- the optional user interface may further include a standard wired Interface, wireless interface.
- the display may be an LED display, a liquid crystal display, a touch-sensitive liquid crystal display, an OLED (Organic Light-Emitting Diode) touch sensor, or the like.
- the display may also be referred to as a display screen or a display unit as appropriate.
- the robot may further include an RF (Radio Frequency) circuit, a sensor, a WiFi module, and the like.
- the sensor may be a light sensor, a distance sensor, or the like.
- a screen state automatic detection program is stored in the memory 11; when the processor 12 executes the program, the following operations are implemented:
- the robot provided in this embodiment may be placed in an unmanned network point area, and the service equipment in the unmanned network point area may have one or more.
- the robot can be moved to the unmanned network point area in real time, timing, in an idle state or upon receiving a detection command, and the detection of the screen display state of each service device in the unmanned network point area is turned on. Wherein, when the robot is in the process of no need to process the service for a preset period of time, it is determined that it is in an idle state.
- Each service device in the unmanned network site area is divided into preset areas, and the preset area of one service device refers to an area where the distance from the service machine is less than or equal to a preset distance, and the robot passes during the moving process.
- Positioning its own position determines the position coordinates in the unmanned network point area, and combines the known position coordinates of each service device in the unmanned network point area to calculate the current position and the predetermined position of each service device.
- Distance if current The distance between the location and the location where the service device is located is less than or equal to the preset distance, and then the robot is determined to move to the preset region of the service device.
- each service device needs to be detected one by one.
- the control robot moves to the preset area of each service device in the unmanned network point area, and two of them are described below.
- the movement of the robot is controlled according to the preset mobile navigation path.
- the control robot controls the mobile navigation according to the movement. The path continues to move to the undetected business device until the display screen of all business devices is detected and the mobile is finished.
- the robot may be controlled to move randomly. After moving to a preset area of a service device and detecting the display screen of the service device to be detected, the control robot marks the service device to be detected as an obstacle. The object, the obstacle avoidance movement, after the obstacle avoidance movement is completed, the control robot continues to perform random movement, moves to another business device for detection, until all the business equipment is marked as an obstacle and ends the movement, and the business equipment The obstacle mark is cleared.
- the principle of the obstacle avoiding movement algorithm is: when the robot detects a business device, shifts the preset angle to the left or right in the current moving direction, and judges Whether it can move unimpeded after the offset angle, if it can pass after the offset angle, continue to move randomly along the direction after the offset, if it is blocked after the offset angle, it will continue to shift in the same direction.
- the algorithm that constantly adjusts the offset angle circumvents obstacles that hinder movement.
- the service device corresponding to the preset area is used as a service device to be detected, and a preset graphic code is sent to the service device to be detected, according to the preset.
- the content displayed on the display screen by the graphic code determines whether a circuit failure has occurred in the service device to be detected.
- the service device corresponding to the preset area is used as a service device to be detected, and the service device corresponding to the preset area is detected.
- the robot can adjust the angle of the camera 15 until the display screen of the current display screen of the service device can be captured, and then send a preset graphic code to the service device to be detected to display the content on the display screen according to the preset graphic code. Determining whether a fault occurs in the service device to be detected.
- the control robot performs wireless communication with the service device to be detected (ie, the service device corresponding to the preset area currently described), and sends a display instruction for displaying the preset graphic code including the preset information to the service device to be detected.
- the service device displays the preset graphic code according to the received display instruction; the control robot scans and analyzes the display content of the display screen of the service device to be detected; if the preset information is scanned and analyzed from the display screen of the service device to be detected, It is determined that the display screen of the service device to be detected does not have a circuit fault; if the preset information is not scanned and analyzed from the display screen of the service device to be detected, it is determined that a circuit fault occurs on the display screen of the service device to be detected.
- the preset graphic code may be a two-dimensional code with preset information.
- the use of two-dimensional code for circuit fault detection has the following advantages: the two-dimensional code has strong fault tolerance and has error correction capability, as shown in Figures 2a to 2d, even if there are stains, spots, cracks, etc. on the display screen. Some information is blocked, and the QR code can still be recognized. It does not affect whether the power can be powered by the QR code, that is, whether the circuit fault occurs.
- the QR code can contain information, such as the ID of the screen.
- the service device corresponding to the preset area can be identified according to the screen ID included in the two-dimensional code, because the outer contour of the two-dimensional code is square,
- the robot is not required to be strictly identifiable to the target screen.
- the robot can easily restore the distortion caused by the perspective, as shown in Figures 3a to 3c.
- the position of the robot may cause a certain degree of perspective, which may cause the outer contour of the two-dimensional code to be displayed in the lens of the robot. Rectangle. It can be seen from Fig. 3b that the two-dimensional code does not show a square, but a quadrilateral, resulting in a "nearly large and small" perspective. However, the two-dimensional code recognition can perform reverse fluoroscopy in the case of generating perspective so that the image can be recognized normally.
- the QR code and the reverse perspective method are existing mature methods and will not be explained in detail here. Referring to Fig. 3c, the effect of the reverse perspective of the two-dimensional code is shown.
- a prompt message is sent to the predetermined monitoring device, indicating that a circuit fault occurs on the display screen of the service device to be detected needs to be processed.
- the prompt information is sent to the background server, where the prompt information includes the unique ID of the service device in which the circuit is faulty, and the network identifier of the unmanned network area where the service device to be detected is located.
- the prompt information format may be "The unattended dot area of the dot is marked as *****, and the display screen of the service device whose ID number is ***** has a circuit failure that needs to be processed.”
- control the service device to be detected to perform image display according to preset display parameters, and analyze an image displayed on the display screen to analyze the to-be-processed Check whether the display screen of the service device has a preset type of abnormality.
- S3 includes the following specific operations:
- the display screens of the service device to be detected are respectively displayed in a solid color display manner, for example, three preset colors of red, green, and blue, wherein different preset colors correspond to display areas of different sizes.
- the shape of the display area corresponding to each preset color corresponds to the shape of the maximum display area of the display screen, and the display area corresponding to one of the preset colors is the maximum display area of the display screen.
- S34 Determine, according to the acquired anti-interference image, a maximum display area of the display screen, and perform image extraction of the actual display area for the mask image corresponding to the preset color of the maximum display area according to the determined maximum display area. Extract the maximum display image, because if there is interference area in the perimeter of the screen, for example, there is “water inlet” around the screen, the screen boundary area cannot be displayed normally, then the maximum map obtained by taking the mask is incomplete, which is equivalent. The boundary interference problem is excluded from the detection, so the maximum display area needs to be obtained in the above manner.
- the screen width of the small-sized rectangle is L1
- the screen width of the medium-size rectangle is L2
- the screen width of the large-size rectangle is L3.
- the proportional relationship between the length and width of the screen is also known data.
- the maximum display area of the display screen is determined according to the anti-interference image of the preset color corresponding to the display area of the smaller size, and the preset color corresponding to the maximum display area is determined according to the determined maximum display area.
- the mask image is used to extract the image of the actual display area, and the principle of extracting the maximum display image is as follows:
- the width of the large-sized rectangle is denoted as L" 3 .
- the average value of the two values is preferably taken as the width of the large-size rectangle according to Formula 1:
- the area range of the large-sized rectangle can be obtained, and the area range of the large-size rectangle is the maximum display area of the screen, that is, the outer contour curve of the largest-size rectangular area shown in FIG. 6.
- the quadrilateral is inversely transformed into a rectangle, and the anti-interference image is acquired according to the position of the mask image rectangle.
- the boundary of the calculated small-sized rectangle and the boundary of the medium-sized rectangle are respectively calculated.
- an extension line of the boundary line can be obtained, and the extension lines of the four boundary lines are respectively obtained by the above method, and then the intersection points of the four boundary lines are determined, according to the intersection of the four boundary lines and the The four boundary lines determine the position of the quadrilateral of the mask image, that is, the range of the medium size rectangle or the small size rectangle, thereby determining the exact values of L1 and L2.
- the principle of filtering and enhancing the maximum display image is as follows:
- the image of the abnormal state of the screen is affected by non-uniform illumination, irregular gaps on the screen surface and equipment, so that the acquired image has Noise, so you need to remove these noises, while maintaining the details of the abnormal state of the screen, and as much as possible to improve the contrast between the crack and the background, which is conducive to later segmentation and recognition. It can be seen from the pre-processing of FIG. 9 and the pre-processed FIG. 10 that the noise-improved graph has less noise interference and the problem is more clearly visible.
- the present application preprocesses an image using a method of a pilot filter.
- the algorithm steps using the pilot filter are as follows: the grayscaled screen image is used as the input image P, and the grayscale screen image is also used as the guide image I, and the image filtered by the noise filter is used as q. Equation 2 enhances the input image P.
- m is the enhancement factor, which can be determined according to the actual situation.
- the filter here can be selected according to the actual situation.
- the ⁇ of the following formula is a satisfaction weight, and the enchanted satisfaction is evaluated according to the P_enhanced image effect, and the highest satisfaction of the enhanced is set to a fixed value, for example, 500.
- T n The nth iteration is denoted as T n
- T n+1 the n+1th time can be recorded as T n+1
- T n+1 the value of m after n iterations is Then there are:
- the iterative process is as follows: First we give an initial weight m as Then, substituting into formula 2 to solve the image after P_enhanced, and then evaluating the satisfaction of P_enhanced according to formula 3, and then digitizing the obtained satisfaction weight ⁇ , and then using the alphaned value and formula 5 to m weight The value is updated. Then repeat the above steps to update m. After several rounds of iteration, you can get an ideal m weight. Because each time the iteration will modify the weight m, the ⁇ value will decrease when the effect of the image P_enhanced is poor, and the final m weight will also decrease, and the weight will increase accordingly.
- Each adjustment will drive the image P_enhanced effect to a better direction, set the number of iterations to a fixed value (here set to 10), and it will have a convergence state after n iterations, that is, the image P_enhanced effect is close to the most Good state.
- the enhanced image P_enhanced is used as the input image P, and the grayscale screen image is used as the guide image I, and the filtered image is obtained as the final enhanced and smoothed image.
- S36 analyzing the enhanced image to analyze whether a preset type of abnormality occurs in a display screen of the service device to be detected, where the abnormality of the preset type includes a bad spot, a bad line, and a crack.
- the identification process of the bad line is as follows: the image preprocessed by the above steps is acquired, and the straight line is detected by using the Hough transform. When the number of bad lines is detected to be greater than 1, it is determined that there is a bad line; otherwise, the judgment does not exist. Bad line.
- the Hough transform is a feature extraction technique.
- the classic Hough transform recognizes the lines in the image and later develops to recognize arbitrary shapes, but the more common identifying shapes are circles and ellipses, which we use to identify the state of the screen loop.
- a straight line is defined as:
- ⁇ 0 is used to represent the algebraic distance of the origin to the line, and ⁇ 0 is the angle between the orthogonal line of the line and the x-axis, then:
- ( ⁇ , ⁇ ) is a representation of polar coordinates. However, if ( ⁇ , ⁇ ) is also expressed in the form of Cartesian coordinates, that is, ⁇ and ⁇ are orthogonally processed, then ( ⁇ , ⁇ ) is called a Hough space.
- a point in the Cartesian coordinate system corresponds to a sinusoid of the Hough space.
- a straight line is composed of an infinite number of points. In the Hough space, there are an infinite number of sinusoids, but these sinusoids intersect at a point ( ⁇ 0 , ⁇ 0 ), and the point is brought into Equation 7 and Equation 8 to obtain the slope of the line. And the intercept, so a straight line is determined. Therefore, when the line is identified by the Hough transform, the maximum value in the Hough space may correspond to a straight line.
- the conventional method of checking the bad line is by manual observation.
- the robot automatically detects whether there is a bad line in the preprocessed image by the above method.
- Performing a bad line detection analysis on the enhanced image to determine whether the display screen of the service device to be detected occurs The bad line. Referring to Fig. 12, there is a case where a bad spot appears on the screen.
- the process of determining the bad spot is to obtain the pre-processed picture, and use the SimpleBlobDetector algorithm to detect the presence of spots, when the number of detected spots is greater than one. , to determine the presence of bad spots on the screen, otherwise it is judged that there is no bad spot on the screen.
- the embodiment also introduces a method for selecting parameters in the SimpleBlobDetector algorithm.
- Table 1 solves the value of feature k
- the meaning of the above formula is that when the feature value is k, the probability that the judgment is that the spot is correct is as high as possible, and the lower the probability that the judgment is the spot but not actually the spot, the larger the weight wk.
- the parameter that needs to be determined is selected as the feature point of the SimpleBlobDetector algorithm with the largest wk value of each feature parameter. Therefore, it is not necessary to artificially determine the specific value of this parameter.
- the conventional method of inspecting bad spots is by manual observation.
- the robot automatically detects whether there is a bad spot on the preprocessed image by the above method.
- the screen is tested for cracks as follows:
- the crack is identified by the segmented image: after the segmentation, some connected regions can be seen, where the connected region refers to the image region formed after the crack is segmented, and when the number of connected regions is greater than 1, the image can be judged. There are cracks, and the cracks may be mesh or a single crack across the screen. If there is only one connected region, we obtain the X-axis projection distance H of the crack and the projection distance R from the y-axis, as shown in Fig. 14. If H ⁇ threshold D and R ⁇ threshold D, there is no crack in the screen.
- the threshold D is a distance value, and when H and R are simultaneously smaller than this value, the line can be considered as a crack, but a noise point. If any of H and R is greater than this threshold, it can be judged that there is a lateral or longitudinal crack on the screen, but the crack does not span the entire screen, as shown in Fig. 14 is a form of lateral crack.
- the conventional method of inspecting cracks is by manual observation, and by the method of the embodiment, the robot passes the above method.
- the preprocessed image automatically detects the presence of cracks.
- control robots are respectively moved to the preset areas of the service devices of the unmanned network point area. If the robot moves to the preset area of the service device, the preset graphic code is sent to the service device to be detected. Determining whether a circuit fault occurs in the service device to be detected according to the content displayed on the display screen by the preset graphic code. If no circuit failure occurs, controlling the service device to be detected to perform graphic display according to the preset display parameter, and displaying on the display screen The image is analyzed according to a preset analysis rule to analyze whether a preset type of abnormality has occurred on the display screen of the service device to be detected. The scheme does not require manual participation, and the circuit fault and the screen display state of the device are moved to the corresponding area by the robot. Perform automatic detection.
- the screen state automatic detection program may also be divided into one or more modules, one or more modules are stored in the memory 11 and executed by one or more processors (this implementation)
- the processor 12 is executed to complete the application
- a module referred to herein refers to a series of computer program instructions that are capable of performing a particular function.
- FIG. 15 it is a schematic diagram of a program module for automatically detecting a screen state in an embodiment of a screen state automatic detecting robot of the present application.
- the screen state automatic detecting program may be divided into a control moving module 10 and a first
- the detection module 20 and the second detection module 30, the functions or operational steps performed by the modules 10-30 are substantially the same as those of the above embodiments, and will not be described in detail herein.
- the control mobile module 10 is configured to control the robot to move to a preset area of each service device of the unmanned network point area, respectively;
- the first detecting module 20 is configured to: if the mobile device moves to a preset area, the service device corresponding to the preset area is used as a service device to be detected, and then send a preset graphic code to the to-be-detected service device, Determining whether a circuit fault occurs in the service device to be detected according to the content displayed on the display screen by the preset graphic code;
- the second detecting module 30 is configured to: if the circuit of the service device to be detected does not have a circuit fault, control the service device to be detected to perform image display according to the preset display parameter, and analyze the image displayed on the display screen, It is analyzed whether a preset type of abnormality has occurred on the display screen of the service device to be detected.
- the present application also provides a method for automatically detecting a screen state.
- a flowchart of a preferred embodiment of the screen state automatic detecting method of the present application is shown. The method can be performed by a device that can be implemented by software and/or hardware.
- the screen state automatic detecting method includes:
- Step S10 controlling the robot to move to a preset area of each service device in the unmanned network point area, respectively;
- the robot in this embodiment can be placed in the unmanned network point area, and the service equipment in the unmanned network point area can have one or more.
- the robot can be moved to the unmanned network point area in real time, timing, in an idle state or upon receiving a detection command, and the detection of the screen display state of each service device in the unmanned network point area is turned on. Wherein, when the robot is in the process of no need to process the service for a preset period of time, it is determined that it is in an idle state.
- Each service device in the unmanned network site area is divided into preset areas, and the preset area of one service device refers to an area where the distance from the service machine is less than or equal to a preset distance, and the robot passes during the moving process.
- Positioning its own position determines the position coordinates in the unmanned network point area, and combines the known position coordinates of each service device in the unmanned network point area to calculate the current position and the predetermined position of each service device. The distance, if the distance between the current location and the location of a service device is less than or equal to the preset distance, determines that the robot moves to the preset area of the service device.
- each service device needs to be detected one by one.
- the control robot moves to the preset area of each service device in the unmanned network point area, and two of them are described below.
- the movement of the robot is controlled according to the preset mobile navigation path.
- the control robot controls the mobile navigation according to the movement. The path continues to move to the undetected business device until the display screen of all business devices is detected and the mobile is finished.
- the robot may be controlled to move randomly. After moving to a preset area of a service device and detecting the display screen of the service device to be detected, the control robot marks the service device to be detected as an obstacle. The object, the obstacle avoidance movement, after the obstacle avoidance movement is completed, the control robot continues to perform random movement, moves to another business device for detection, until all the business equipment is marked as an obstacle and ends the movement, and the business equipment The obstacle mark is cleared.
- the principle of the obstacle avoiding movement algorithm is: when the robot detects a business device, shifts the preset angle to the left or right in the current moving direction, and judges Whether it can move unimpeded after the offset angle, if it can pass after the offset angle, continue to move randomly along the direction after the offset, if it is blocked after the offset angle, it will continue to shift in the same direction.
- the algorithm that constantly adjusts the offset angle circumvents obstacles that hinder movement.
- step S20 if the robot moves to a preset area, the service device corresponding to the preset area is used as a service device to be detected, and a preset graphic code is sent to the service device to be detected, according to the The content displayed on the display screen by the graphic code determines whether a circuit fault has occurred in the service device to be detected.
- the service device corresponding to the preset area is used as a service device to be detected, and the service device corresponding to the preset area is detected.
- the robot can adjust the angle of the camera 15 until the display screen of the current display screen of the service device can be captured, and then send a preset graphic code to the service device to be detected to display the content on the display screen according to the preset graphic code. Determining whether a fault occurs in the service device to be detected.
- the control robot performs wireless communication with the service device to be detected (ie, the service device corresponding to the preset area currently described), and sends a display instruction for displaying the preset graphic code including the preset information to the service device to be detected.
- the service device displays the preset graphic code according to the received display instruction; the control robot scans and analyzes the display content of the display screen of the service device to be detected; if the preset information is scanned and analyzed from the display screen of the service device to be detected, It is determined that the display screen of the service device to be detected does not have a circuit fault; if the preset information is not scanned and analyzed from the display screen of the service device to be detected, it is determined that a circuit fault occurs on the display screen of the service device to be detected.
- the preset graphic code may be a two-dimensional code with preset information.
- the use of two-dimensional code for circuit fault detection has the following advantages: the two-dimensional code has strong fault tolerance and has error correction capability, as shown in Figures 2a to 2d, even if there are stains, spots, cracks, etc. on the display screen. Some information is blocked, and the QR code can still be recognized. It does not affect whether the power can be powered by the QR code, that is, whether the circuit fault occurs.
- the QR code can contain information, such as the ID of the screen.
- the service device corresponding to the preset area can be identified according to the screen ID included in the two-dimensional code; since the outer contour of the two-dimensional code is square, Therefore, it is not required that the robot can be recognized strictly against the target screen.
- the robot can easily restore the distortion caused by the perspective, as shown in Figs. 3a to 3c.
- the position of the robot may cause a certain degree of perspective, which may cause the outer contour of the two-dimensional code to be displayed in the lens of the robot. Rectangle. It can be seen from Fig. 3b that the two-dimensional code does not show a square, but a quadrilateral, resulting in a "nearly large and small" perspective.
- the two-dimensional code recognition can perform reverse fluoroscopy in the case of generating perspective so that the two-dimensional code can be recognized normally.
- the reverse fluoroscopy method is an existing mature method and will not be explained in detail here. Referring to Fig. 3c, the effect of the reverse perspective of the two-dimensional code is shown.
- a prompt message is sent to the predetermined monitoring device, indicating that a circuit fault occurs on the display screen of the service device to be detected needs to be processed.
- the prompt information is sent to the background server, where the prompt information includes the unique ID of the service device in which the circuit is faulty, and the network identifier of the unmanned network area where the service device to be detected is located.
- the prompt information format may be "The unattended dot area of the dot is marked as *****, and the display screen of the service device whose ID number is ***** has a circuit failure that needs to be processed.”
- Step S30 If the display screen of the service device to be detected does not have a circuit failure, control the to-be-detected industry.
- the device displays the image according to the preset display parameter, and analyzes the image displayed on the display screen to analyze whether a preset type of abnormality has occurred on the display screen of the service device to be detected.
- step S30 in the preferred embodiment of the screen state automatic detecting method of the present application, and step S30 includes:
- Step S301 the display screen of the service device to be detected is controlled to display a plurality of preset colors in a solid color display manner, for example, three preset colors of red, green, and blue, wherein different preset colors correspond to different size displays.
- the area, the shape of the display area corresponding to each preset color corresponds to the shape of the maximum display area of the display screen, and the display area corresponding to one of the preset colors is the maximum display area of the display screen.
- step S302 when a preset color is displayed in a solid color display manner, a mask is displayed on the display graphic of the display screen for the preset color, that is, the other colors of the display screen except the preset color are masked to obtain a mask.
- Code image when a preset color is displayed in a solid color display manner, a mask is displayed on the display graphic of the display screen for the preset color, that is, the other colors of the display screen except the preset color are masked to obtain a mask.
- Step S303 performing anti-interference processing on the mask image corresponding to the preset color of the display area of the smaller size to obtain an anti-interference image, wherein the display area of the smaller size is other sizes than the maximum display area. Display area.
- Step S304 determining a maximum display area of the display screen according to the acquired anti-interference image, and performing image extraction of the actual display area for the mask image of the preset color corresponding to the maximum display area according to the determined maximum display area.
- extract the maximum display image because if there is interference area in the perimeter of the screen, for example, there is “water inlet” around the screen, the screen boundary area cannot be displayed normally, then the maximum map obtained by taking the mask is incomplete, quite In order to exclude the boundary interference problem from the detection, it is necessary to obtain the maximum display area by the above method.
- the screen width of the small-sized rectangle is L1
- the screen width of the medium-size rectangle is L2
- the screen width of the large-size rectangle is L3.
- the proportional relationship between the length and width of the screen is also known data.
- the maximum display area of the display screen is determined according to the anti-interference image of the preset color corresponding to the display area of the smaller size, and the preset color corresponding to the maximum display area is determined according to the determined maximum display area.
- the mask image is used to extract the image of the actual display area, and the principle of extracting the maximum display image is as follows:
- the width of the large-sized rectangle is denoted as L" 3 .
- the average value of the two values is preferably taken as the width of the large-size rectangle according to Formula 1:
- the area range of the large-sized rectangle can be obtained, and the area range of the large-size rectangle is the maximum display area of the screen, that is, the outer contour curve of the largest-size rectangular area shown in FIG. 6.
- the quadrilateral is inversely transformed into a rectangle, and the anti-interference image is acquired according to the position of the mask image rectangle.
- the boundary of the calculated small-sized rectangle and the boundary of the medium-sized rectangle are respectively calculated.
- an extension line of the boundary line can be obtained, and the extension lines of the four boundary lines are respectively obtained by the above method, and then the intersection points of the four boundary lines are determined, according to the intersection of the four boundary lines and the The four boundary lines determine the position of the quadrilateral of the mask image, that is, the range of the medium size rectangle or the small size rectangle, thereby determining the exact values of L1 and L2.
- Step S305 performing noise filtering enhancement preprocessing on the maximum display image to obtain an enhanced image.
- the collected abnormal image of the screen is affected by non-uniform illumination, irregular gaps on the screen surface, equipment, etc., so that the acquired image is noisy, it is necessary to remove the noise while maintaining the details of the abnormal state of the screen, and As much as possible to improve the contrast between the crack and the background, which is conducive to the later segmentation and recognition. It can be seen from the pre-processing of FIG. 9 and the pre-processed FIG. 10 that the noise-improved graph has less noise interference and the problem is more clearly visible.
- the present application preprocesses an image using a method of a pilot filter.
- the algorithm steps using the bootstrap filter are as follows:
- the grayscale image of the screen is used as the input image P.
- the grayscale screen image is used as the guide image I
- the image filtered by the noise filter is q
- the input image P is enhanced by Equation 2.
- m is the enhancement factor, which can be determined according to the actual situation.
- the filter here can be selected according to the actual situation.
- the ⁇ of the following formula is a satisfaction weight, and the enchanted satisfaction is evaluated according to the P_enhanced image effect, and the highest satisfaction of the enhanced is set to a fixed value, for example, 500.
- T n The nth iteration is denoted as T n
- T n+1 the n+1th time can be recorded as T n+1
- T n+1 the value of m after n iterations is Then there are:
- the iterative process is as follows: First we give an initial weight m as Then, substituting into formula 2 to solve the image after P_enhanced, and then evaluating the satisfaction of P_enhanced according to formula 3, and then digitizing the obtained satisfaction weight ⁇ , and then using the alphaned value and formula 5 to m weight The value is updated. Then repeat the above steps to update m. After several rounds of iteration, you can get an ideal m weight. Because each time the iteration will modify the weight m, the ⁇ value will decrease when the effect of the image P_enhanced is poor, and the final m weight will also decrease, and the weight will increase accordingly.
- Each adjustment will drive the image P_enhanced effect to a better direction, set the number of iterations to a fixed value (here set to 10), and it will have a convergence state after n iterations, that is, the image P_enhanced effect is close to the most Good state.
- the enhanced image P_enhanced is used as the input image P, and the grayscale screen image is used as the guide image I, and the filtered image is obtained as the final enhanced and smoothed image.
- step S306 the enhanced image is analyzed to analyze whether a preset type of abnormality has occurred in the display screen of the service device to be detected, and the abnormality of the preset type includes bad spots, bad lines and cracks.
- the identification process of the bad line is as follows: the image preprocessed by the above steps is acquired, and the straight line is detected by using the Hough transform. When the number of bad lines is detected to be greater than 1, it is determined that there is a bad line; otherwise, the judgment does not exist. Bad line.
- the Hough transform is a feature extraction technique.
- the classic Hough transform recognizes the lines in the image and later develops to recognize arbitrary shapes, but the more common identifying shapes are circles and ellipses, which we use to identify the state of the screen loop.
- a straight line is defined as:
- ⁇ 0 is used to represent the algebraic distance of the origin to the line, and ⁇ 0 is the angle between the orthogonal line of the line and the x-axis, then:
- ( ⁇ , ⁇ ) is a representation of polar coordinates. However, if ( ⁇ , ⁇ ) is also expressed in the form of Cartesian coordinates, that is, ⁇ and ⁇ are orthogonally processed, then ( ⁇ , ⁇ ) is called a Hough space.
- a point in the Cartesian coordinate system corresponds to a sinusoid of the Hough space.
- a straight line is composed of an infinite number of points. In the Hough space, there are an infinite number of sinusoids, but these sinusoids intersect at a point ( ⁇ 0 , ⁇ 0 ), and the point is brought into Equation 7 and Equation 8 to obtain the slope of the line. And the intercept, so a straight line is determined. Therefore, when the line is identified by the Hough transform, the maximum value in the Hough space may correspond to a straight line.
- the conventional method of checking the bad line is by manual observation.
- the robot automatically detects whether there is a bad line in the preprocessed image by the above method.
- the process of determining the bad spot is to obtain the pre-processed picture, and use the SimpleBlobDetector algorithm to detect the presence of spots, when the number of detected spots is greater than one. , to determine the presence of bad spots on the screen, otherwise it is judged that there is no bad spot on the screen.
- the embodiment also introduces a A method for selecting parameters in the SimpleBlobDetector algorithm.
- the meaning of the above formula is that when the feature value is k, the probability that the judgment is that the spot is correct is as high as possible, and the lower the probability that the judgment is the spot but not actually the spot, the larger the weight wk.
- the parameter that needs to be determined is selected as the feature point of the SimpleBlobDetector algorithm with the largest wk value of each feature parameter. Therefore, it is not necessary to artificially determine the specific value of this parameter.
- the conventional method of inspecting bad spots is by manual observation.
- the robot automatically detects whether there is a bad spot on the preprocessed image by the above method.
- the screen is tested for cracks as follows:
- the crack is identified by the segmented image: after the segmentation, some connected regions can be seen, where the connected region refers to the image region formed after the crack is segmented, and when the number of connected regions is greater than 1, the image can be judged. There are cracks, and the cracks may be mesh or a single crack across the screen. If there is only one connected region, we obtain the X-axis projection distance H of the crack and the projection distance R from the y-axis, as shown in Fig. 14. If H ⁇ threshold D and R ⁇ threshold D, there is no crack in the screen.
- the threshold D is a distance value, and when H and R are simultaneously smaller than this value, the line can be considered as a crack, but a noise point. If any of H and R is greater than this threshold, it can be judged that there is a lateral or longitudinal crack on the screen, but the crack does not span the entire screen, as shown in Fig. 14 is a form of lateral crack.
- the conventional method for inspecting cracks is by manual observation.
- the robot automatically detects whether there is a crack in the preprocessed image by the above method.
- the control robot moves to a preset area of each service device in the unmanned network point area, and if the robot moves to a preset area of a special service device, the service device to be detected is Sending a preset graphic code to determine whether a circuit fault occurs in the service device to be detected according to the content displayed on the display screen by the preset graphic code. If no circuit failure occurs, controlling the service device to be detected to perform graphic display according to the preset display parameter. And analyzing the image displayed on the display screen according to the preset analysis rule to analyze whether the display screen of the service device to be detected has a preset type of abnormality, the scheme does not need manual participation, and the robot moves to the corresponding area to the device. The circuit fault and the screen display status are automatically detected.
- the embodiment of the present application further provides a computer readable storage medium, where the screen state automatic detection program is stored, and the screen state automatic detection program can be executed by one or more processors, Implement the following operations:
- the service device corresponding to the preset area is used as a service device to be detected, and a preset graphic code is sent to the service device to be detected, according to the preset graphic code. Determining, by the content displayed on the display screen, whether the circuit to be detected has a circuit failure;
- the device to be detected is controlled to perform image display according to the preset display parameter, and the image displayed on the display screen is analyzed to analyze the service to be detected. Whether the preset type of exception has occurred on the display screen of the device.
- the robot performs scan analysis on the content displayed on the display screen by the service device to be detected based on the display instruction; if the preset information is scanned and analyzed from the display screen of the service device to be detected, the display of the service device to be detected is determined. No circuit failure occurs on the screen; if the preset information is not scanned and analyzed from the display screen of the service device to be detected, it is determined that a circuit fault has occurred on the display screen of the service device to be detected.
- the display screens of the service device to be detected are respectively displayed in a plurality of preset colors in a manner of displaying in a solid color, wherein different preset colors correspond to display areas of different sizes, shapes of the display areas corresponding to the respective preset colors, and the display
- the shape of the maximum display area of the screen corresponds to, and the display area corresponding to one of the preset colors is the maximum display area of the display screen; when a preset color is displayed in a solid color display manner, for the preset color Masking the display graphic of the display screen to obtain a mask image; performing anti-interference processing on the mask image corresponding to the preset color of the display area of the smaller size to obtain an anti-interference image, wherein the smaller size
- the display area is a display area of a size other than the maximum display area; the maximum display area of the display screen is determined according to the acquired anti-interference image, and the preset corresponding to the maximum display area is determined according to the determined maximum display area.
- the mask image of the color is used to extract the image of the actual display area, and the maximum display image is extracted; Maximum enhancement display image noise filtering pretreatment, to obtain an enhanced image; enhancement of the image is analyzed to analyze whether the display screen of the device to be detected traffic abnormality occurs in the preset type.
- the technical solution of the present application which is essential or contributes to the prior art, may be embodied in the form of a software product stored in a storage medium (such as ROM/RAM as described above). , a disk, an optical disk, including a number of instructions for causing a terminal device (which may be a mobile phone, a computer, a server, or a network device, etc.) to perform the methods described in the various embodiments of the present application.
- a terminal device which may be a mobile phone, a computer, a server, or a network device, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Aviation & Aerospace Engineering (AREA)
- Automation & Control Theory (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Electromagnetism (AREA)
- Quality & Reliability (AREA)
- Artificial Intelligence (AREA)
- Toxicology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Image Processing (AREA)
- Manipulator (AREA)
- Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)
- Financial Or Insurance-Related Operations Such As Payment And Settlement (AREA)
- Image Analysis (AREA)
Abstract
Description
图3a为发生了透视的二维码;图3b为透视后的二维码的外轮廓模型;图3c为将透视后的二维码进行反透视后的效果;
Claims (20)
- 一种屏幕状态自动检测机器人,其特征在于,所述机器人包括存储器和处理器,所述存储器上存储有屏幕状态自动检测程序,所述屏幕状态自动检测程序被所述处理器执行时实现如下操作:控制所述机器人分别移动到无人网点区的各个业务设备的预设区域内;若所述机器人移动到一个预设区域内,则将该预设区域对应的业务设备作为待检测业务设备,并向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障;若所述待检测业务设备的显示屏幕未发生电路故障,则控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求1所述的屏幕状态自动检测机器人,其特征在于,所述处理器还用于执行所述屏幕状态自动检测程序,以在向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤之后,还实现如下步骤:若所述待检测业务设备的显示屏幕发生了电路故障,则向预先确定的监控设备发送提示信息,提示所述待检测业务设备的显示屏幕发生了电路故障需要进行处理。
- 根据权利要求1所述的屏幕状态自动检测机器人,其特征在于,所述控制机器人分别移动到无人网点区的各个业务设备的预设区域内包括:按照预设移动导航路径控制所述机器人移动,在移动到一个业务设备的预设区域内且对所述待检测业务设备的显示屏幕检测完毕后,控制所述机器人按照所述移动导航路径继续移动,直到所有业务设备的显示屏幕均检测完毕;或者,控制所述机器人进行随机移动,在移动到一个业务设备的预设区域内且对所述待检测业务设备的显示屏幕检测完毕后,将所述待检测业务设备标记为障碍物,控制机器人进行规避移动,在规避移动完成后,控制所述机器人继续进行随机移动,直至所有业务设备均被标记为障碍物后结束移动,并将业务设备的障碍物标记清除。
- 根据权利要求1所述的屏幕状态自动检测机器人,其特征在于,所述向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤包括:控制所述机器人与待检测的业务设备进行无线通信,以向所述待检测业务设备发送显示预设图形码的显示指令,其中,所述预设图像码中包含有预设信息;控制所述机器人对所述待检测业务设备基于所述显示指令在显示屏幕上显示的内容进行扫描分析;若从待检测的业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕没有发生电路故障;若没有从待检测业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕发生了电路故障。
- 根据权利要求1所述的屏幕状态自动检测机器人,其特征在于,所述控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像按照预设分析规则进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常包括:控制所述待检测业务设备的显示屏幕分别以纯色显示的方式显示多种预设颜色,其中,不同预设颜色对应不同尺寸的显示区域,各个预设颜色对应的显示区域的形状和所述显示屏幕的最大显示区域的形状对应,且其中一个所述预设颜色对应的显示区域为所述显示屏幕的最大显示区域;在以纯色显示的方式显示一种预设颜色时,针对该预设颜色对显示屏幕的显示图形取 掩码,以获取掩码图像;为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像,其中,所述较小尺寸的显示区域为除最大显示区域之外的其它尺寸的显示区域;根据获取的抗干扰图像,确定出所述显示屏幕的最大显示区域,并根据确定的最大显示区域,为对应最大显示区域的预设颜色的取掩码图像进行实际显示区域的图像提取,提取出最大显示图像;对所述最大显示图像进行滤噪增强预处理,获得增强图像;对所述增强图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求2所述的屏幕状态自动检测机器人,其特征在于,所述控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像按照预设分析规则进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常的步骤包括:控制所述待检测业务设备的显示屏幕分别以纯色显示的方式显示多种预设颜色,其中,不同预设颜色对应不同尺寸的显示区域,各个预设颜色对应的显示区域的形状和所述显示屏幕的最大显示区域的形状对应,且其中一个所述预设颜色对应的显示区域为所述显示屏幕的最大显示区域;在以纯色显示的方式显示一种预设颜色时,针对该预设颜色对显示屏幕的显示图形取掩码,以获取掩码图像;为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像,其中,所述较小尺寸的显示区域为除最大显示区域之外的其它尺寸的显示区域;根据获取的抗干扰图像,确定出所述显示屏幕的最大显示区域,并根据确定的最大显示区域,为对应最大显示区域的预设颜色的取掩码图像进行实际显示区域的图像提取,提取出最大显示图像;对所述最大显示图像进行滤噪增强预处理,获得增强图像;对所述增强图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求6所述的屏幕状态自动检测机器人,其特征在于,所述为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像的步骤包括:获取对应较小尺寸的显示区域的预设颜色的掩码图像的每条边界线上的多个点的坐标;分别基于所述各个边界线上的多个点的坐标,通过最小二乘法直线拟合算法获取所述掩码图像的每条边界线的延长线;根据所述掩码图像的每条边界线的延长线获取所述四条边界线的交点,根据所述四条边界线的交点以及所述四条边界线确定出所述掩码图像四边形的位置;将所述四边形进行反透视变换为矩形,根据所述掩码图像矩形的位置获取所述抗干扰图像。
- 一种屏幕状态自动检测方法,其特征在于,所述方法包括:控制所述机器人分别移动到无人网点区的各个业务设备的预设区域内;若所述机器人移动到一个预设区域内,则将该预设区域对应的业务设备作为待检测业务设备,并向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障;若所述待检测业务设备的显示屏幕未发生电路故障,则控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求8所述的屏幕状态自动检测方法,其特征在于,向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤之后,所述方法还包括如下步骤:若所述待检测业务设备的显示屏幕发生了电路故障,则向预先确定的监控设备发送提示信息,提示所述待检测业务设备的显示屏幕发生了电路故障需要进行处理。
- 根据权利要求8所述的屏幕状态自动检测方法,其特征在于,所述控制机器人分别移动到无人网点区的各个业务设备的预设区域内的步骤包括:按照预设移动导航路径控制所述机器人移动,在移动到一个业务设备的预设区域内且对所述待检测业务设备的显示屏幕检测完毕后,控制所述机器人按照所述移动导航路径继续移动,直到所有业务设备的显示屏幕均检测完毕;或者,控制所述机器人进行随机移动,在移动到一个业务设备的预设区域内且对所述待检测业务设备的显示屏幕检测完毕后,将所述待检测业务设备标记为障碍物,控制机器人进行规避移动,在规避移动完成后,控制所述机器人继续进行随机移动,直至所有业务设备均被标记为障碍物后结束移动,并将业务设备的障碍物标记清除。
- 根据权利要求8所述的屏幕状态自动检测方法,其特征在于,所述向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤包括:控制所述机器人与待检测的业务设备进行无线通信,以向所述待检测业务设备发送显示预设图形码的显示指令,其中,所述预设图像码中包含有预设信息;控制所述机器人对所述待检测业务设备基于所述显示指令在显示屏幕上显示的内容进行扫描分析;若从待检测的业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕没有发生电路故障;若没有从待检测业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕发生了电路故障。
- 根据权利要求8所述的屏幕状态自动检测方法,其特征在于,所述控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像按照预设分析规则进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常的步骤包括:控制所述待检测业务设备的显示屏幕分别以纯色显示的方式显示多种预设颜色,其中,不同预设颜色对应不同尺寸的显示区域,各个预设颜色对应的显示区域的形状和所述显示屏幕的最大显示区域的形状对应,且其中一个所述预设颜色对应的显示区域为所述显示屏幕的最大显示区域;在以纯色显示的方式显示一种预设颜色时,针对该预设颜色对显示屏幕的显示图形取掩码,以获取掩码图像;为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像,其中,所述较小尺寸的显示区域为除最大显示区域之外的其它尺寸的显示区域;根据获取的抗干扰图像,确定出所述显示屏幕的最大显示区域,并根据确定的最大显示区域,为对应最大显示区域的预设颜色的取掩码图像进行实际显示区域的图像提取,提取出最大显示图像;对所述最大显示图像进行滤噪增强预处理,获得增强图像;对所述增强图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求9所述的屏幕状态自动检测方法,其特征在于,所述控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像按照预设分析规则 进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常的步骤包括:控制所述待检测业务设备的显示屏幕分别以纯色显示的方式显示多种预设颜色,其中,不同预设颜色对应不同尺寸的显示区域,各个预设颜色对应的显示区域的形状和所述显示屏幕的最大显示区域的形状对应,且其中一个所述预设颜色对应的显示区域为所述显示屏幕的最大显示区域;在以纯色显示的方式显示一种预设颜色时,针对该预设颜色对显示屏幕的显示图形取掩码,以获取掩码图像;为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像,其中,所述较小尺寸的显示区域为除最大显示区域之外的其它尺寸的显示区域;根据获取的抗干扰图像,确定出所述显示屏幕的最大显示区域,并根据确定的最大显示区域,为对应最大显示区域的预设颜色的取掩码图像进行实际显示区域的图像提取,提取出最大显示图像;对所述最大显示图像进行滤噪增强预处理,获得增强图像;对所述增强图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求13所述的屏幕状态自动检测方法,其特征在于,所述为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像的步骤包括:获取对应较小尺寸的显示区域的预设颜色的掩码图像的每条边界线上的多个点的坐标;分别基于所述各个边界线上的多个点的坐标,通过最小二乘法直线拟合算法获取所述掩码图像的每条边界线的延长线;根据所述掩码图像的每条边界线的延长线获取所述四条边界线的交点,根据所述四条边界线的交点以及所述四条边界线确定出所述掩码图像四边形的位置;将所述四边形进行反透视变换为矩形,根据所述掩码图像矩形的位置获取所述抗干扰图像。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有屏幕状态自动检测程序,所述屏幕状态自动检测程序可被一个或者多个处理器执行,以实现如下步骤:控制所述机器人分别移动到无人网点区的各个业务设备的预设区域内;若所述机器人移动到一个预设区域内,则将该预设区域对应的业务设备作为待检测业务设备,并向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障;若所述待检测业务设备的显示屏幕未发生电路故障,则控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求15所述的计算机可读存储介质,其特征在于,所述屏幕状态自动检测程序还可被一个或者多个处理器执行,以在向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤之后,还实现如下步骤:若所述待检测业务设备的显示屏幕发生了电路故障,则向预先确定的监控设备发送提示信息,提示所述待检测业务设备的显示屏幕发生了电路故障需要进行处理。
- 根据权利要求15所述的计算机可读存储介质,其特征在于,所述控制机器人分别移动到无人网点区的各个业务设备的预设区域内的步骤包括:按照预设移动导航路径控制所述机器人移动,在移动到一个业务设备的预设区域内且 对所述待检测业务设备的显示屏幕检测完毕后,控制所述机器人按照所述移动导航路径继续移动,直到所有业务设备的显示屏幕均检测完毕;或者,控制所述机器人进行随机移动,在移动到一个业务设备的预设区域内且对所述待检测业务设备的显示屏幕检测完毕后,将所述待检测业务设备标记为障碍物,控制机器人进行规避移动,在规避移动完成后,控制所述机器人继续进行随机移动,直至所有业务设备均被标记为障碍物后结束移动,并将业务设备的障碍物标记清除。
- 根据权利要求15所述的计算机可读存储介质,其特征在于,所述向所述待检测业务设备发送预设图形码,以根据所述预设图形码在显示屏幕上显示的内容判断所述待检测业务设备是否发生了电路故障的步骤包括:控制所述机器人与待检测的业务设备进行无线通信,以向所述待检测业务设备发送显示预设图形码的显示指令,其中,所述预设图像码中包含有预设信息;控制所述机器人对所述待检测业务设备基于所述显示指令在显示屏幕上显示的内容进行扫描分析;若从待检测的业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕没有发生电路故障;若没有从待检测业务设备的显示屏幕上扫描分析出预设信息,则确定待检测业务设备的显示屏幕发生了电路故障。
- 根据权利要求15所述的计算机可读存储介质,其特征在于,所述控制所述待检测业务设备按照预设显示参数进行图像显示,并对显示屏幕显示的图像按照预设分析规则进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常的步骤包括:控制所述待检测业务设备的显示屏幕分别以纯色显示的方式显示多种预设颜色,其中,不同预设颜色对应不同尺寸的显示区域,各个预设颜色对应的显示区域的形状和所述显示屏幕的最大显示区域的形状对应,且其中一个所述预设颜色对应的显示区域为所述显示屏幕的最大显示区域;在以纯色显示的方式显示一种预设颜色时,针对该预设颜色对显示屏幕的显示图形取掩码,以获取掩码图像;为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像,其中,所述较小尺寸的显示区域为除最大显示区域之外的其它尺寸的显示区域;根据获取的抗干扰图像,确定出所述显示屏幕的最大显示区域,并根据确定的最大显示区域,为对应最大显示区域的预设颜色的取掩码图像进行实际显示区域的图像提取,提取出最大显示图像;对所述最大显示图像进行滤噪增强预处理,获得增强图像;对所述增强图像进行分析,以分析出所述待检测业务设备的显示屏幕是否发生了预设类型的异常。
- 根据权利要求19所述的计算机可读存储介质,其特征在于,所述为对应较小尺寸的显示区域的预设颜色的掩码图像进行抗干扰处理以获取抗干扰图像的步骤包括:获取对应较小尺寸的显示区域的预设颜色的掩码图像的每条边界线上的多个点的坐标;分别基于所述各个边界线上的多个点的坐标,通过最小二乘法直线拟合算法获取所述掩码图像的每条边界线的延长线;根据所述掩码图像的每条边界线的延长线获取所述四条边界线的交点,根据所述四条边界线的交点以及所述四条边界线确定出所述掩码图像四边形的位置;将所述四边形进行反透视变换为矩形,根据所述掩码图像矩形的位置获取所述抗干扰图像。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG11202001417WA SG11202001417WA (en) | 2017-08-29 | 2017-10-31 | Automatic screen state detection robot, method and computer-readable storage medium |
US16/632,859 US11361422B2 (en) | 2017-08-29 | 2017-10-31 | Automatic screen state detection robot, method and computer-readable storage medium |
JP2019556812A JP6911149B2 (ja) | 2017-08-29 | 2017-10-31 | スクリーン状態自動検出ロボット、方法及びコンピュータ読み取り可能な記憶媒体 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710754580.7 | 2017-08-29 | ||
CN201710754580.7A CN107688345B (zh) | 2017-08-29 | 2017-08-29 | 屏幕状态自动检测机器人、方法及计算机可读存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019041516A1 true WO2019041516A1 (zh) | 2019-03-07 |
Family
ID=61155532
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2017/108740 WO2019041516A1 (zh) | 2017-08-29 | 2017-10-31 | 屏幕状态自动检测机器人、方法及计算机可读存储介质 |
Country Status (5)
Country | Link |
---|---|
US (1) | US11361422B2 (zh) |
JP (1) | JP6911149B2 (zh) |
CN (1) | CN107688345B (zh) |
SG (1) | SG11202001417WA (zh) |
WO (1) | WO2019041516A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111339298A (zh) * | 2020-02-25 | 2020-06-26 | 北京松果电子有限公司 | 一种分类预测方法、装置及存储介质 |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017156046A1 (en) | 2016-03-07 | 2017-09-14 | Hyla, Inc. | Screen damage detection for devices |
US11210777B2 (en) * | 2016-04-28 | 2021-12-28 | Blancco Technology Group IP Oy | System and method for detection of mobile device fault conditions |
US10876914B2 (en) | 2016-04-28 | 2020-12-29 | Blancco Technology Group IP Oy | Systems and methods for detection of mobile device fault conditions |
US12013346B2 (en) | 2018-02-19 | 2024-06-18 | Blancco Technology Group IP Oy | System and method for detection of mobile device fault conditions |
CN110340882A (zh) * | 2018-04-08 | 2019-10-18 | 苏州宝时得电动工具有限公司 | 机器人避障方法、装置、计算机设备及存储介质 |
CN108921843B (zh) * | 2018-07-05 | 2022-08-12 | 深圳市路畅科技股份有限公司 | 一种检测车载导航仪性能的系统及方法 |
WO2021055457A1 (en) | 2019-09-16 | 2021-03-25 | Assurant, Inc. | System, method, apparatus, and computer program product for utilizing machine learning to process an image of a mobile device to determine a mobile device integrity status |
US11580627B2 (en) | 2020-01-06 | 2023-02-14 | Assurant, Inc. | Systems and methods for automatically grading pre-owned electronic devices |
CN112069043A (zh) * | 2020-08-04 | 2020-12-11 | 北京捷通华声科技股份有限公司 | 一种终端设备状态检测方法、模型生成方法以及装置 |
CN113221881B (zh) * | 2021-04-30 | 2022-03-22 | 浙大城市学院 | 一种多层级的智能手机屏幕缺陷检测方法 |
CN113709289A (zh) * | 2021-08-31 | 2021-11-26 | 展讯通信(上海)有限公司 | 终端测试的监控系统及方法 |
CN114973612A (zh) * | 2022-03-28 | 2022-08-30 | 深圳市揽讯科技有限公司 | Led显示屏故障自动告警监控系统及监控方法 |
CN115375645B (zh) * | 2022-08-17 | 2024-08-16 | 武汉海微科技股份有限公司 | 车载显示屏显示异常的测试方法、装置、设备及存储介质 |
CN117078672B (zh) * | 2023-10-13 | 2024-07-09 | 南京极景微半导体有限公司 | 基于计算机视觉的手机屏幕缺陷智能检测方法 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1765111A (zh) * | 2004-01-23 | 2006-04-26 | 日本诺贝尔株式会社 | 设备检查装置、基于其上的设备检查系统以及蜂窝电话保持器 |
CN1871846A (zh) * | 2003-10-21 | 2006-11-29 | 皇家飞利浦电子股份有限公司 | 屏幕图像验证 |
US20100053604A1 (en) * | 2008-08-31 | 2010-03-04 | Paul Rice | Broken screen detector |
CN101782456A (zh) * | 2009-01-19 | 2010-07-21 | 深圳市新超亮特种显示设备有限公司 | 一种远程检测显示器工作状态的方法及系统 |
CN103838674A (zh) * | 2014-03-21 | 2014-06-04 | 徐禄勇 | 基于数字图像的智能测试机器人及其使用方法 |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5600574A (en) * | 1994-05-13 | 1997-02-04 | Minnesota Mining And Manufacturing Company | Automated image quality control |
JP2001142512A (ja) * | 1999-11-16 | 2001-05-25 | Mitsubishi Electric Corp | ロボットの遠隔操作システム |
DE10137056C5 (de) * | 2001-07-28 | 2018-09-06 | Rieter Ingolstadt Gmbh | Verfahren zur Wartung einer Textilmaschine |
KR101166843B1 (ko) * | 2005-08-30 | 2012-07-19 | 엘지디스플레이 주식회사 | Lcd 검사 장비 및 lcd 검사 방법 |
JP4703480B2 (ja) * | 2006-05-17 | 2011-06-15 | 株式会社日立製作所 | 映像内の移動体検出方法、映像システムの異常発生原因分析支援方法及び支援システム |
KR101546850B1 (ko) * | 2008-06-27 | 2015-08-24 | 삼성전자주식회사 | 디스플레이장치의 화질평가방법 및 화질평가장치 |
US8643662B2 (en) * | 2009-04-22 | 2014-02-04 | Samsung Electronics Co., Ltd. | Video entertainment picture quality adjustment |
FR3012425B1 (fr) * | 2013-10-24 | 2017-03-24 | European Aeronautic Defence & Space Co Eads France | Robot collaboratif d'inspection visuelle d'un aeronef |
CN105336287B (zh) * | 2014-08-06 | 2019-08-06 | 珠海格力电器股份有限公司 | 显示屏的测试方法、装置和系统 |
CN104318879A (zh) * | 2014-10-20 | 2015-01-28 | 京东方科技集团股份有限公司 | 显示装置、显示装置故障分析系统和方法 |
US9955150B2 (en) * | 2015-09-24 | 2018-04-24 | Qualcomm Incorporated | Testing of display subsystems |
CN106406289A (zh) * | 2016-11-17 | 2017-02-15 | 北京中科汇联科技股份有限公司 | 一种机器人故障检修系统及方法 |
CN106681333A (zh) * | 2017-03-02 | 2017-05-17 | 刘伟豪 | 一种提高变电站巡检机器人稳定性的方法及系统 |
CN106960646A (zh) * | 2017-03-30 | 2017-07-18 | 上海与德科技有限公司 | 一种屏幕故障检测方法和装置 |
-
2017
- 2017-08-29 CN CN201710754580.7A patent/CN107688345B/zh active Active
- 2017-10-31 WO PCT/CN2017/108740 patent/WO2019041516A1/zh active Application Filing
- 2017-10-31 SG SG11202001417WA patent/SG11202001417WA/en unknown
- 2017-10-31 US US16/632,859 patent/US11361422B2/en active Active
- 2017-10-31 JP JP2019556812A patent/JP6911149B2/ja active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1871846A (zh) * | 2003-10-21 | 2006-11-29 | 皇家飞利浦电子股份有限公司 | 屏幕图像验证 |
CN1765111A (zh) * | 2004-01-23 | 2006-04-26 | 日本诺贝尔株式会社 | 设备检查装置、基于其上的设备检查系统以及蜂窝电话保持器 |
US20100053604A1 (en) * | 2008-08-31 | 2010-03-04 | Paul Rice | Broken screen detector |
CN101782456A (zh) * | 2009-01-19 | 2010-07-21 | 深圳市新超亮特种显示设备有限公司 | 一种远程检测显示器工作状态的方法及系统 |
CN103838674A (zh) * | 2014-03-21 | 2014-06-04 | 徐禄勇 | 基于数字图像的智能测试机器人及其使用方法 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111339298A (zh) * | 2020-02-25 | 2020-06-26 | 北京松果电子有限公司 | 一种分类预测方法、装置及存储介质 |
CN111339298B (zh) * | 2020-02-25 | 2024-04-09 | 北京小米松果电子有限公司 | 一种分类预测方法、装置及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
SG11202001417WA (en) | 2020-03-30 |
CN107688345A (zh) | 2018-02-13 |
US11361422B2 (en) | 2022-06-14 |
US20200167907A1 (en) | 2020-05-28 |
JP6911149B2 (ja) | 2021-07-28 |
CN107688345B (zh) | 2019-04-19 |
JP2020517926A (ja) | 2020-06-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019041516A1 (zh) | 屏幕状态自动检测机器人、方法及计算机可读存储介质 | |
CN110097542B (zh) | 芯片气泡的检测方法、装置及存储介质 | |
JP5699788B2 (ja) | スクリーン領域検知方法及びシステム | |
CN102750555B (zh) | 一种基于即时通讯工具的表情识别装置 | |
CN110199317A (zh) | 使用手持设备对木板的自动检测、计数和测量 | |
US9946947B2 (en) | System and method for finding saddle point-like structures in an image and determining information from the same | |
JP6197340B2 (ja) | 画像処理装置、画像処理方法、およびプログラム | |
US11216905B2 (en) | Automatic detection, counting, and measurement of lumber boards using a handheld device | |
CN116883661B (zh) | 一种基于目标识别与图像处理的动火作业检测方法 | |
CN101146236A (zh) | 用以测试影像撷取装置的方法、相关的测试平台及影像卡 | |
CN111259891A (zh) | 一种自然场景下身份证识别方法、装置、设备和介质 | |
CN108154506A (zh) | 一种六角螺母自动检测方法 | |
CN108665495A (zh) | 图像处理方法及装置、移动终端 | |
EP4171051A1 (en) | Collation device, program, and collation method | |
US11825190B2 (en) | Collation device | |
CN113139399B (zh) | 一种图像线框识别方法及服务器 | |
CN110458857A (zh) | 中心对称图元检测方法、装置、电子设备及可读存储介质 | |
KR101637977B1 (ko) | 레이저 비전 시스템을 이용한 용접 조인트의 특징점 검출 방법 | |
JP2023061880A (ja) | 照合装置及びプログラム | |
CN115424062B (zh) | 一种自动识别斜线网的方法及装置、设备及存储介质 | |
JP7613093B2 (ja) | 情報処理装置及び情報処理プログラム | |
CN110769221B (zh) | 在拍摄图像中检测投影图像的角点的方法和投影仪 | |
CN118368408B (zh) | 图像采集设备的检测方法、设备和存储介质 | |
CN116879292B (zh) | 一种光触媒硅藻泥板的质量评估方法及装置 | |
US20240153126A1 (en) | Automatic image cropping using a reference feature |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 17923772 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019556812 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 24/09/2020) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 17923772 Country of ref document: EP Kind code of ref document: A1 |