[go: up one dir, main page]

CN113602799A - A kind of airport luggage handling system and control method thereof - Google Patents

A kind of airport luggage handling system and control method thereof Download PDF

Info

Publication number
CN113602799A
CN113602799A CN202110897488.2A CN202110897488A CN113602799A CN 113602799 A CN113602799 A CN 113602799A CN 202110897488 A CN202110897488 A CN 202110897488A CN 113602799 A CN113602799 A CN 113602799A
Authority
CN
China
Prior art keywords
straight line
grid
luggage
industrial camera
gripper
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110897488.2A
Other languages
Chinese (zh)
Other versions
CN113602799B (en
Inventor
祝会龙
张静
刘满禄
单毛毛
田凤莲
王姮
张华�
段淇昱
周建
白克强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southwest University of Science and Technology
Original Assignee
Southwest University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southwest University of Science and Technology filed Critical Southwest University of Science and Technology
Priority to CN202110897488.2A priority Critical patent/CN113602799B/en
Publication of CN113602799A publication Critical patent/CN113602799A/en
Application granted granted Critical
Publication of CN113602799B publication Critical patent/CN113602799B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B65CONVEYING; PACKING; STORING; HANDLING THIN OR FILAMENTARY MATERIAL
    • B65GTRANSPORT OR STORAGE DEVICES, e.g. CONVEYORS FOR LOADING OR TIPPING, SHOP CONVEYOR SYSTEMS OR PNEUMATIC TUBE CONVEYORS
    • B65G47/00Article or material-handling devices associated with conveyors; Methods employing such devices
    • B65G47/74Feeding, transfer, or discharging devices of particular kinds or types
    • B65G47/90Devices for picking-up and depositing articles or materials
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B65CONVEYING; PACKING; STORING; HANDLING THIN OR FILAMENTARY MATERIAL
    • B65GTRANSPORT OR STORAGE DEVICES, e.g. CONVEYORS FOR LOADING OR TIPPING, SHOP CONVEYOR SYSTEMS OR PNEUMATIC TUBE CONVEYORS
    • B65G47/00Article or material-handling devices associated with conveyors; Methods employing such devices
    • B65G47/52Devices for transferring articles or materials between conveyors i.e. discharging or feeding devices
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/80Geometric correction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Mechanical Engineering (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Manipulator (AREA)

Abstract

本发明公开了一种航空港行李箱搬运系统及其控制方法,包括控制单元、与控制单元信号连接的检测单元和用于运输行李箱的运输单元;控制单元包括控制箱和固定安装于机械臂上、且受控于控制单元的抓手;检测单元包括网格激光、工业相机和压力薄膜传感器;工业相机和网格激光均安装于抓手的顶板上,且工业相机的四周环设有若干个点光源灯;压力薄膜传感器固定安装于抓手的内侧侧壁上;运输单元包括传送带和设置于传送带一侧的运输小车。本发明利用机械臂末端的抓手与周围点光源灯配合工业相机结合机器视觉中的物体检测,对行李箱进行尺寸及体积的估计,机械臂控制抓手去抓取行李箱放至行李箱搬运小车上。

Figure 202110897488

The invention discloses an airport luggage handling system and a control method thereof, comprising a control unit, a detection unit signally connected with the control unit, and a transport unit for transporting luggage; the control unit includes a control box and is fixedly installed on a mechanical arm , and controlled by the gripper of the control unit; the detection unit includes a grid laser, an industrial camera and a pressure film sensor; the industrial camera and the grid laser are installed on the top plate of the gripper, and the surrounding rings of the industrial camera are provided with several The point light source lamp; the pressure film sensor is fixedly installed on the inner side wall of the gripper; the transport unit includes a conveyor belt and a transport trolley arranged on one side of the conveyor belt. The invention uses the gripper at the end of the mechanical arm and the surrounding point light source lamps to cooperate with the industrial camera and the object detection in machine vision to estimate the size and volume of the luggage, and the robotic arm controls the gripper to grab the luggage and put it into the luggage for transportation. on the car.

Figure 202110897488

Description

Airport luggage case carrying system and control method thereof
Technical Field
The invention belongs to the technical field of robot handling, and particularly relates to an airport luggage case handling system and a control method thereof.
Background
In the face of increasingly developed air transportation business, the mobility requirements of personnel, luggage and goods are continuously improved, and the workload of airport operation is greatly increased. The luggage consignment is carried in and out port, the sorting task is increasingly heavy, and the manual sorting mode is adopted, so that the labor intensity of workers is high, the luggage sorting operation efficiency of an airport is seriously hindered, and the operation cost is rapidly increased.
Therefore, the automatic baggage handling and sorting system of the airport is required to be intelligently transformed, so that the automation and the intellectualization of the baggage handling of the airport are realized, the bag clamping, the mistaken sorting and the damage in the baggage handling and sorting process are avoided, the baggage and cargo handling efficiency of the airport is improved, and the intelligent construction level and the image of the airport are improved.
Present luggage transport mostly leans on the mode of manual sorting, and its work efficiency is very low, because the size and size of suitcase is unified not the sorting that can not be fine when carrying out the suitcase transport, seriously hinders airport luggage letter sorting operation efficiency more, leads to operation cost to rise sharply etc..
Disclosure of Invention
The present invention is directed to solving the above-mentioned problems, and provides a baggage handling system for airport and a control method thereof.
In order to achieve the purpose, the invention adopts the technical scheme that:
in one aspect, a luggage handling system for an airport and a control method thereof, which includes a control unit, a detection unit in signal connection with the control unit, and a transportation unit for transporting luggage;
the control unit comprises a control box and a gripper which is fixedly arranged on the mechanical arm and is controlled by the control unit;
the detection unit comprises a grid laser, an industrial camera and a pressure film sensor; the industrial camera and the grid laser are both arranged on the top plate of the gripper, and a plurality of point light source lamps are arranged around the industrial camera; the pressure film sensor is fixedly arranged on the side wall of the inner side of the hand grip;
the transport unit comprises a conveyor belt and a transport trolley arranged on one side of the conveyor belt.
Furthermore, the grid laser, the industrial camera and the film pressure sensor are connected with a small embedded industrial personal computer and are in communication connection with the control box through the small embedded industrial personal computer.
Further, the point light source lamp is an LED light-emitting source lamp.
In one aspect, a method of controlling an airport baggage handling system comprising the steps of:
s1, industrial camera calibration, including parameter calibration in the industrial camera and calibration of hand-eye coordinate conversion relation;
s2, the industrial camera collects the image information of the luggage case on the conveyor belt, and transmits the collected image information to the control unit for processing to obtain the position information of the luggage case in the image;
s3, controlling the hand grip to move to a target position according to the position information of the luggage case, controlling the hand grip to move according to the moving track of the luggage case, and setting the vertical downward moving distance and the horizontal displacement of the hand grip until the hand grip moves to two sides of the luggage case;
s4, grabbing the luggage case according to the weight of the luggage case and the pressure value fed back by the film sensor, and adjusting the grabbing strength of the grab in real time;
s5, the control unit drives the mechanical arm to rotate, the luggage case is transported to the upper part of the transport trolley, the mechanical arm is moved downwards continuously, the gripper is controlled to release the luggage case, and the luggage case falls on the transport trolley;
and S6, the control unit controls the mechanical arm to return to the initial position.
Further, the step S2, the industrial camera acquires image information of the luggage on the conveyor belt, and transmits the acquired image information to the control unit for processing, so as to obtain the position information of the luggage in the image, including:
marking and extracting the acquired image information data by adopting a silhouette processing method;
carrying out binarization processing on the extracted image information, and carrying out image segmentation on the image subjected to binarization processing to obtain image information containing a target area of the trunk;
filtering the segmented image information;
and adopting skeletonization to identify the position information of the target area, and transmitting the position information to the control unit.
Further, the step S3 of calculating the moving track of the luggage case includes:
the grid laser projects a group of parallel grid red laser lines on the surface of the suitcase to be grabbed;
calculating a horizontal straight line equation and a vertical straight line equation in the grid by adopting standard Hough transform, and then calculating coordinates of start points and stop points of grid lines through intersection points of the horizontal straight lines and the vertical straight lines;
and drawing straight lines on the basis of the binary image of the calculated coordinates of the start point and the stop point, and obtaining a result graph of grid line detection.
Further, according to the result graph of the grid line detection, calculating the coordinates of the starting point of the grid horizontal line:
keeping the vertical straight line equation unchanged, sequentially traversing all the horizontal straight line equations, calculating the coordinates of the intersection point of each horizontal straight line and the set vertical line, and calculating to obtain the coordinates of the initial point of the horizontal straight line of the actual grid;
keeping the horizontal straight line equation unchanged, sequentially traversing all the vertical straight line equations, calculating the coordinates of the intersection points of the horizontal straight line and the vertical straight line, and calculating to obtain the coordinates of the initial points of the straight lines in the vertical direction of the actual grid;
and calculating the coordinates of the end points of the grid horizontal lines according to a result graph of grid line detection:
keeping the horizontal straight line equation unchanged, sequentially traversing all the vertical straight line equations, calculating the intersection point coordinate of each vertical straight line and the set horizontal straight line, and calculating to obtain the end point coordinate of the actual grid horizontal straight line;
keeping the horizontal straight line equation unchanged, sequentially traversing all the vertical straight line equations, and calculating the coordinates of the intersection points of the horizontal straight line and the vertical straight line to obtain the coordinates of the end points of the straight lines in the vertical direction of the actual grid.
Further, the industrial camera calibration in step S1 includes:
carrying out parameter calibration and distortion correction in the industrial camera by adopting a Zhangyou calibration algorithm;
and performing hand-eye calibration by adopting a Tsai-Lenz calibration method.
Further, the calculation and correction of the volume of the suitcase is also included:
when the grid laser irradiates on the target luggage case, the industrial camera photographs the target luggage case, identifies the laser grid on the photographed picture, and takes the edge of the luggage case with the farthest bent grid line of the luggage case as the farthest edge;
calculating the area of the uppermost surface of the farthest luggage case, and further estimating the initial volume of the luggage case;
the calculated volume of the case is filled with an edge dilation filling of a total of 5cm of gray border outwardly filled at the most distal curved gridlines of the case.
The airport luggage case carrying system and the control method thereof provided by the invention have the following beneficial effects:
the size and the volume of the luggage case are estimated by utilizing the tongs at the tail end of the mechanical arm and the peripheral point light source lamps in combination with the object detection in the machine vision, and the mechanical arm controls the tongs to grab the luggage case and place the luggage case on the luggage case carrying trolley; the pressure film sensor is utilized to realize real-time detection and feedback of the strength of the gripper at the tail end of the mechanical arm, so that the luggage case can be gripped without damage; it can adapt to the suitcase that snatchs not unidimensional to the harmless of the suitcase and snatching and carrying that has realized snatching of controllable dynamics has been realized.
Drawings
FIG. 1 is a schematic diagram of a airport luggage handling system.
FIG. 2 is a view showing the construction of the gripper.
Fig. 3 is a block diagram of image processing.
Fig. 4 is a diagram showing an example of a straight line in a polar coordinate system.
FIG. 5 is an illustration of calculating grid line coordinates.
Fig. 6 is a diagram of hand-eye calibration and visual guidance, in which the left diagram is a diagram of hand-eye calibration coordinate transformation, and the right diagram is a diagram of visual guidance process.
Fig. 7 is a flow chart of hand-eye calibration.
FIG. 8 is a three-dimensional information measurement and localization diagram based on structured light.
Fig. 9 is a schematic diagram of luggage volume estimation.
Wherein, 1, a conveyor belt; 2. a top plate; 3. a mechanical arm; 4. a control box; 5. a luggage case; 6. transporting the trolley; 7. an industrial camera; 8. grid laser; 9. a pressure membrane sensor; 10. and (4) a hand grip.
Detailed Description
The following description of the embodiments of the present invention is provided to facilitate the understanding of the present invention by those skilled in the art, but it should be understood that the present invention is not limited to the scope of the embodiments, and it will be apparent to those skilled in the art that various changes may be made without departing from the spirit and scope of the invention as defined and defined in the appended claims, and all matters produced by the invention using the inventive concept are protected.
According to the first embodiment of the application, referring to fig. 1 and 2, the airport luggage 5 handling system and the control method thereof in the present scheme comprise a control unit, a detection unit in signal connection with the control unit, and a transportation unit for transporting the luggage 5.
The control unit comprises a control box 4, a mechanical arm 3 and a gripper 10, the gripper 10 is mounted on the mechanical arm 3 and controlled by the control box 4, and the mechanical arm 3 is a mechanical arm 3 with 6 degrees of freedom.
The control box 4 is connected with the mechanical arm 3 with 6 degrees of freedom through a communication line; the gripper 10 at the tail end of the mechanical arm 3 is connected with the control box 4 through a communication line; the grid laser 8, the industrial camera 7 and the film pressure sensor are connected with the small embedded industrial personal computer firstly and then are in wired communication with the machine box controlled by the small embedded industrial personal computer and the mechanical arm 3.
The industrial camera 7 positions and classifies the luggage case 5 according to categories, the luggage case 5 needs to be sorted after the classification and the positioning, and the force of the pressure film sensor 9 on the end gripper 10 is fed back to the mechanical arm 3 to control the gripper 10 to carry out force-controlled sorting.
The control unit is used for controlling the mechanical arm 3 and the gripper 10 to realize the grabbing of the luggage case 5, placing the grabbed luggage case 5 on the transport trolley 6, and finally returning to the initial position to wait for the next grabbing task.
The detection unit comprises a grid laser 8, an industrial camera 7 and a pressure film sensor 9, the industrial camera 7 and the grid laser 8 are both arranged on the top plate 2 of the gripper 10, the grid laser 8 projects a group of parallel grid red laser lines on the surface of an object to be gripped, and the object is generally judged.
And the periphery of the industrial camera 7 is provided with a plurality of point light source lamps in a surrounding manner, the point light source lamps select the characteristics for projecting the shooting target, so that the different parts of the point light source lamps have enough contrast, and the LED luminous electro-optic is selected, and meanwhile, the defect that the industrial camera 7 can be matched well for good detection in some dim environments can be overcome.
The industrial camera 7 collects the object marked by the grid laser 8, and converts the visual image and the characteristics of the target object into a series of data which can be processed by calculation. The pressure film sensor 9 is fixedly arranged on the inner side wall of the hand grip 10.
The transport unit comprises a conveyor belt 1 and a transport trolley 6 arranged on one side of the conveyor belt 1.
According to the second embodiment of the present application, a method for controlling a system for handling a luggage case 5 at an airport comprises the following steps:
step S1, calibrating the industrial camera 7, including calibrating parameters in the industrial camera 7 and calibrating a hand-eye coordinate transformation relationship, which specifically includes:
the high-precision hand-eye calibration is a key for ensuring the precision of the system, and in the embodiment, the calibration comprises two parts of camera internal parameter calibration and hand-eye coordinate conversion relation calibration, wherein the camera internal parameter calibration and distortion correction are realized by adopting an improved Zhang Zhengyou calibration algorithm.
The calibration of the hand-eye coordinate relationship of the mechanical arm 3 mainly comprises the step of solving a hand-eye mapping model, namely a nonlinear mapping model from a robot visual space to a robot working space, such as a camera coordinate system O in the left diagram of FIG. 6cTo the manipulator base coordinate system OrThe conversion relationship of (1). The right diagram of fig. 6 shows the coordinate transformation relationship in the visual guidance process.
Referring to fig. 7, in the present solution, a Tsai-Lenz calibration method is adopted to perform hand-eye calibration, so as to obtain a spatial pose between a robot base coordinate and a camera.
Step S2, the industrial camera 7 collects the image information of the luggage case 5 on the conveyor belt 1, and transmits the collected image information to the control unit for processing, so as to obtain the position information of the luggage case 5 in the image;
referring to fig. 3, the image acquisition includes:
marking and extracting the acquired image information data by adopting a silhouette processing method;
carrying out binarization processing on the extracted image information, and carrying out image segmentation on the image subjected to binarization processing to obtain image information containing a target area of the trunk 5;
filtering the segmented image information;
adopting skeletonization to identify the position information of the target area, and transmitting the position information to a control unit;
the control unit issues commands to the control box 4 and controls the operation of the mechanical arm 3 and the gripper 10 according to the position information of the luggage box 5.
Step S3, controlling the hand grip 10 to move to the target position according to the position information of the luggage 5, controlling the hand grip 10 to move according to the moving track of the luggage 5, and setting the vertical downward movement distance and the horizontal displacement of the hand grip 10 until the hand grip 10 moves to the two sides of the luggage 5, which specifically comprises:
the posture and the volume of the luggage case 5 are intelligently identified, a group of parallel grid red laser lines are projected on the surface of an object to be grabbed by the grid laser 8, and the coordinate position of the luggage case 5 in the grid is firstly calculated by combining point light source lamps with an industrial camera 7 and object detection in machine vision;
analyzing the hough transform detection result shows that there are some problems in directly using hough transform to detect the grid lines, such as overlapping of a plurality of straight lines at the same position, shielding of light, and the fact that the start and end points of the straight lines are not actual start and end points. The problem of linear overlapping is solved by the idea of classifying and combining Hough transform detection results. Because the non-parallel straight lines have intersection points, the coordinates of the start point and the end point of the actual grid line can be obtained by calculating the intersection points of the straight lines in the horizontal direction and the vertical direction, so that the problems of partial shielding, non-true coordinates of the start point and the end point of the grid line and the like can be solved.
And (4) Hough transform, namely comparing two Hough transform functions to know, and selecting standard Hough transform. Because the result of standard Hough transform is (theta, rho) parameters under a polar coordinate system, and the theta parameter is the included angle between a straight line and a horizontal axis, the method provides a convenient way for horizontal and vertical classification arrangement of the straight line. In addition, the standard Hough transform function only has one dynamically changing parameter, and the parameter value is easy to set.
Referring to fig. 4, calculation is performed by using a hough transform straight line method:
point A (x)0,y0) The straight line passing through the point A satisfies the equation y ═ k x0+ b. (k is slope, b is intercept);
then the point A (x) is crossed in the XOY plane0,y0) The linear clusters of (a) may be represented by y ═ k × x0+ b, however, since the slope of the line perpendicular to the X-axis is infinite and cannot be represented, the special case can be solved by converting the rectangular coordinate system to the polar coordinate system;
in a polar coordinate system, the linear equation can be expressed as ρ ═ xcos θ + ysin θ (ρ represents the distance from the origin to the line), and the calculation method is shown in fig. 4.
Judging the linear direction: by judging the parameter theta, when theta belongs to a horizontal line between two epsilon (pi/4, 3 x pi/4), or is a vertical line, then sorting and combining straight line data with small difference, eliminating the problem that a plurality of straight lines are overlapped at the same position, and finally obtaining the coordinate information of the straight lines in a rectangular coordinate system through coordinate system conversion.
And calculating a horizontal straight line equation and a vertical straight line equation by using the results, and then calculating the coordinates of the start and stop points of the grid lines through the intersection points of the horizontal straight line and the vertical straight line, so that the problems caused by the fact that the coordinates of the start and stop points of the grid lines are not actual in Hough transformation and lamplight shielding can be solved.
And drawing the straight line on the basis of the binary image according to the coordinates of the straight line start and stop points obtained by calculation, and obtaining a result graph of grid line detection. The straight line is drawn on the binary image because not only can the laser line of the original position be obtained, but also the laser area deformed under the influence of the barrier can be reserved, and the problem of shielding of light on the grid laser 8 lines can be solved.
In FIG. 5, line LhsAnd a straight line LheIs a straight line in the horizontal direction in the detection result of the Hough transform of the binary image, wherein the straight line LhsIn the presentation of the results of the detectionFirst horizontal straight line, straight line LheRepresenting the last horizontal line in the test results. Straight line LvsAnd a straight line LveIs a straight line in the vertical direction in the detection result of the Hough transform of the binary image, wherein the straight line LvsRepresents the first vertical straight line, straight line L in the detection resultveRepresenting the last vertical line in the test result. And in fig. 5, the coordinate point (x)1,y1),(x2,y2),(x3,y3) And (x)4,y4) Are the start and end point coordinates of the actual grid laser 8 line.
Wherein, the coordinate point (x)v1,yv1) And (x)v2,yv2) In xv1And xv2Is calculated by Hough transform, and yv1And yv2The value of (d) is 0. Coordinate point (x)v3,yv3) And (x)v4,yv4) In xv3And xv4Calculated by Hough transform, and yh1And yh3Is the height of the image. Coordinate point (x)h1,yh1) And (x)h3,yh3) In xh1And xh3Is 0 and y ish1And yh3Is calculated by Hough transform. Coordinate point (x)h2,yh2) And (x)h4,yh4) In xh2And xh4Is the width of the image, and yh2And yh4Is calculated by Hough transform.
Then, according to the calculation method of the grid horizontal line starting point coordinates, keeping the vertical straight line equations unchanged, sequentially traversing all the horizontal straight line equations, calculating the intersection point coordinates of each horizontal straight line and the set vertical line, and finally obtaining the actual grid horizontal line starting point coordinates. For example (x)1,y1) And (x)3,y3) And the coordinates are equal. And similarly, keeping the horizontal straight line equation unchanged, sequentially traversing all the vertical straight line equations, then calculating the coordinates of the intersection points of the horizontal straight line and the vertical straight line, and finally obtaining the coordinates of the initial point of the straight line in the vertical direction of the actual grid. For example (x)1,y1) And (x)2,y2) And the coordinates are equal.
Then, according to the calculation method of the grid horizontal line terminal point coordinates, keeping the horizontal straight line equations unchanged, traversing all the vertical straight line equations in sequence, calculating the intersection point coordinates of each vertical straight line and the set horizontal straight line, and finally obtaining the terminal point coordinates of the actual grid horizontal straight lines. For example (x)2,y2) And (x)4,y4) And the coordinates are equal. And similarly, keeping the horizontal straight line equation unchanged, sequentially traversing all the vertical straight line equations, then calculating the coordinates of the intersection points of the horizontal straight line and the vertical straight line, and finally obtaining the coordinates of the end point of the straight line in the vertical direction of the actual grid. For example (x)3,y3) And (x)4,y4) And the coordinates are equal.
The size information measuring and positioning method based on the industrial camera 7 comprises the following steps:
referring to fig. 8, a reference light triggering method based on active grid laser 8 is adopted, and a size information measuring and positioning method based on an industrial camera 7 is utilized, so that the high precision of luggage size information measurement can be met, the time loss and the computing resource consumption caused by dense point cloud operation are avoided, and the real-time performance of the system is ensured. A schematic diagram of the dimensional measurement based on the industrial camera 7 is shown in fig. 7.
A circle of point light sources are arranged at the tail end of the mechanical arm 3 and matched with the industrial camera 7, and when the environmental light is dark, the point light sources can provide enough light for the industrial camera 7; meanwhile, when the grid laser 8 irradiates the surface of the luggage case 5, the point light source can well highlight the lines of the grid laser 8, and the lines which are clear enough are provided for the industrial camera 7 to photograph, so that the size and the volume of the luggage case 5 can be estimated and calculated in the rear direction.
Calculating and correcting the volume of the luggage case 5, when the grid laser 8 irradiates the target luggage case 5, the industrial camera 7 photographs the target luggage case 5, and identifies the laser grid on the photographed picture, with the edge of the luggage case 5 of the farthest curved grid line of the luggage case 5 as the farthest edge;
a first step of calculating the area of the preliminary uppermost face of the furthest-side luggage case 5, and then calculating the volume of the preliminary luggage case 5;
the second step is to perform an edge expansion filling of the calculated volume of the luggage case 5 on the basis of the first step, in order for the gripper 10 at the end of the robot arm 3 to be able to completely wrap the edge of the luggage case 5, thus allowing both ends of the gripper 10 to be transferred to the luggage case 5.
Edge filling is an optimization of the industrial camera 7 in making the target size inspection relative to the gripper 10; filling the edges of the gray frames of 5cm in total outwards on the basis of the first step, wherein the virtual modeling is performed when the size of the luggage case 5 is estimated, so that the opening degree of the hand grip 10 is greater than the lengths of the two actually longest sides of the posture of the luggage case 5 during detection; if the distance between the longest sides is just taken as the degree of opening of the hand grip 10, then the hand grip 10 may be blocked by the thickness of the hand grip 10, so that the hand grip 10 cannot be gripped, and therefore, in practical situations, the gripping state of the hand grip 10 can be well adapted by performing a size expansion, so that each luggage case 5 can be gripped.
Take 20 inch luggage case 5 as an example:
since different sizes of the luggage case 5 are fixed, the grid laser 8 is irradiated above the luggage case 5, the industrial camera 7 performs detection shooting on the luggage case 5, the area of the front face of the case is estimated, 34 x 50cm in fig. 9 represents the real area of the front face of the luggage case 5(20 inches), the length and the width are respectively expanded by 2.5cm on the basis of the length and the width calculated by the camera to reserve enough space for grabbing, the broken line (39 x 55cm) in fig. 9 is an optimization made by the industrial camera 7 relative to the grab 10 when the industrial camera performs target size detection, and the edge filling does not exist actually but is a virtual modeling made when the size of the luggage case 5 is estimated, so that the expansion degree of the grab 10 is greater than the length of the actual longest two sides of the posture of the luggage case 5 during detection. Since the height of the luggage case 5 corresponds to the corresponding size, the volume of the luggage case 5 can be estimated by multiplying the height by the area of the front surface.
And S4, grabbing the luggage case 5 according to the weight of the luggage case 5 and the pressure value fed back by the film sensor, adjusting the grabbing strength of the grab 10 in real time, detecting and feeding back the strength of the grab 10 at the tail end of the mechanical arm 3 in real time by using the pressure film sensor 9, and grabbing the luggage case 5 without damage.
Step S5, the control unit drives the mechanical arm 3 to rotate, the luggage case 5 is transported to the upper part of the transport trolley 6, the mechanical arm 3 is moved downwards continuously, the gripper 10 is controlled to release the luggage case 5, and the luggage case 5 falls on the transport trolley 6;
step S6, the control unit controls the robot arm 3 to return to the initial position.
While the embodiments of the invention have been described in detail in connection with the accompanying drawings, it is not intended to limit the scope of the invention. Various modifications and changes may be made by those skilled in the art without inventive step within the scope of the appended claims.

Claims (9)

1.一种航空港行李箱搬运系统,其特征在于:包括控制单元、与控制单元信号连接的检测单元和用于运输行李箱的运输单元;1. An airport luggage handling system is characterized in that: comprising a control unit, a detection unit signally connected with the control unit and a transport unit for transporting luggage; 所述控制单元包括控制箱和固定安装于机械臂上、且受控于控制单元的抓手;The control unit includes a control box and a gripper fixedly mounted on the robotic arm and controlled by the control unit; 所述检测单元包括网格激光、工业相机和压力薄膜传感器;所述工业相机和网格激光均安装于抓手的顶板上,且工业相机的四周环设有若干个点光源灯;所述压力薄膜传感器固定安装于抓手的内侧侧壁上;The detection unit includes a grid laser, an industrial camera and a pressure film sensor; the industrial camera and the grid laser are both mounted on the top plate of the gripper, and a number of point light sources are arranged around the industrial camera; the pressure The film sensor is fixedly installed on the inner sidewall of the gripper; 所述运输单元包括传送带和设置于传送带一侧的运输小车。The transport unit includes a conveyor belt and a transport trolley arranged on one side of the conveyor belt. 2.根据权利要求1所述的航空港行李箱搬运系统,其特征在于:所述网格激光、工业相机、薄膜压力传感器与小型嵌入式工控机连接,并通过小型嵌入式工控机与控制箱通信连接。2. The airport luggage handling system according to claim 1, wherein the grid laser, the industrial camera and the film pressure sensor are connected to a small embedded industrial computer, and communicate with the control box through the small embedded industrial computer connect. 3.根据权利要求1所述的航空港行李箱搬运系统,其特征在于:所述点光源灯为LED发光光源灯。3 . The airport luggage handling system according to claim 1 , wherein the point light source lamp is an LED light source lamp. 4 . 4.根据权利要求1-3任一所述的航空港行李箱搬运系统的控制方法,其特征在于,包括以下步骤:4. according to the control method of the arbitrary described airport luggage handling system of claim 1-3, it is characterized in that, comprising the following steps: S1、工业相机标定,包括工业相机机内参数标定和手眼坐标转换关系标定;S1. Industrial camera calibration, including industrial camera internal parameter calibration and hand-eye coordinate conversion relationship calibration; S2、工业相机采集传送带上行李箱的图像信息,并将采集的图像信息传送至控制单元进行处理,得到图像中行李箱的位置信息;S2. The industrial camera collects the image information of the luggage on the conveyor belt, and transmits the collected image information to the control unit for processing, so as to obtain the position information of the luggage in the image; S3、根据行李箱的位置信息控制抓手移动至目标位置,根据行李箱移动轨迹,控制抓手移动,设定抓手垂直下移距离和水平位移,直至抓手移动至行李箱的两侧;S3. Control the gripper to move to the target position according to the position information of the suitcase, control the gripper to move according to the movement track of the suitcase, and set the vertical downward movement distance and horizontal displacement of the gripper until the gripper moves to both sides of the suitcase; S4、根据行李箱重量和薄膜传感器反馈的压力值抓取行李箱,并实时调整抓手的抓取力度;S4. Grab the suitcase according to the suitcase weight and the pressure value fed back by the film sensor, and adjust the grasping force of the gripper in real time; S5、控制单元驱动机械臂旋转作业,将行李箱运输至运输小车上方,继续下移机械臂,控制抓手释放行李箱,行李箱落在运输小车上;S5. The control unit drives the robotic arm to rotate, transports the luggage to the top of the transport trolley, continues to move the robotic arm down, controls the gripper to release the luggage, and the luggage falls on the transport trolley; S6、控制单元控制机械臂回调至初始位置。S6, the control unit controls the mechanical arm to return to the initial position. 5.根据权利要求4所述的航空港行李箱搬运系统的控制方法,其特征在于,所述步骤S2中工业相机采集传送带上行李箱的图像信息,并将采集的图像信息传送至控制单元进行处理,得到图像中行李箱的位置信息,包括:5. The control method of the airport luggage handling system according to claim 4, wherein in the step S2, the industrial camera collects the image information of the luggage on the conveyor belt, and transmits the collected image information to the control unit for processing , get the location information of the luggage in the image, including: 采用剪影处理法对采集的图像信息数据进行标记提取;Use the silhouette processing method to mark and extract the collected image information data; 对提取的图像信息进行二值化处理,并对二值化处理后的图像进行图像分割,得到含有行李箱目标区域的图像信息;Perform binarization processing on the extracted image information, and perform image segmentation on the binarized image to obtain image information containing the target area of the suitcase; 对分割后的图像信息进行滤波处理;Filter the segmented image information; 采用骨架化识别目标区域位置信息,并将位置信息传送至控制单元。Use skeletonization to identify the location information of the target area, and transmit the location information to the control unit. 6.根据权利要求4所述的航空港行李箱搬运系统的控制方法,其特征在于,所述步骤S3中计算行李箱的移动轨迹,包括:6. The control method of the airport luggage handling system according to claim 4, characterized in that, calculating the movement trajectory of the luggage in the step S3, comprising: 网格激光对待抓取行李箱的表面投射一组平行网格红色激光线,工业相机配合点光源进行图像拍摄,并将拍摄的图像信息传输至控制单元;The grid laser projects a set of parallel grid red laser lines on the surface of the suitcase to be grabbed, the industrial camera cooperates with the point light source to capture images, and transmits the captured image information to the control unit; 采用标准霍夫变换计算网格中水平直线方程和垂直直线方程,然后通过水平直线和垂直直线的交点计算网格线的起止点坐标;The standard Hough transform is used to calculate the horizontal line equation and the vertical line equation in the grid, and then the coordinates of the starting and ending points of the grid lines are calculated through the intersection of the horizontal and vertical lines; 将计算得到的起止点坐标在二值化图像的基础上进行直线绘制,并得出网格线检测的结果图。The calculated starting and ending point coordinates are drawn on the basis of the binarized image, and the result graph of grid line detection is obtained. 7.根据权利要求6所述的航空港行李箱搬运系统的控制方法,其特征在于,根据网格线检测的结果图,计算网格水平线起始点坐标:7. The control method of the airport luggage handling system according to claim 6, is characterized in that, according to the result graph of grid line detection, calculate grid horizontal line starting point coordinates: 保持垂直直线方程不变,依次遍历所有水平直线方程,并计算每一条水平直线和既定的垂直线的交点坐标,计算得到实际网格水平直线的起始点坐标;Keep the vertical straight line equation unchanged, traverse all the horizontal straight line equations in turn, and calculate the intersection coordinates of each horizontal straight line and a predetermined vertical line, and calculate the starting point coordinates of the actual grid horizontal straight line; 保持水平直线方程不变,依次遍历所有垂直直线方程,计算水平直线和垂直直线的交点坐标,计算得到实际网格垂直方向直线的起始点坐标;Keep the horizontal straight line equation unchanged, traverse all the vertical straight line equations in turn, calculate the intersection coordinates of the horizontal straight line and the vertical straight line, and calculate the starting point coordinates of the vertical straight line of the actual grid; 根据网格线检测的结果图,计算网格水平线终点坐标:Calculate the coordinates of the end point of the grid horizontal line according to the result of grid line detection: 保持水平直线方程不变,依次遍历所有垂直直线方程,并计算每一条垂直直线和既定的水平直线的交点坐标,计算得到实际网格水平直线的终点坐标;Keep the horizontal straight line equation unchanged, traverse all vertical straight line equations in turn, and calculate the intersection coordinates of each vertical straight line and a predetermined horizontal straight line, and calculate the end point coordinates of the actual grid horizontal straight line; 保持水平直线方程不变,依次遍历所有垂直直线方程,计算水平直线和垂直直线的交点坐标,得到实际网格垂直方向直线的终点坐标。Keep the equation of the horizontal straight line unchanged, traverse all the equations of the vertical straight line in turn, calculate the coordinates of the intersection point of the horizontal straight line and the vertical straight line, and obtain the coordinates of the end point of the straight line in the vertical direction of the actual grid. 8.根据权利要求4所述的航空港行李箱搬运系统的控制方法,其特征在于,所述步骤S1中工业相机标定,包括:8. The control method of the airport luggage handling system according to claim 4, wherein the calibration of the industrial camera in the step S1 comprises: 采用张正友标定算法进行工业相机机内参数标定和畸变校正;The calibration algorithm of Zhang Zhengyou is used to calibrate the internal parameters of the industrial camera and correct the distortion; 采用Tsai-Lenz标定法进行手眼标定。The hand-eye calibration was performed using the Tsai-Lenz calibration method. 9.根据权利要求4所述的航空港行李箱搬运系统的控制方法,其特征在于,还包括行李箱的体积的计算与校正:9. The control method of airport luggage handling system according to claim 4, is characterized in that, also comprises the calculation and correction of the volume of luggage: 当网格激光照射到目标行李箱上时,工业相机对目标行李箱进行拍照,对拍照的图片上的激光网格进行识别,以行李箱的最远的弯曲的网格线的行李箱的边缘为最远边;When the grid laser irradiates the target suitcase, the industrial camera takes a picture of the target suitcase, identifies the laser grid on the photographed picture, and takes the edge of the suitcase with the furthest curved grid line of the suitcase. for the farthest edge; 计算以最远边行李箱的最上面的一个面的面积,进而估算出行李箱的初步体积;Calculate the area of the uppermost face of the farthest luggage, and then estimate the initial volume of the luggage; 对计算得到的行李箱的体积做边缘扩张填充,边缘填充为在行李箱的最远的弯曲的网格线向外填充总共5cm的灰色边框。The calculated volume of the suitcase is filled with edge expansion, and the edge fill is filled with a total of 5cm of gray borders at the farthest curved grid line of the suitcase.
CN202110897488.2A 2021-08-05 2021-08-05 Airport luggage case carrying system and control method thereof Active CN113602799B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110897488.2A CN113602799B (en) 2021-08-05 2021-08-05 Airport luggage case carrying system and control method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110897488.2A CN113602799B (en) 2021-08-05 2021-08-05 Airport luggage case carrying system and control method thereof

Publications (2)

Publication Number Publication Date
CN113602799A true CN113602799A (en) 2021-11-05
CN113602799B CN113602799B (en) 2022-09-13

Family

ID=78307207

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110897488.2A Active CN113602799B (en) 2021-08-05 2021-08-05 Airport luggage case carrying system and control method thereof

Country Status (1)

Country Link
CN (1) CN113602799B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115108316A (en) * 2022-08-29 2022-09-27 民航成都物流技术有限公司 Baggage pickup clamp, apparatus and method
CN115112508A (en) * 2022-08-29 2022-09-27 民航成都物流技术有限公司 Device and method for identifying soft and hard bags of consigned luggage in civil aviation airport
CN118544361A (en) * 2024-07-25 2024-08-27 中航信移动科技有限公司 Luggage handling system based on computer vision and mechanical arm

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1067068A1 (en) * 1999-07-08 2001-01-10 ABBPATENT GmbH Method and device for luggage transport in airports
NL1023904C2 (en) * 2003-07-11 2005-01-12 Csi Ind B V Automatic airport baggage handling process, has automatic loading controlled using measured characteristics of baggage pieces and baggage container loading condition
CN106927079A (en) * 2017-03-21 2017-07-07 长春理工大学 A kind of industrial detonator crawl and packaging system and method based on machine vision
CN108161913A (en) * 2017-12-31 2018-06-15 柳州福能机器人开发有限公司 A kind of intelligent mobile transfer robot and its method of work
CN208795188U (en) * 2018-09-05 2019-04-26 无锡维胜威信息科技有限公司 A kind of structured light binocular vision detection system
CN110509281A (en) * 2019-09-16 2019-11-29 中国计量大学 Device and method for pose recognition and grasping based on binocular vision
CN110660104A (en) * 2019-09-29 2020-01-07 珠海格力电器股份有限公司 Industrial robot visual identification positioning grabbing method, computer device and computer readable storage medium
CN112047113A (en) * 2020-08-26 2020-12-08 苏州中科全象智能科技有限公司 3D visual stacking system and method based on artificial intelligence technology
CN112136506A (en) * 2020-09-27 2020-12-29 哈尔滨理工大学 Robotic arm device with fruit ripeness discrimination function
CN112374119A (en) * 2020-11-05 2021-02-19 泉州装备制造研究所 Self-adaptive airport logistics system

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1067068A1 (en) * 1999-07-08 2001-01-10 ABBPATENT GmbH Method and device for luggage transport in airports
NL1023904C2 (en) * 2003-07-11 2005-01-12 Csi Ind B V Automatic airport baggage handling process, has automatic loading controlled using measured characteristics of baggage pieces and baggage container loading condition
CN106927079A (en) * 2017-03-21 2017-07-07 长春理工大学 A kind of industrial detonator crawl and packaging system and method based on machine vision
CN108161913A (en) * 2017-12-31 2018-06-15 柳州福能机器人开发有限公司 A kind of intelligent mobile transfer robot and its method of work
CN208795188U (en) * 2018-09-05 2019-04-26 无锡维胜威信息科技有限公司 A kind of structured light binocular vision detection system
CN110509281A (en) * 2019-09-16 2019-11-29 中国计量大学 Device and method for pose recognition and grasping based on binocular vision
CN110660104A (en) * 2019-09-29 2020-01-07 珠海格力电器股份有限公司 Industrial robot visual identification positioning grabbing method, computer device and computer readable storage medium
CN112047113A (en) * 2020-08-26 2020-12-08 苏州中科全象智能科技有限公司 3D visual stacking system and method based on artificial intelligence technology
CN112136506A (en) * 2020-09-27 2020-12-29 哈尔滨理工大学 Robotic arm device with fruit ripeness discrimination function
CN112374119A (en) * 2020-11-05 2021-02-19 泉州装备制造研究所 Self-adaptive airport logistics system

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115108316A (en) * 2022-08-29 2022-09-27 民航成都物流技术有限公司 Baggage pickup clamp, apparatus and method
CN115112508A (en) * 2022-08-29 2022-09-27 民航成都物流技术有限公司 Device and method for identifying soft and hard bags of consigned luggage in civil aviation airport
CN118544361A (en) * 2024-07-25 2024-08-27 中航信移动科技有限公司 Luggage handling system based on computer vision and mechanical arm

Also Published As

Publication number Publication date
CN113602799B (en) 2022-09-13

Similar Documents

Publication Publication Date Title
CN108555908B (en) A method for gesture recognition and picking of stacked workpieces based on RGBD cameras
CN108399639B (en) Rapid automatic grabbing and placing method based on deep learning
CN109230580B (en) Unstacking robot system and unstacking robot method based on mixed material information acquisition
CN113602799A (en) A kind of airport luggage handling system and control method thereof
CN110580725A (en) A kind of box sorting method and system based on RGB-D camera
JP5429614B2 (en) Box-shaped workpiece recognition apparatus and method
CN106737665B (en) Manipulator control system and implementation method based on binocular vision and SIFT feature matching
CN112010024B (en) Automatic container grabbing method and system based on laser and vision fusion detection
JPWO2009028489A1 (en) Object detection method, object detection apparatus, and robot system
CN115070781B (en) Object grabbing method and two-mechanical-arm cooperation system
CN111311691A (en) Depalletizing robot depalletizing method and system
CN110980276B (en) A method for automatic blanking of castings with three-dimensional vision and robots
CN113666028B (en) Garbage can detecting and grabbing method based on fusion of laser radar and camera
JP7398763B2 (en) Robot system equipped with overlap processing mechanism and its operating method
CN113021391A (en) Integrated vision robot clamping jaw and using method thereof
CN117124302B (en) Parts sorting method, device, electronic equipment and storage medium
CN115582827A (en) A Grasping Method of Unloading Robot Based on 2D and 3D Vision Positioning
JP7408107B2 (en) Systems and methods for robotic systems with object handling
CN116758017A (en) A visual real-time detection method for the precise quantity and abnormality of bagged foods densely arranged vertically on the assembly line
JP2010210511A (en) Recognition device of three-dimensional positions and attitudes of objects, and method for the same
CN116572253A (en) Method and device for grasping and controlling a test tube
CN113731860A (en) Automatic sorting system and method for objects stacked in container
CN119005255B (en) A method for sorting disordered stacked materials based on deep learning and point cloud
CN216511500U (en) Object unstacking system
CN115837985B (en) Disordered grabbing method based on machine vision

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant