Kim et al., 2024 - Google Patents
Adnet: Non-local affinity distillation network for lightweight depth completion with guidance from missing lidar pointsKim et al., 2024
- Document ID
- 3834304578153299325
- Author
- Kim J
- Noh J
- Jeong M
- Lee W
- Park Y
- Park J
- Publication year
- Publication venue
- IEEE Robotics and Automation Letters
External Links
Snippet
Depth completion is one of the crucial methods to estimate dense depth information of surrounding environments for various real-world applications such as autonomous driving, robotics, and augmented reality. In these real-world applications, it is strictly required for a …
- 238000004821 distillation 0 title abstract description 52
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/02—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using radio waves
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00624—Recognising scenes, i.e. recognition of a whole field of perception; recognising scene-specific objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S3/00—Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Yuan et al. | RGGNet: Tolerance aware LiDAR-camera online calibration with geometric deep learning and generative model | |
| Zhou et al. | Towards deep radar perception for autonomous driving: Datasets, methods, and challenges | |
| Abbasi et al. | Lidar point cloud compression, processing and learning for autonomous driving | |
| Cho et al. | A large rgb-d dataset for semi-supervised monocular depth estimation | |
| Qu et al. | An outline of multi-sensor fusion methods for mobile agents indoor navigation | |
| Ding et al. | Novel pipeline integrating cross-modality and motion model for nearshore multi-object tracking in optical video surveillance | |
| Shen et al. | YCANet: Target detection for complex traffic scenes based on camera-LiDAR fusion | |
| Kim et al. | Adnet: Non-local affinity distillation network for lightweight depth completion with guidance from missing lidar points | |
| Tong et al. | Edge-assisted epipolar transformer for industrial scene reconstruction | |
| Li et al. | MonoIndoor++: Towards better practice of self-supervised monocular depth estimation for indoor environments | |
| Lee et al. | $^{2} $: LiDAR-camera loop constraints for cross-modal place recognition | |
| Nguyen et al. | Roi-based lidar sampling algorithm in on-road environment for autonomous driving | |
| Kurniawan et al. | ClusterFusion: Leveraging radar spatial features for radar-camera 3D object detection in autonomous vehicles | |
| Xing et al. | ROIFormer: semantic-aware region of interest transformer for efficient self-supervised monocular depth estimation | |
| Lang et al. | BEV-TP: end-to-end visual perception and trajectory prediction for autonomous driving | |
| Rohan et al. | A systematic literature review on deep learning-based depth estimation in computer vision | |
| Alaba et al. | Multi-sensor fusion 3D object detection for autonomous driving | |
| Zhou et al. | Resolution-sensitive self-supervised monocular absolute depth estimation: Y. Zhou et al. | |
| Mondal et al. | Fusion of color and hallucinated depth features for enhanced multimodal deep learning-based damage segmentation | |
| Liu et al. | Reprojection errors as prompts for efficient scene coordinate regression | |
| Wang et al. | Unsupervised scale network for monocular relative depth and visual odometry | |
| Hong et al. | Real-time 3D visual perception by cross-dimensional refined learning | |
| Yu et al. | DMFusion: LiDAR-camera fusion framework with depth merging and temporal aggregation: X. Yu et al. | |
| Wang et al. | Perceptual enhancement for unsupervised monocular visual odometry | |
| Xu et al. | Depth-Vision-Decoupled Transformer With Cascaded Group Convolutional Attention for Monocular 3-D Object Detection |