ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations
"> Figure 1
<p>Example image of background variations in the ABOships dataset: (<b>a</b>) View of maritime vessels on Aura river including the urban landscape; (<b>b</b>) View of a maritime vessel in the Finnish Archipelago.</p> "> Figure 2
<p>Example image of a occlusion: (<b>a</b>) Boat in front of a militaryship; (<b>b</b>) Several sailboats occluding each other while docked, on the right half of the image.</p> "> Figure 3
<p>Example image of a sailboat, view from two perspectives: (<b>a</b>) Lateral; (<b>b</b>) Frontal.</p> "> Figure 4
<p>Example images of annotated objects in the ABOships dataset: (<b>a</b>) boat, (<b>b</b>) cargoship, (<b>c</b>) cruiseship, (<b>d</b>) ferry, (<b>e</b>) militaryship, (<b>f</b>) miscboat, (<b>g</b>) miscellaneous (floater), (<b>h</b>) motorboat, (<b>i</b>) passengership, (<b>j</b>) sailboat and (<b>k</b>) seamark.</p> "> Figure 5
<p>The video collection was separated into 48 workpackages of images (<b>1</b>), which were labelled in an initial labelling step (<b>2</b>). Using the OpenCV Tracker, the objects were tracked across frames to produce traces (<b>3</b>) and then relabelled to fix inconsistencies and fill in the labels that might have been skipped (<b>4</b>). The resulting labels were then compiled into the maritime imagery dataset (<b>5</b>).</p> "> Figure 6
<p>The relabelling process utilized our relabelling software application. Its GUI (graphical user-interface) shows the annotator traces of tracked images between annotation frames (<b>1</b>). The annotator is required to either relabel every instance by selecting the correct label from the right panel, or edit an annotation (by selecting a label that emerged distinct from others (<b>2</b>)) and change the label of each image individually and possibly fix the bounding box to fit the object more tightly (<b>3</b>). Special attention was required in certain situations when the tracker would drift onto other objects, in which case that particular entry of the trace might have had a different label from the rest (<b>4</b>). When all labels belonging to a trace were verified and steps (<b>1</b>)–(<b>4</b>) were completed (<b>5</b>), the changes were saved into a new file and the annotator was provided with the next trace.</p> "> Figure 7
<p>Histograms of occupied pixel area at <math display="inline"><semantics> <mrow> <mi>l</mi> <mi>o</mi> <msub> <mi>g</mi> <mn>2</mn> </msub> </mrow> </semantics></math>-scale for all annotated objects by object category, divided into three groups for each category: small, medium and large according to Microsoft COCO variants (small: log<math display="inline"><semantics> <msub> <mrow/> <mn>2</mn> </msub> </semantics></math>(area) < 10, medium: 10 < log<math display="inline"><semantics> <msub> <mrow/> <mn>2</mn> </msub> </semantics></math>(area) < <math display="inline"><semantics> <mrow> <mn>13.16</mn> </mrow> </semantics></math> and large: log<math display="inline"><semantics> <msub> <mrow/> <mn>2</mn> </msub> </semantics></math>(area) > <math display="inline"><semantics> <mrow> <mn>13.16</mn> </mrow> </semantics></math>). The vertical colored lines represent the following values: the red line—represents the mean of the distribution, the yellow line represents the threshold for small objects and the green vertical line delineates the threshold for large objects. In each histogram, respectively, entries to the left of the yellow line represent the small objects group, entries in between the yellow and the green line show the medium-sized objects group and those to the right of the green line depict the large objects group.</p> "> Figure 8
<p>Qualitative detection results for the ABOships dataset on (<b>a</b>) Faster R-CNN and Inception-ResNet-v2 as feature extractor, (<b>b</b>) EfficientDet with EfficientNet as feature extractor, (<b>c</b>) R-FCN with ResNet101 as feature extractor, and (<b>d</b>) SSD with ResNet101 as feature extractor. The ground truth bounding-boxes are shown as red rectangles. Predicted boxes by these methods are depicted as green bounding boxes. Each output box is associated with a class label and a score with a value in the interval [0, 1].</p> ">
Abstract
:1. Introduction
2. Related Work
2.1. Object Detection
- Hand-annotated visual descriptors provided large number of proposals, which caused high rates of false positives.
- Visual descriptors (as mentioned above) extract low-level features, but are unsuitable for high-level features.
- Each step of a detection pipeline is optimized separately, so global optimization is difficult to attain.
2.2. General Object Detection Datasets
2.3. Maritime Vessel Detection Datasets
3. Materials and Methods
3.1. Camera System
3.2. Dataset Diversity
3.3. Dataset Design
3.4. Annotation
- boat—rowing boats or oval-shaped boats (from a lateral perspective), or small-sized boats, visual distinction—rowing-like boats even if they possess engine power;
- cargoship—large-scale ships used for cargo transportation, visual distinction—long ship with cargo containers or designed with container carrying capacity;
- cruiseship—large ship that transports passengers and/or cars on longer distances (assumed at least some hundreds of km);
- ferry—medium-sized ship, used to transport people and cars, a.k.a. waterbus/watertaxi, another appropriate term would be cableferry, visual distinction—it includes entrances on two opposite sides and a cabin in the middle;
- militaryship—an official ship that is either military or Coast Guard and includes a special hull with antennas. For Coast Guard fleets, usually the hulls of their ships read “Coast Guard” and the military ones are dark gray/metallic/black/brown in colour;
- miscboat—miscellaneous maritime vessel, visual distinction—generic boat that does not include any visual distinction mentioned in the other ship categories;
- miscellaneous—identified floaters (birds, other objects floating in the water) or unidentified/unidentifiable floaters;
- motorboat—primarily a speedboat, visual distinction—sleek, aerodynamic features;
- passengership—medium-sized ship, used to transport people on short distances, ex. restaurant boat, visual distinction-usually it has multiple noticeable lateral windows;
- sailboat—sails-propelled boat or a boat which exhibits sails, visual distinction—sails;
- seamark—green/red/blue/black/yellow cone-shaped metal/plastic floater or pipe emerging from the sea.
3.5. Relabelling Algorithm
3.6. Dataset Statistics
4. Results
4.1. Evaluation Criteria
4.2. Baseline Detection
5. Qualitative Results
6. Discussion
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Shao, Z.; Wang, L.; Wang, Z.; Du, W.; Wu, W. Saliency-aware convolution neural network for ship detection in surveillance video. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 781–794. [Google Scholar] [CrossRef]
- Liu, S. Remote Sensing Satellite Image Acquisition Planning: Framework, Methods and Application. Ph.D. Thesis, University of South Carolina, Columbia, SC, USA, 2014. [Google Scholar]
- Shao, Z.; Wu, W.; Wang, Z.; Du, W.; Li, C. Seaships: A large-scale precisely annotated dataset for ship detection. IEEE Trans. Multimed. 2018, 20, 2593–2604. [Google Scholar] [CrossRef]
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Druzhkov, P.; Kustikova, V. A survey of deep learning methods and software tools for image classification and object detection. Pattern Recognit. Image Anal. 2016, 26, 9–15. [Google Scholar] [CrossRef]
- Zhang, J.; Li, W.; Ogunbona, P.; Xu, D. Recent advances in transfer learning for cross-dataset visual recognition: A problem-oriented perspective. ACM Comput. Surv. (CSUR) 2019, 52, 1–38. [Google Scholar] [CrossRef]
- Wu, X.; Sahoo, D.; Hoi, S.C. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
- Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep learning for generic object detection: A survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef] [Green Version]
- Pathak, A.R.; Pandey, M.; Rautaray, S. Application of deep learning for object detection. Procedia Comput. Sci. 2018, 132, 1706–1717. [Google Scholar] [CrossRef]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In European Conference on Computer Vision; Springer: Zurich, Switzerland, 2014; pp. 740–755. [Google Scholar]
- Miller, G.A. WordNet: A lexical database for English. Commun. ACM 1995, 38, 39–41. [Google Scholar] [CrossRef]
- Kanjir, U.; Greidanus, H.; Oštir, K. Vessel detection and classification from spaceborne optical images: A literature survey. Remote. Sens. Environ. 2018, 207, 1–26. [Google Scholar] [CrossRef]
- Arshad, N.; Moon, K.S.; Kim, J.N. Multiple ship detection and tracking using background registration and morphological operations. In Signal Processing and Multimedia; Springer: Jeju Island, Korea, 2010; pp. 121–126. [Google Scholar]
- Wijnhoven, R.; van Rens, K.; Jaspers, E.; de With, P.H.N. Online learning for ship detection in maritime surveillance. In Proceedings of the 31th Symposium on Information Theory in the Benelux, Rotterdam, The Netherlands, 11–12 May 2010; pp. 73–80. [Google Scholar]
- Lee, S.J.; R., M.I.; Lee, H.W.; Ha, J.S.; Woo, I.G. Image-Based Ship Detection and Classification for Unmanned Surface Vehicle Using Real-Time Object Detection Neural Networks. In Proceedings of the 28th International Ocean and Polar Engineering Conference, International Society of Offshore and Polar Engineers, Sapporo, Hokkaido, Japan, 10–15 June 2018. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems; ACM: Montreal, QC, Canada, 2015; pp. 91–99. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Amsterdam, The Netherlands, 2016; pp. 21–37. [Google Scholar]
- Zheng, Y.; Zhang, S. Mcships: A Large-Scale Ship Dataset for Detection Furthermore, Fine-Grained Categorization in the Wild. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
- Lukezic, A.; Vojir, T.; Cehovin Zajc, L.; Matas, J.; Kristan, M. Discriminative correlation filter with channel and spatial reliability. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6309–6318. [Google Scholar]
- Padilla, R.; Netto, S.L.; da Silva, E.A. A survey on performance metrics for object-detection algorithms. In Proceedings of the 2020 International Conference on Systems, Signals and Image Processing (IWSSIP), Niteroi, Brazil, 1–3 July 2020; pp. 237–242. [Google Scholar]
- Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 10781–10790. [Google Scholar]
- Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. In Advances in Neural Information Processing Systems; ACM: New York, NY, USA, 2016; pp. 379–387. [Google Scholar]
- TensorFlow Object Detection API. Available online: https://github.com/tensorflow/models/tree/master/research/object_detection (accessed on 4 February 2021).
- Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
General Object Detection Dataset | |||
---|---|---|---|
Dataset | Total Images | Total Classes | Annotations |
ImageNet | 14,197,122 | 1000 | 1,034,908 |
COCO | 330,000 | 91 | 2,500,000 |
OpenImage (V6) | 9,000,000 | 600 | 16,000,000 |
PASCAL VOC (2012) | 11,530 | 20 | 27,450 |
Maritime Vessel Instances | |
---|---|
Dataset | Vessel Count |
ImageNet | 1071 |
COCO | 3146 |
OpenImage | 1000 |
PASCAL VOC | 353 |
Datasets for Ship Detection | |||
---|---|---|---|
Name | Total Images | Annotations | Ship Types Included |
SeaShips | 31,455 | 40,077 | 6 |
Singapore | 17,450 | 192,980 | 6 |
MCShips | 14,709 | 26,529 | 13 |
ABOShips | 9880 | 41,967 | 9 |
Number of Images and Annotations for Every Object Category | ||||
---|---|---|---|---|
Class | Images | Percentage | Objects | Percentage |
Seamark | 3744 | 37.89% | 7670 | 18.27% |
Boat | 2034 | 20.58% | 2913 | 6.94% |
Sailboat | 3842 | 38.88% | 8147 | 19.41% |
Motorboat | 4062 | 41.11% | 7092 | 16.89% |
Passengership | 2639 | 26.71% | 4464 | 10.63% |
Cargoship | 157 | 1.58% | 161 | 0.38% |
Ferry | 945 | 9.56% | 1046 | 2.49% |
Miscboat | 2797 | 28.30% | 4642 | 11.06% |
Miscellaneous | 129 | 1.30% | 200 | 0.47% |
Militaryship | 2559 | 25.90% | 4128 | 9.83% |
Cruiseship | 1347 | 13.63% | 1504 | 3.58% |
Detection Performance of Different Detectors on the ABOships Dataset | |||||
---|---|---|---|---|---|
Method | Feature Extractor | ||||
Faster RCNN | Inception ResNet V2 | 23.16 | 30.86 | 46.84 | 35.18 |
ResNet50 V1 | 9.76 | 20.94 | 41.65 | 26.49 | |
ResNet101 | 18.42 | 25.07 | 38.17 | 30.26 | |
SSD | ResNet101 V1 FPN | 21.39 | 31.18 | 42.07 | 30.03 |
MobileNet V1 FPN | 12.34 | 27.61 | 37.83 | 28.59 | |
MobileNet V2 | 3.01 | 17.05 | 27.37 | 17.48 | |
EfficientDet | EfficientNet D1 | 10.94 | 29.68 | 55.48 | 33.83 |
RFCN | ResNet101 | 18.05 | 26.20 | 41.61 | 32.46 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Iancu, B.; Soloviev, V.; Zelioli, L.; Lilius, J. ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations. Remote Sens. 2021, 13, 988. https://doi.org/10.3390/rs13050988
Iancu B, Soloviev V, Zelioli L, Lilius J. ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations. Remote Sensing. 2021; 13(5):988. https://doi.org/10.3390/rs13050988
Chicago/Turabian StyleIancu, Bogdan, Valentin Soloviev, Luca Zelioli, and Johan Lilius. 2021. "ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations" Remote Sensing 13, no. 5: 988. https://doi.org/10.3390/rs13050988
APA StyleIancu, B., Soloviev, V., Zelioli, L., & Lilius, J. (2021). ABOships—An Inshore and Offshore Maritime Vessel Detection Dataset with Precise Annotations. Remote Sensing, 13(5), 988. https://doi.org/10.3390/rs13050988