40 - Benchmarking Deep Learning Models For Object
40 - Benchmarking Deep Learning Models For Object
{daghash.alqahtani@student, adel.toosi}@unimelb.edu.au
2
Monash University, Melbourne, Victoria, Australia
Aamir.Cheema@monash.edu
1 Introduction
Object detection is crucial in computer vision for identifying and locating ob-
jects in images or videos. It helps organizations optimize and automate processes
and has diverse applications in fields like autonomous vehicles, surveillance,
retail, healthcare, agriculture, manufacturing, sports analytics, environmental
monitoring, and smart cities. Object detection is fundamental to autonomous
transportation, enabling precise recognition of pedestrians, obstacles, and other
vehicles to ensure safe operation. In autonomous vehicles, for example, object
2 D. Alqahtani et al.
detectors identify the vehicle’s position, track other objects, and facilitate route
planning. Onboard computing units and Internet of Things (IoT) sensors enable
local, real-time data processing, allowing for rapid responses to avoid collisions.
These capabilities highlight the significant value of object detection in enhancing
the performance and safety of automated systems [1].
The field of object detection is marked by continuous technological progress,
with ongoing refinement of detection algorithms to improve accuracy and speed
in complex environments. The emergence of edge computing enables real-time
object detection on devices like smartphones, drones, and IoT devices, reducing
latency and cloud dependence. However, significant challenges remain in devel-
oping robust object detection systems for edge computing due to constrained
resources and varying energy requirements. Researchers and industries must se-
lect appropriate models and edge devices to balance accuracy, processing speed,
and energy consumption.
In this paper, we aim to evaluate the performance of most popular deep
learning models for object detection across prominent edge devices, collecting key
metrics such as energy consumption, inference time, and accuracy. Additionally,
we provide insights for deploying these models on the investigated edge devices.
Our key contributions can be summarized as follows:
TPU Accelerator: The Coral USB Accelerator [6] is a USB device that func-
tions as an Edge TPU co-processor for computational devices. It features a
USB-C port, allowing it to connect to a host computer and accelerate machine
learning inference tasks. The Edge TPU, an Application-Specific Integrated Cir-
cuit developed by Google, is designed to execute machine learning models on
edge computing devices like the Raspberry Pi.
NVIDIA Jetson: The NVIDIA Jetson Orin Series represents the latest in de-
veloper boards for energy-efficient autonomous machinery. With up to 275 tril-
lion operations per second (TOPS) and an 8X performance improvement over
the previous generation, the Jetson Orin modules enable numerous concurrent AI
inference pipelines. Their support for high-speed interfaces and multiple sensors
makes them ideal for advanced robotics. The Orin Nano, the entry-level module
in this series, focuses on low power consumption and cost efficiency while main-
taining AI processing capabilities. It is suitable for edge AI devices, IoT devices,
and embedded systems where space, power, and cost efficiency are critical [16].
You Only Look Once (YOLO): Introduced in 2015 by Redmon et al. [17],
the YOLO algorithm represents a significant advancement in real-time object de-
tection. Unlike conventional methods that apply a classifier to multiple regions
within an image, YOLO models treat detection as a single regression problem,
4 D. Alqahtani et al.
predicting bounding boxes and class probabilities from entire images in a sin-
gle evaluation. This approach significantly improves detection speed, making
YOLO ideal for real-time applications. The algorithm has evolved through sev-
eral iterations, including YOLOv2, YOLOv3, YOLOv4, and the latest YOLOv8
developed by Ultralytics [19], each enhancing accuracy, speed, and the ability to
detect a wider range of object sizes.
Single shot multibox detector (SSD): The SSD algorithm predicts multiple
bounding boxes and their corresponding class scores in a single pass. It uses mul-
tiple feature maps from different network layers to perform detections at various
scales. SSD’s key innovation is using feature maps from various network layers
to predict detections at multiple scales, addressing the challenge of detecting
objects of different sizes. Smaller feature maps detect larger objects, while larger
feature maps detect smaller objects [14].
Deep learning frameworks play a crucial role in developing and deploying object
detection models by providing comprehensive tools and functionalities. These
frameworks facilitate building, training, and implementing object detection sys-
tems with pre-built models, data augmentation techniques, and utilities for var-
ious development stages. Examples include TensorFlow Lite, PyTorch, and Ten-
sorRT. They enable researchers and developers to efficiently manage the com-
plexities of object detection tasks, from data preprocessing and model training
to evaluation and deployment.
3 Performance Evaluation
3.1 Metrics
Inference Time: This metric measures the time taken by each model from re-
ceiving the input image to producing detection results, excluding pre-processing
or post-processing steps. The inference time is crucial for real-time object detec-
tion. We report the inference time in milliseconds for each model on each device,
averaging over a series of test images for consistent and reliable measurements.
edge devices as Table 1 displays that. The choice of software and frameworks
is influenced by the need to optimize performance for each specific device, in-
cluding those with CPUs, TPUs, and GPUs. We use PyTorch to deploy and run
YOLOv8 models on the Raspberry Pi 3, Raspberry Pi 4, and Raspberry Pi 5.
To leverage the capabilities of TPUs, the YOLOv8 models are converted from
PyTorch to TensorFlow Lite (TFLite) format and compiled to run on Raspberry
Pis with TPUs. Please note that to run YOLOv8 on a TPU, we have compressed
the model size to ensure it is executable on the TPU. This means the image size
is reduced from 640x640 to 320x320. For deployment on the NVIDIA Jetson
Orin Nano, the YOLOv8 models are converted to TensorRT format to optimize
for the GPU.
EfficientDet Lite and SSD models are initially in TFLite format from Coral [7]
to be deployed on the Raspberry Pi devices. These models are compiled to run
on TPUs in the respective Raspberry Pi devices. For the NVIDIA Jetson Orin
Nano, EfficientDet Lite and SSD models are converted from TFLite to TensorRT
format for optimized performance on the GPU.
The operating systems used are Raspberry Pi OS (Bookwork - 64 bit) for the
Raspberry Pi devices and Jetson Linux (Ubuntu-based) for the NVIDIA Jetson
Orin Nano. Additional libraries and tools, such as OpenCV for image processing
and FiftyOne for model evaluation, are utilized to facilitate the experiments. To
detect objects in an image, we write Python code and utilize the Flask-RESTful
library to run this code as a service with an API URL. This approach allow
us to deploy the object detection functionality as a web service, enabling easy
integration and testing across different edge devices.
file. Next, to automate the testing, a bash script on the agent device ran the
object detection service on the edge device, followed by the Locust file. Upon
completion, the script terminated the service and proceeded to the next model,
ensuring consistent testing across all devices. Experiments were repeated three
times for each model on each device, with the average values used for the final
analysis. Finally, the accuracy measurement involved separate Python scripts
using the FiftyOne tool to download the COCO validation dataset, and the ob-
ject detection models were run on this dataset to calculate accuracy metrics,
including Precision, Recall, F1 score, and mean Average Precision, which were
recorded in a CSV file.
Agent Device
Locust File Automation Test File
Write response in
CSV file Kill the proccess
Inference Time: This section analyzes the inference times of object detec-
tion models. The measurements, reported in milliseconds, reveal distinct perfor-
mance patterns across these platforms. Beginning with Pi3, the SSD_v1 model
exhibits the lowest inference time at 427 ms among all evaluated models. While
the Det_lite models require longer inference times compared to SSD_v1 and
SSD_lite, they still outperform all variants of the YOLO8 model. Notably, the
YOLO8 models demonstrate the highest inference times, with the maximum
recorded at 12,960 ms. When the Coral USB Accelerator is integrated with the
Benchmarking Object Detection Models on Edge Devices 9
U
Pi5
Or TPU
no
SS v1
SS v1
De ite0
De ite1
SS e2
Yol te
Yol _n
Yol _s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol _s
_m
_TP
_TP
P
Na
li
o8
o8
o8
o8
D_
D_
it
it
l
_
o8
o8
D_
D_
Pi4
Pi5
t_l
t_l
t_l
t_l
t_l
t_l
in
De
De
Device Name Model Name Model Name
(a) Base Energy (b) Raspberry Pi3 (c) Pi3 + TPU
6 0.8 Excl. Base Energy
5
Excl. Base Energy Excl. Base Energy
SS v1
SS v1
De te0
De te1
SS 2
Yol te
Yol n
Yol _s
_m
De 0
De ite1
SS e2
Yol ite
Yol _n
Yol _s
_m
De te0
De te1
SS 2
Yol te
Yol n
Yol _s
_m
ite
ite
_
li
li
o8
o8
o8
o8
o8
o8
D_
D_
D_
t
it
l
o8
o8
o8
D_
D_
D_
i
i
i
i
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
De
De
De
Model Name Model Name Model Name
(d) Raspberry Pi4 (e) Pi4 + TPU (f) Raspberry Pi5
SS v1
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol _s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol _s
_m
o8
o8
o8
o8
D_
D_
it
it
l
l
o8
o8
D_
D_
t_l
t_l
t_l
t_l
t_l
t_l
De
De
Fig. 2: Base Energy Energy Consumption per request excluding the base energy
/ Total Energy consumption per request for different edge devices.
Raspberry Pi3, as shown in Fig. 3, the inference times for SSD and YOLO8
models improve significantly, with SSD_v1 remaining the fastest at 61 ms. In
contrast, Det_lite2 exhibits the highest inference time, taking 1,576 ms.
The results reveal that on Pi4, the SSD_v1 and SSD_lite models achieve
the fastest inference times at 209 ms and 292 ms, respectively. Conversely, the
YOLOv8 models across all versions are slower than the Det_lite0 and Lite1
models, with YOLO8_m representing the slowest at 3671 ms, as shown in Fig. 3.
The addition of the Edge TPU to Pi4 significantly reduces the inference times
of the SSD and YOLO8 models, with the SSD_v1 model achieving the lowest
inference time of 12 ms, while the Det_lite2 model has the highest at 188 ms.
Similarly, on Pi5, the SSD_v1 and SSD_lite models are the fastest, with in-
ference times of 93 ms and 127 ms, respectively. Although the Det_lite models
are slower than the SSD models, they are still faster than the YOLO8 models,
with YOLO8_m exhibiting the highest inference time of 1348 ms. The integra-
tion of the Edge TPU to Pi5 further improves the performance of the SSD and
10 D. Alqahtani et al.
SS v1
SS _v1
SS v1
De lite0
De lite1
SS e2
Yol lite
Yol _n
Yol 8_s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol 8_s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol 8_s
_m
o8
o8
o8
D_
D_
it
it
it
l
l
o8
o8
o8
D_
D_
D_
o
o
D
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_
t_
De
De
De
Model Name Model Name Model Name
(a) Raspberry Pi3 (b) Pi3 + TPU (c) Raspberry Pi4
188 1348 139
1250
125
150 1000
115
100 84
97
112
100 750 644 75
79 57
50
500 380 50 36
50 34 316
20 250 126 242 25 15 21
12 93 127 10
0 0 0
SS v1
SS _v1
SS v1
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol 8_s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol 8_s
_m
De ite0
De ite1
SS e2
Yol ite
Yol _n
Yol 8_s
_m
o8
o8
o8
D_
D_
it
it
it
l
l
o8
o8
o8
D_
D_
D_
o
o
D
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
De
De
De
Model Name Model Name Model Name
(d) Pi4 + TPU (e) Raspberry Pi5 (f) Pi5 + TPU
50 50
Inference Time (ms)
40
30 29
23 22 22
20 20 19
16
10
0
SS v1
De ite0
De ite1
SS e2
Yol te
Yol _n
Yol _s
_m
li
o8
o8
D_
it
o8
D_
t_l
t_l
t_l
De
Model Name
(g) Jetson Orin Nano
YOLO8 models, with SSD_v1 achieving the lowest inference time of 10 ms and
Det_lite2 having the highest at 139 ms.
Finally, on Jetson Orin Nano, the YOLO8_n model demonstrates the mini-
mum inference time of 16 ms, while the Det_lite and SSD models have similar
inference times within the range of 20 ms. The YOLO8_m model remains the
slowest at 50 ms, as presented in Fig. 3.
Key insights: SSD_v1 exhibits the most rapid inference times when de-
ployed across various edge devices. The incorporation of TPU, substantially
enhances the performance of the evaluated models. Conversely, YOLO8_m gen-
erally demonstrates the slowest inference times among the tested configurations.
Accuracy: This subsection presents the accuracy measurements. The mean Av-
erage Precision (mAP) on the Raspberry Pi devices varied across model sizes,
as shown in Fig. 4. While the SSD_v1 model has the lowest mAP at 19, the
YOLO8_m model achieves the highest mAP of 44 among all evaluated mod-
els. The Det_lite0, lite1, and lite2 models exhibit medium mAPs ranging from
26 to 33. Both SSD_lite and YOLO8_n have similar mAPs around 30, while
Benchmarking Object Detection Models on Edge Devices 11
YOLO8_s has a higher mAP of approximately 40. When deploying these mod-
els on the Pi devices equipped with TPU accelerators, the Det_lite and SSD
models, including all their versions, demonstrate comparable mAPs to those ob-
served on the standalone Pi devices, as illustrated in Fig. 4. However, running
the YOLO8 models on Pis with TPU accelerators resulted in a reduction in ac-
curacy, with YOLO8_n having the lowest mAP of 16. Furthermore, the mAP of
YOLO8 object detection models on Jetson Orin Nano follow a similar accuracy
pattern as on the Raspberry Pi, ranging from 31 to 44, as presented in Fig. 4. In
contrast, the SSD_v1, SSD_lite, Det_lite0, Det_lite1, and Det_lite2 models
exhibit a slight decrease in mAP compared to the Raspberry Pi and Edge TPU
results. The SSD_v1 model has the lowest mAP at 16, while the SSD_lite model
achieves 27. The Det_lite0 model has an accuracy of 23, while the Det_lite1
and Det_lite2 models perform better, with mAPs of 28 and 32, respectively.
Key insights: YOLO8_m demonstrates consistently superior accuracy com-
pared to other evaluated models across various device platforms. Conversely, the
SSD_v1 model often exhibits the lowest mean Average Precision (mAP) among
the tested models. The use of TPU accelerators on the Pi devices yields sim-
ilar accuracy levels for the Det_lite and SSD model families, but results in a
reduction in accuracy for the YOLO8 models. Jetson Orin Nano exhibits com-
parable accuracy patterns for the YOLO8 models to the other setups, but shows
a slightly lower mAP for the remaining models in comparison to the Raspberry
Pi and TPU-equipped configurations.
50 44
50 50 44
40 39 40 40 39
Accuracy (mAP)
Accuracy (mAP)
Accuracy (mAP)
33 31 33 32 32 31
30 30 30 30
30 26 30 26 25 30 28 27
23
20 19 20 19 16 20 16
10 10 10
0 0 0
SS v1
SS v1
SS v1
De ite0
De ite1
SS e2
Yol te
Yol _n
Yol _s
_m
De ite0
De ite1
SS e2
Yol te
Yol _n
Yol _s
_m
De ite0
De ite1
SS e2
Yol te
Yol _n
Yol _s
_m
li
li
li
o8
o8
o8
o8
o8
o8
D_
D_
D_
it
it
it
o8
o8
o8
D_
D_
D_
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
t_l
De
De
De
Pi3 Pi5 Pi4_TPU Orin Nano Pi3 Pi5 Pi4_TPU Orin Nano Pi3 Pi5 Pi4_TPU Orin Nano
Pi3 Pi5 Pi4_TPU Orin Nano Pi3 Pi5 Pi4_TPU Orin Nano Pi3 Pi5 Pi4_TPU Orin Nano
Energy Cons. Per Req. (mWh)
Pi3 Pi5 Pi4_TPU Orin Nano Pi3 Pi5 Pi4_TPU Orin Nano
Energy Cons. Per Req. (mWh)
Fig. 5: Energy consumption per request (excluding base energy) versus inference
time for various object detection models (A fitted linear regression line is shown,
which appears curved due to the logarithmic scale of the inference time).
has better inference times, Jetson Orin Nano demonstrates better energy effi-
ciency. Pi3 with TPU comes next, and Pi4 and Pi5 are comparable, before the
worst performance by Pi3. When running the Det_Lite and YOLO8 models, the
edge devices exhibit a linear correlation between the two metrics. Jetson Orin
Nano outperforms the other devices, while Pi3 demonstrates the poorest perfor-
mance in terms of both inference time and energy consumption. However, Pi5
shows outlier results with YOLO8, deviating from the overall regression trend
as Fig. 5(c-h) presents.
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
50 50 50
40
Accuracy (mAP) 40 40
Accuracy (mAP)
Accuracy (mAP)
30 30 30
20 20 20
10 10 10
00.0 0.1 0.2 0.3 0 0.1 0.2 0.3 0.4 0 0.1 0.2 0.3 0.4
Energy Consumption (mWh) Energy Consumption (mWh) Energy Consumption (mWh)
(a) SSD_v1 (b) SSD_lite (c) Det_lit0
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
50 50 50
40 40 40
Accuracy (mAP)
Accuracy (mAP)
Accuracy (mAP)
30 30 30
20 20 20
10 10 10
0 0.2 0.4 0.6 0 0.2 0.4 0.6 0.8 1.0 0 0.25 0.50 0.75 1.00 1.25
Energy Consumption (mWh) Energy Consumption (mWh) Energy Consumption (mWh)
(d) Det_lite1 (e) Det_lite2 (f) Yolo8_n
Accuracy (mAP)
30 30
20 20
10 10
0 0.5 1.0 1.5 2.0 2.5 00 2 4 6
Energy Consumption (mWh) Energy Consumption (mWh)
(g) Yolo8_s (h) Yolo8_m
Fig. 6: Energy consumption per request (excluding base energy) versus accuracy
for various object detection models
devices, except for Pis with TPUs, which show a significant reduction due to the
compression process required for execution on TPUs. Overall, Jetson Orin Nano
achieves the best results in terms of accuracy and energy consumption, whereas
Pi3 performs the worst as Fig. 6 displays.
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
50 50 50
40
Accuracy (mAP) 40 40
Accuracy (mAP)
Accuracy (mAP)
30 30 30
20 20 20
10 10 10
01
10 102 103 104 01
10 102 103 104 01
10 102 103 104
Inference Time (ms) Inference Time (ms) Inference Time (ms)
(a) SSD_v1 (b) SSD_lite (c) Det_lit0
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
50 50 50
40 40 40
Accuracy (mAP)
Accuracy (mAP)
Accuracy (mAP)
30 30 30
20 20 20
10 10 10
01
10 102 103 104 01
10 102 103 104 01
10 102 103 104
Inference Time (ms) Inference Time (ms) Inference Time (ms)
(d) Det_lite1 (e) Det_lite2 (f) Yolo8_n
Accuracy (mAP)
30 30
20 20
10 10
01
10 102 103 104 01
10 102 103 104
Inference Time (ms) Inference Time (ms)
(g) Yolo8_s (h) Yolo8_m
Fig. 7: Inference time versus accuracy for various object detection models
the best performance in terms of accuracy and inference time for the YOLOv8
models, while Pi3 exhibits the poorest results as shown in Fig. 7.
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
20
Accuracy (mAP)
Accuracy (mAP)
Accuracy (mAP)
30 25
15 25 20
20 15
10 15
10 10
5 5 5
0 0 0
400 600 600
500 500
)
300 400
ms
ms
ms
0.000.05 0.0 0.0 400
300 300
e(
e(
e(
Energ0.100.15 200 0.1
Energ 0.1
Energ
y Co0.2 200 y C0.2 200
Tim
Tim
Tim
y Con 0.200.25 100 0.3
ns. (m 100 ons. (m
0.3 100
s. (m 0.30 0
er
er
er
Wh) Wh)0.4 0 Wh0.4 0
Inf
Inf
Inf
)
Accuracy (mAP)
Accuracy (mAP)
30 30 30
25 25
20 20 20
15 15
10 10 10
5 5
0 0 0
1200 2000 2500
1000 1500 2000
)
)
800
ms
ms
ms
0.0 600 0.0 0.2 1000 0.0 0.2 1500
e(
e(
e(
Energ0.2 400 Energ 0.4 Energ0.4 0.6 1000
im
im
im
y Con0.4 y Con 0.6 0.8 500 y Con 0.8 1.0
200 s. (m 1.2 0500
T
T
s. (m 0.6 s. (m 1.0 0
er
er
er
Wh) 0 Wh) Wh)
Inf
Inf
Inf
(d) Det_lite1 (e) Det_lite2 (f) Yolo8_n
Pi4 Pi5 Pi3_TPU Pi4_TPU Pi4 Pi5 Pi3_TPU Pi4_TPU
Pi5_TPU Orin Nano Pi3 Pi5_TPU Orin Nano Pi3
40
Accuracy (mAP)
Accuracy (mAP)
40
30 30
20 20
10 10
0 0
6000 12500
10000
)
)
ms
ms
0.0 0.5 4000 0 1 7500
e(
e(
Energ 1.0 2000 Energ 2 3 5000
im
im
y Con 1.5 2.0 y Con 4
s. (m 5 6 02500
T
s. (m 2.5 0 T
er
er
Wh) Wh)
Inf
Inf
Fig. 8: Energy consumption per request (excluding base energy) versus inference
time versus accuracy for various object detection models
4 Related Work
This section provides an overview of the most relevant research on object de-
tection models for edge computing devices and compares our work with existing
related work as shown in Table 2. To the best of our knowledge, our work is
a unique study due to its comprehensive evaluation of various object detection
models and edge devices.
Cantero et al. [4] examine quantization levels and model architectures (SSD,
CenterNet, EfficientDet, Faster R-CNN) on the NXP i-MX8M-PLUS and Google
Coral Dev Board with EdgeTPU, using metrics like warm-up time, inference
time, and accuracy. In contrast, our study measures energy consumption and
evaluates YOLOv8 on Raspberry Pi 3, Pi 4, Pi 5, and Jetson Orin Nano. Also,
Kamath and Renuka [10] examine EfficientDet models with integer quantization
on Raspberry Pi, recommending EfficientDet0 and EfficientDet1 for resource-
constrained devices. Our study includes a broader range of models and also
measures energy consumption.
16 D. Alqahtani et al.
References
1. Balasubramaniam, A., Pasricha, S.: Object detection in autonomous vehicles: Sta-
tus and open challenges. arXiv preprint arXiv:2201.07706 (2022)
2. Baller, S.P., Jindal, A., Chadha, M., Gerndt, M.: Deepedgebench: Benchmarking
deep neural networks on edge devices. In: 2021 IEEE International Conference on
Cloud Engineering (IC2E). pp. 20–30. IEEE (2021)
3. Bulut, A., Ozdemir, F., Bostanci, Y.S., Soyturk, M.: Performance evaluation of
recent object detection models for traffic safety applications on edge. In: Proceed-
ings of the 2023 5th International Conference on Image Processing and Machine
Vision. pp. 1–6 (2023)
4. Cantero, D., Esnaola-Gonzalez, I., Miguel-Alonso, J., Jauregi, E.: Benchmarking
object detection deep learning models in embedded devices. Sensors 22(11), 4205
(2022)
5. Chen, C.W., Ruan, S.J., Lin, C.H., Hung, C.C.: Performance evaluation of edge
computing-based deep learning object detection. In: Proceedings of the 2018 VII
International Conference on Network, Communication and Computing. pp. 40–43
(2018)
6. Coral: Usb accelerator datasheet. Tech. rep., Google LLC,
https://coral.ai/docs/accelerator/datasheet/ (2019)
7. Coral: Object detection (May 2024), https://coral.ai/models/
object-detection/
8. Foundation, R.P.: About us (May 2024), https://www.raspberrypi.org/about/
9. Galliera, R., Suri, N.: Object detection at the edge: Off-the-shelf deep learning
capable devices and accelerators. Procedia Computer Science 205, 239–248 (2022)
18 D. Alqahtani et al.
10. Kamath, V., Renuka, A.: Performance analysis of the pretrained efficientdet for
real-time object detection on raspberry pi. In: 2021 International Conference on
Circuits, Controls and Communications (CCUBE). pp. 1–6. IEEE (2021)
11. Kang, P., Somtham, A.: An evaluation of modern accelerator-based edge devices
for object detection applications. Mathematics 10(22), 4299 (2022)
12. Lema, D.G., Usamentiaga, R., García, D.F.: Quantitative comparison and perfor-
mance evaluation of deep learning-based object detection models on edge comput-
ing devices. Integration 95, 102127 (2024)
13. Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Dollár, P.,
Zitnick, C.L.: Microsoft coco: Common objects in context. In: Computer Vision–
ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12,
2014, Proceedings, Part V 13. pp. 740–755. Springer (2014)
14. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: Ssd:
Single shot multibox detector. In: Computer Vision–ECCV 2016: 14th European
Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part
I 14. pp. 21–37. Springer (2016)
15. Magalhães, S.C., dos Santos, F.N., Machado, P., Moreira, A.P., Dias, J.: Bench-
marking edge computing devices for grape bunches and trunks detection using
accelerated object detection single shot multibox deep learning models. Engineer-
ing Applications of Artificial Intelligence 117, 105604 (2023)
16. Nvidia: Nvidia jetson orin (May 2024), https://www.nvidia.com/en-us/
autonomous-machines/embedded-systems/jetson-orin/
17. Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified,
real-time object detection. In: Proceedings of the IEEE conference on computer
vision and pattern recognition. pp. 779–788 (2016)
18. Tan, M., Pang, R., Le, Q.V.: Efficientdet: Scalable and efficient object detection. In:
Proceedings of the IEEE/CVF conference on computer vision and pattern recog-
nition. pp. 10781–10790 (2020)
19. Ultralytics: Home (2024), https://docs.ultralytics.com/
20. Voxel51: Fiftyone (May 2024), https://voxel51.com/fiftyone/
21. Zagitov, A., Chebotareva, E., Toschev, A., Magid, E.: Comparative analysis of
neural network models performance on low-power devices for a real-time object
detection task. Computer 48(2) (2024)