[go: up one dir, main page]

0% found this document useful (0 votes)
18 views37 pages

sensors-23-06202-v2

Download as pdf or txt
Download as pdf or txt
Download as pdf or txt
You are on page 1/ 37

sensors

Review
Augmented Reality (AR) for Surgical Robotic and Autonomous
Systems: State of the Art, Challenges, and Solutions
Jenna Seetohul 1, * , Mahmood Shafiee 1,2, * and Konstantinos Sirlantzis 3,4

1 Mechanical Engineering Group, School of Engineering, University of Kent, Canterbury CT2 7NT, UK
2 School of Mechanical Engineering Sciences, University of Surrey, Guildford GU2 7XH, UK
3 School of Engineering, Technology and Design, Canterbury Christ Church University,
Canterbury CT1 1QU, UK; konstantinos.sirlantzis@canterbury.ac.uk
4 Intelligent Interactions Group, School of Engineering, University of Kent, Canterbury CT2 7NT, UK
* Correspondence: jls56@kent.ac.uk (J.S.); m.shafiee@surrey.ac.uk (M.S.)

Abstract: Despite the substantial progress achieved in the development and integration of augmented
reality (AR) in surgical robotic and autonomous systems (RAS), the center of focus in most devices
remains on improving end-effector dexterity and precision, as well as improved access to minimally
invasive surgeries. This paper aims to provide a systematic review of different types of state-of-the-
art surgical robotic platforms while identifying areas for technological improvement. We associate
specific control features, such as haptic feedback, sensory stimuli, and human–robot collaboration,
with AR technology to perform complex surgical interventions for increased user perception of the
augmented world. Current researchers in the field have, for long, faced innumerable issues with
low accuracy in tool placement around complex trajectories, pose estimation, and difficulty in depth
perception during two-dimensional medical imaging. A number of robots described in this review,
such as Novarad and SpineAssist, are analyzed in terms of their hardware features, computer vision
systems (such as deep learning algorithms), and the clinical relevance of the literature. We attempt
to outline the shortcomings in current optimization algorithms for surgical robots (such as YOLO
and LTSM) whilst providing mitigating solutions to internal tool-to-organ collision detection and
image reconstruction. The accuracy of results in robot end-effector collisions and reduced occlusion
remain promising within the scope of our research, validating the propositions made for the surgical
Citation: Seetohul, J.; Shafiee, M.;
clearance of ever-expanding AR technology in the future.
Sirlantzis, K. Augmented Reality
(AR) for Surgical Robotic and Keywords: augmented reality (AR); machine learning (ML); navigation; planning; robotic and
Autonomous Systems: State of the autonomous systems (RAS); surgery
Art, Challenges, and Solutions.
Sensors 2023, 23, 6202. https://
doi.org/10.3390/s23136202
1. Introduction
Academic Editor: Bijan Shirinzadeh
Over the past couple of decades, significant advancements in the performance of
Received: 22 April 2023 robotic platforms have been achieved by researchers in the academic community, with
Revised: 9 June 2023
the deployment of such robots soaring amidst the COVID-19 pandemic. Studies show
Accepted: 3 July 2023
that the high probability of a resurgence in COVID-19 cases necessitates cost-effective and
Published: 6 July 2023
self-deploying telepresence robots to ensure pathogen control worldwide [1]. According
to Raje et al. [2], the market size of healthcare robots was over 9 billion in 2022, exceeding
the current fleet number more than twofold in comparison to values in 2019. Today,
Copyright: © 2023 by the authors.
robotic platforms such as the Davinci robot have significantly improved the way in which
Licensee MDPI, Basel, Switzerland. surgeons perform complex interventions, reducing the need for patient re-admission due
This article is an open access article to its minimally invasive nature. Novel surgical robots are, today, the most sought-after
distributed under the terms and approach in performing repetitive tasks in an accurate manner. Imaging technology has
conditions of the Creative Commons significantly changed the world of robotic surgery, especially when it comes to biopsies,
Attribution (CC BY) license (https:// the examination of complex vasculature for catheterization, and the visual estimation of
creativecommons.org/licenses/by/ target points for port placement. There is a great need for the image analysis of CT scans
4.0/). and X-rays for the identification of the correct position of an anatomical landmark such

Sensors 2023, 23, 6202. https://doi.org/10.3390/s23136202 https://www.mdpi.com/journal/sensors


Sensors 2023, 23, 6202 2 of 37

as a tumor or polyp. This information is at the core of most augmented reality systems,
where development starts with the reconstruction and localization of targets. Hence, the
primary role of augmented reality (AR) applications in surgery would be to visualize and
guide a user towards a desired robot configuration with the help of intelligent computer
vision algorithms.
The application of such cutting-edge robotic technologies remains diverse in various
sectors, from carrying out military or manufacturing tasks to airborne or underwater oper-
ations, due to their dexterity, ease of operation, high adaptability, and multi-functionality.
The widespread demand for AR in surgery is the impetus for our work, with a core focus
on the challenges encountered in their deployment in the existing literature as well as our
proposed solutions in counteracting these issues, emphasizing the precision of end-effector
placement and feedback from control systems. The field of surgery has seen a quantum
leap in the evolution of procedural ergonomics and the level of autonomy in robots during
an intervention. Since the first robotic-assisted surgery was successfully used to treat neuro-
logical tumors from a spin-out industrial robot called the PUMA 200 in 1985 [3], scientists
across the globe have found the need for increased precision in robotic arm positioning and
orientation to relieve surgeons of their long hours in operation theaters. From the AESOP
robotic arm built by Computer Motion for laparoscopic camera positioning in 1993 to the
DaVinci system cleared for use in 2000 for countless segmentectomies of various organs,
each platform has been improved in terms of hardware and software features, introducing
the 3D visualization of the inner anatomy to surgeons via a see-through display. From
this evolutionary hierarchy, scientists have seen the use of AR as a blessing in the surgical
setting, reducing the surgeon’s cognitive load whilst performing complex surgeries such as
cardiothoracic, colorectal, head and neck, and urological resections.
The collaboration between AR and RAS is a breakthrough in the world of minimally
invasive robotic surgeries, with the earliest publications on this principle dating back to
the 2000s, by Worn et al. [4]. More recently, in the media, the healthcare startup company
Novarad introduced an AR-based surgical navigation system called VisAR, which operates
based on virtual organ superposition with submillimeter accuracy [5]. Various other
startups, such as Proximie [6], have also emphasized the importance of AR for surgical
guidance through their extensive work on virtual “scrubbing in” and the augmented
visualization of end-effector operations. These platforms provide an incentive to surgical
robot manufacturers to integrate similar collaborative software packages into their control
systems to obviate the risk of hand tremors, improving the synergy in human–robot
arm placement and enabling telepresence via microphone communication throughout the
procedure. This type of collaboration remains in its early pilot stages, although it is of
increasing relevance in addressing the gaps in non-contact surgery during the pre- and
post-pandemic eras.
Although most advanced robots perform pre-programmed repetitive tasks with mini-
mal artificial intelligence (AI) training, existing surgical robots with supplementary visual,
olfactory, and haptic modalities prove to augment human–robot interaction and hence im-
prove overall system performance. In this paper, we evaluate the types of surgical scenarios
that involve AR technology, their corresponding navigation strategies, and the DL methods
used in their operation. We also focus on identifying the loopholes in the existing literature,
involving the levels of autonomy (LoA) in surgical RAS, accuracy in GPU performance,
experiment genericity, and clinical evaluation, amongst others. In the conventional robot
hierarchy adapted from works by Haidegger et al. [7] and Attanasio et al. [8], the level
of autonomy (LoA) framework enables researchers to adapt the control system to terrain
irregularities and exploit the force requirements for more efficient robot kinematics. Using
such robot classification, researchers across the academic community have explored the pos-
sibilities of enhancing the control systems of surgical robots with novel visual modalities,
for better output efficiency within the accepted medical standards. Lower-level robots (0–1)
are employed for assigned tasks within a defined scope, requiring a surgeon’s guidance and
pre-programming with no or limited external support, such as active constraints for user
Sensors 2023, 23, 6202 3 of 37

navigation or virtual fixtures to improve user visualization of the surrounding anatomy.


In the higher-level entities (2–5), more comprehensive systems have been developed for
varying surgical complexity, which are environment-aware and perform cognitive decision
making whilst adapting to external changes in stimuli. Such systems can provide certain
capabilities to the human–robot interface to relieve the surgeon of certain responsibilities
whilst switching from operator to robot for the duration of the task to be executed. To some
degree, it must be noted that algorithmic approaches are included in each LoA (Level 0
has a degree of tremor filtering and redundancy resolution) but mostly the higher-level
platforms are able to perform preoperative planning and devise an interventional algorithm
to allow complete control of the surgery under a surgeon’s supervision.
The ability to perform surgeries autonomously has been debated by several law courts
against Intuitive Surgical, with up to 3000 cases in 2016 [9]. This is because certain surgical
robots require a degree of human input in line with the medical equipment safety design
concept, which includes the Medical Electrical Equipment Standard (IEC 80601-2-77 [10]
(https://www.iso.org/standard/68473.html) [accessed on 20 March 2023] as well as the
IEC60601-1 [11] (https://www.iso.org/standard/65529.html) [accessed on 20 March 2023]
safety standard series [9]. The latter describes the standards that manufacturers must follow
to ensure patient (target) and surgical team safety (surgeon as main user), incorporating
risk control measures while controlling the robot in a surgery. Furthermore, the patient
safety guidance under the standards states that, owing to the vulnerability of patients,
the surgical team needs to be well-versed in hazard and risk prevention in case of an
accident. For example, any uninitiated motion or swerving away from the trajectory,
excessive speed in the motors, or faulty safeguards may pose life-threatening risks during
a pre-planned autonomous or semi-autonomous thoracotomy through 3D image overlay.
Our review does not include fully autonomous robotic platforms that can operate without
human intervention and employ sensory feedback systems for a decentralized network,
due to the high risks associated with complete robot autonomy. Instead, we classify and
identify the features within the exhaustive list of surgical robotic platforms, adapted from
works by Simaan et al. [12] and Hoeckelmann et al. [13], which are or have the potential
of providing a certain LoA to the surgeon through visualization methods or other DL
algorithmic approaches. These robots are either on the market currently as commercial
robots or proof-of-concept devices that have deployable potential in the future.

1.1. Current Knowledge of XR, AR, and VR Platforms


The umbrella term for platforms used for immersive visualization, interaction, sim-
ulation, and improved user perception of scenarios is called extended reality (XR). AR
systems have proven to be an indispensable medium for human interaction with the ex-
ternal virtual world by bridging the gap between the required task and assisting tools
through unobscured user display. According to Krevelen et al. [14], AR provides user im-
mersion by augmenting the field of view of the real world with computerized data such as
graphics and audiovisual content, as well as other sensory reinforcement methods. Several
interfaces have been used in robots, such as head-mounted displays, smart glasses such
as the HoloLens 1 and 2 [15], and handheld devices such as smartphones and overhead
projectors [16]. The HoloLens is likened to a personal computing system, designed with
an optical see-through mechanism: virtual data are projected onto a translucent screen in
the user’s field of view (FoV) while preserving the real-world setting in the background;
this enables the instant synchronization of proprioceptive stimuli, as well as complete
situational awareness. This allows the device to fit into different sectors, such as gaming,
manufacturing, and surgery, due to its high-resolution imaging, albeit with sub-optimal
spatial coherence. It is fitted with tracking sensors, pose estimation sensors, 3D coordinate
mapping sensors, environment-sensing cameras, speakers, inertial measurement units
(IMU), and holographic processing units. In other words, an AR model for surgery is a rev-
olutionary platform aiming to create and display digital information in real time, primarily
superposed over the actual organ. The three main components of this model include a
Sensors 2023, 23, 6202 4 of 37

physical object such as forceps or grippers, used as a prototype for the virtual design and
interpretation; ML algorithm-driven sensors with cameras for the visual depiction of output
images; and modeling software that processes the input signals from the cameras [17].
On the other hand, virtual reality (VR) creates a computer-generated back-end scene
for complete immersion, such that the user can experience real-world scenes in a completely
virtual environment. The supporting device system for the generation of a virtual world
consists of joysticks, controllers, or trackpads for navigation; head-tracking devices for
pose estimation; and microphones for voice recognition. VR headsets such as the Oculus
Rift (Facebook, Menlo Park, CA, USA) [18] and Oculus Quest [19] tend to blur the user’s
real-life environment and create a completely immersive virtual scenario, controlled by
a stereoscopic binocular system. The virtual scenario is then developed for the user by
projecting different 2D images onto each eye, with varying perspectives and fields of
view (FoV) between 90◦ and 210◦ and a speed of 90 frames per second, for an enhanced
immersive experience [20]. VR pose estimation in surgical settings includes the use of
clinically acquired imaging datasets, which are reconstructed in a dexterous workspace with
geometrical x, y, and z planes. This enables motion tracking using fiducial cues registered
onto specific coordinate planes that have been isolated from a reconstructed virtual scenario
and replaced in the exact positions after removing the back-end background [21].

1.2. Definition and Scope of Augmented Reality in Surgery


Since its introduction to the scientific world in the 2000s, AR in surgery has been
developing at a soaring rate, although it has been criticized by many due to the heavy
wearable devices, limited sensory input, and inefficient real-time object registration due to
tissue deformation [22]. The new era of AR in RAS has seen a leap in computer-vision-based
decision making, as stated in a paper by Nilsson et al. [23], hence proving its efficacy in
fields including, but not limited to, machinery, manufacturing, surgery, and education.
Following Halsted’s approach of training, “see one, do one, teach one”, a scientist may
observe a particular task being performed through an augmented visualization device,
practice this task several times until mastery is achieved, and eventually demonstrate
this concept to trainees [24]. To provide an accurate representation of the role of AR, we
compile and examine the definitions stated by Milgram et al. [25] and Azuma et al. [26],
who claimed that AR is defined as “the augmentation of natural feedback such as visual,
haptic and olfactory feedback to the surgeon using fiducial cues”. We decided to follow this
definition and classify the existing literature papers in our meta-analysis according to this
principle, excluding VR and AV papers, as well as the side-by-side visualization of medical
images during a surgical procedure without superposition or virtual-to-real tool alignment.
Despite the prevalence of AR technologies in several sectors, there is a significant gap
in their performance in handling cross-modalities during surgical manipulation, which
may lead to targeting errors and inaccuracies such as false negatives. This paper aims to
conduct a systematic review of different types of state-of-the-art surgical robotic platforms
while identifying areas for technological improvement. We associate specific control fea-
tures, such as haptic feedback, sensory stimuli, and human–robot collaboration, with AR
technology to perform complex surgical interventions for increased user perception of the
augmented world. Current researchers in the field have, for long, faced innumerable issues
with low accuracy in tool placement around complex trajectories, pose estimation, and
difficulty in depth perception during two-dimensional medical imaging. The plethora of
robots described in this review, such as Novarad and SpineAssist, are analyzed in terms of
their hardware features, computer vision systems (such as deep learning algorithms), and
the clinical relevance of the literature. We attempt to outline the shortcomings in current
optimization algorithms for surgical robots (such as YOLO and LTSM) whilst providing
mitigating solutions to internal tool-to-organ collision detection and image reconstruc-
tion. Our paper presents a stepping stone for researchers to explore the possibilities of
adapting AR to RAS for the navigation and control of surgical tools within a plethora of
anatomical environments.
Sensors 2023, 23, 6202 5 of 37

The organization of this paper is as follows. Section 2 presents the data collected
during the literature review, including the commercial robotic platforms, proof-of-concept
systems, their operating principles, and the corresponding AR human–robot interfaces.
Section 3 outlines the working principles of AR devices and the categories of hardware
devices used in line with the AR systems for accurate visualization. Section 4 emphasizes
the software implementation in these AR models, as well as the corresponding input
and output data obtained. Section 5 introduces the novel DL framework used for object
detection, path planning, and the data analysis of medical image datasets. Section 6 opens
the floor to a discussion about the various challenges faced by the robot platform whilst
interfacing with AR technology, such as risks of collisions, reduced collaboration, and
divergence in trajectories, as well as some solutions to combat these issues. Section 6
provides a summary of the paper while addressing future research possibilities in AR for
surgery. Finally, a concluding statement in Section 7 provides an incentive to surgeons and
researchers to elaborate and improve the given solutions in this discussion.

2. Research Background
2.1. Classification of AR–RAS Collaboration in Meta-Analysis Study
We started our search with papers obtained from several peer-reviewed databases,
such as IEEExplore, Google Scholar, SCOPUS, and PubMed, to perform a thorough initial
literature search (Figure 1). We focused on articles published from the last decade till
March 2023, due to the rapid advancement of AR technologies, which was marked by
the groundbreaking release of the Microsoft HoloLens in 2016 [27]. The key search terms
used to triage the papers from these databases were from the title and the abstract, such
as “Augmented Reality” AND “Robots” AND “Surgery” OR “Surgical Robot” OR “Sur-
gical Robotics” OR “Surgical Navigation” OR “Robot Assisted” OR “Minimally Invasive
Surgery”. Ultimately, the methodological segregation of papers was performed, dividing
them into clusters in line with the ATRA framework in [28]. They were then divided into
different groups, such as “Software Tools for Surgical Applications”, “Hardware Speci-
fications”, and “DL Algorithms” (see Figure 2), which led to a total of 425 papers from
SCOPUS and 200 papers from PubMed, excluding articles that were not relevant to our
review, duplicated, or published in non-English languages. Considering the fact that the
papers reviewed were published over the last 20 years, there was a significant gap in the
literature regarding high-level AR and AI applications for robots in surgery. The number
of papers published on surgical robots based on AR before the year 2013 was less than
20 per year, and the accrued number of papers published was lower than 500. The literature
review conducted in 2023 saw the highest increase in papers. The techniques listed above
underwent a detailed review amongst the wealth of peer-reviewed publications, whereby
the advantages and disadvantages of the available robotic systems, their AR-based control
features, and the computational algorithms were analyzed. Parameters such as 3D image
reconstruction, types of hardware features, and the potential gaps found in clinical evalua-
tion and path planning were reviewed. Some authors focused on image-guided control and
navigation using intelligent predictive systems, while others studied the orientation and
positioning of robots in an augmented view, from point cloud mesh generation to various
algorithmic approaches.
Sensors
Sensors2023,
2023,23,
23,x6202
FOR PEER REVIEW 6 of 38
6 of 37

Sensors 2023, 23, x FOR PEER REVIEW 6 of 38

Figure
Figure 1. Systematic
Figure 1.
1. Systematicreview
Systematic reviewresults
review results in
results PRISMA
inin flowchart
PRISMA
PRISMA format,
flowchart
flowchart identifying
format,
format, the
the duplicates
identifying
identifying and
the duplicates
duplicates and and
excluded
excluded papers.
papers.
excluded papers.

Figure 2.
Figure 2. Pie chart showing the distribution
distribution and
and taxonomy
taxonomy of
of retrieved
retrieved papers
papers from
from literature.
literature.
Figure 2. Pie chart showing the distribution and taxonomy of retrieved papers from literature.
2.2.
2.2. Review
Review ofof Commercial
Commercial Robots
Robots and
and Proof-of-Concept
Proof-of-Concept Systems
Systems
2.2. Review
The of Commercial Robots and Proof-of-Concept Systems visualization in surgical
The focus of our paper remains the increasing demand
focus of our paper remains the increasing demand forfor visualization in surgical
Theespecially
robotics,
robotics, focus of our
especially in paper remains
in preoperative
preoperative the increasing
scenarios,
scenarios, onto
onto the demand
the real-world
real-world for visualization
patient
patient from thein
from the surgical
object
object
detection
robotics, process
especially of landmarks
in of
preoperative interest beyond
scenarios, onto the
the visible surface
real-world and
patient
detection process of landmarks of interest beyond the visible surface and by merging pre- by merging
from the object
preoperative
detection and
operative and
process real-world
of landmarks
real-world images
images of together. A
interestAbeyond
together. higher rate of
the visible
higher rate deployment
surface and
of deployment of
of AR AR devices
bydevices
merging in pre-
in surgery is encouraged more than ever today, stemming from the expected
surgery is encouraged more than ever today, stemming from the expected increase in the in
operative and real-world images together. A higher rate of deployment ofincrease
AR in
devices
the market
surgery demand of at
is encouraged leastthan
more 18.02%
ever by the end of 2023.from
Thethe
most common areas of
market demand of at least 18.02% by thetoday,
end ofstemming
2023. The most expected
common areasincrease in the
of robotic
market demand
surgery employing of at
AR least
at 18.02%
present by the end
include of 2023. The cardiothoracic,
neurosurgery, most common areas of robotic
orthopedic,
surgery employing AR at present include neurosurgery, cardiothoracic, orthopedic,
Sensors 2023, 23, x FOR PEER REVIEW 7 of 38
Sensors 2023, 23, 6202 7 of 37

gastrointestinal, and ENT, amongst others [29]. For example, a surgeon may use the pre-
robotic imaging
operative surgery employing AR at present
from a patient’s medicalinclude neurosurgery,
database to locate a cardiothoracic,
cancerous tumororthopedic,
and pro-
gastrointestinal, and ENT, amongst others [29]. For example, a surgeon
ject this reconstruction onto the real anatomy to help them to find its exact position. may use the
For
preoperative imaging from a patient’s medical database to locate a cancerous tumor and
further reading on the types of robotic surgeries performed, readers may refer to Robotic
project this reconstruction onto the real anatomy to help them to find its exact position. For
Assisted Minimally Invasive Surgery: A Comprehensive Textbook [30]. Expanding on the
further reading on the types of robotic surgeries performed, readers may refer to Robotic
works by Barcali et al. [31], Appendix A classifies the types of commercial robots and
Assisted Minimally Invasive Surgery: A Comprehensive Textbook [30]. Expanding on
proof-of-concept systems
the works by Barcali that
et al. weAppendix
[31], concentrate on in this
A classifies paper
the typesinofterms of the robots
commercial parameters
and
studied, AR interfaces,
proof-of-concept anatomical
systems that welocation, andonCE
concentrate inmarking
this paperawarded.
in terms of the parameters
In this AR
studied, paper, we aimanatomical
interfaces, to contribute to future
location, and research
CE marking by building
awarded.a foundation on the
current Instate of the art and proof of concept in AR for surgical
this paper, we aim to contribute to future research by building robotics, whilst addressing
a foundation on the
thecurrent
followingstateresearch questions:
of the art and proof of concept in AR for surgical robotics, whilst addressing
• the following
What is the research questions:
current state-of-the-art research in integrating AR technologies with sur-
• gical
What is the current state-of-the-art research in integrating AR technologies with
robotics?
• surgical
What robotics?
are the various hardware and software components used in the development
• of AR-assisted
What are the surgical
various hardware
robots andand software
how components
are they used in the development of
intertwined?
• AR-assisted
What are somesurgical robots application
of the current and how areparadigms
they intertwined?
that have enhanced these robotic
• platforms?
What are Howsome can
of the current application paradigms that have enhanced
we solve the research gaps in previous literaturethese robotic
reviews and
platforms? How can we solve the research gaps in previous literature reviews and
promote faster performance and accuracy in image reconstruction and encourage
promote faster performance and accuracy in image reconstruction and encourage high
high LoA surgical robots with computer vision methods?
LoA surgical robots with computer vision methods?
To To
understand and elaborate on the methodologies used in AR-based robotic surger-
understand and elaborate on the methodologies used in AR-based robotic surgeries,
ies,we
wedecided
decidedtotoclassify
classifythe
thesystems
systemsininterms
termsofoftheir
theirhardware
hardwareand andsoftware
softwarefeatures
featuresasas
an an
initial
initial literature search from our meta-analysis and based on the logic relationship inin
literature search from our meta-analysis and based on the logic relationship
Figure
Figure3. This section
3. This focuses
section focusesononthe
thefeatures
featuresof
ofthe
theAR
AR interfaces in RAS
interfaces in RASthatthatcontribute
contribute
to the hardware development of the system. The papers are categorized
to the hardware development of the system. The papers are categorized in terms of in terms oftheir
their
different marker trackers and sensors, their image registration and alignment
different marker trackers and sensors, their image registration and alignment methods, and methods,
andthethetypes
types ofof displays
displays usedused
forfor visualization.
visualization.

Figure 3. Logic relationship between the different sections evaluated in the literature review, such
Figure 3. Logic
as hardware relationship
(robotic between
platforms), the different
software (machine sections evaluated
learning in theand
algorithms literature review,technolo-
calibration such as
hardware (robotic platforms), software
gies), and augmented reality headsets. (machine learning algorithms and calibration technologies),
and augmented reality headsets.
3. Hardware Components
3.1. Patient-to-Image Registration Devices
sors 2023, 23, x FOR PEER REVIEW 8 of

In surgical navigation systems, AR-based scenes require seamless interaction b


Sensors 2023, 23, 6202 8 of 37
tween the real world and the digital world, increasing the need for precise motion tracki
via marker-based or location-based mechanisms [32]. Often, pose estimation via moti
tracking enablesComponents
3. Hardware the user to perform the accurate manipulation of tools and geometrica
position end-effectors Registration
3.1. Patient-to-Image for the cutting,
Devices outlining, and extraction of anatomical landmar
such as shoulder
In surgicalblades
navigationor internal organs. scenes
systems, AR-based Location-based
require seamlesstriggers maybetween
interaction be used in co
the real world and the digital world, increasing the need
junction with, but not limited to, pose estimation sensors such as IMUs, which provifor precise motion tracking
severalviameasurements,
marker-based or location-based mechanisms
such as acceleration, [32]. Often,
magnetic posestrength,
field estimationandvia motion
orientation a
tracking enables the user to perform the accurate manipulation of tools and geometrically
gles. There is also the possibility of obtaining accurate geographical locations of speci
position end-effectors for the cutting, outlining, and extraction of anatomical landmarks
clinical
suchpersonnel
as shoulderthrough
blades or ARinternalscreens
organs.such as smartphones,
Location-based triggers may HMDs,
be usedand even sma
in con-
glasses [33]. These
junction markers
with, but provide
not limited to, posea basis for the
estimation initial
sensors alignment
such of the provide
as IMUs, which virtual world
several measurements, such as acceleration, magnetic
the real world, with respect to a generic reference frame in space towardsfield strength, and orientation angles.
the target
There is also the possibility of obtaining accurate geographical locations of specific clinical
interest.
personnel through AR screens such as smartphones, HMDs, and even smart glasses [33].
Contrary
These markers to marker-based
provide a basis AR calibration
for the systems,
initial alignment which
of the virtualuse pre-defined
world to the real tracki
markers such
world, withasrespect
QR codes to leverage
to a generic referenceobjects
frame onto
in space a real-world scene,ofmarkerless
towards the target interest. system
Contrary
tend to enable to marker-based
user-friendly AR calibration
referencing cuessystems,
to positionwhich anuse pre-defined
object in space.tracking
They opera
markers such as QR codes to leverage objects onto a real-world
by experimenting with different human skin textures, internal vessel structures, and gescene, markerless systems
tend to enable user-friendly referencing cues to position an object in space. They operate by
metrical features from
experimenting medical
with different scansskin
human of textures,
a patient [34]. vessel
internal The user can prescribe
structures, and geometri- the locati
of thecalmodel
features and
fromnavigate around
medical scans the scene
of a patient [34]. without
The user can necessarily
prescribe the disturbing
location of thethe extern
aspects of their surroundings, collating relayed data from accelerometers and visual, ha
model and navigate around the scene without necessarily disturbing the external aspects
of their
tic, and surroundings,
olfactory sensors, collating
as well relayed
as GPS datasystems.
from accelerometers
Such AR models and visual, haptic, on
depend andcompu
olfactory sensors, as well as GPS systems. Such AR models depend on computer vision
vision algorithms such as convolutional neural networks (CNN) to perceive target obje
algorithms such as convolutional neural networks (CNN) to perceive target objects without
without fiducial
fiducial markers,
markers, commonly commonly
trained usingtrained usingprogram
a software a software
calledprogram
TensorFlowcalled
API. TheTensorFlo
API. The specific
specific referencing
referencing points arepoints
passedarethrough
passedsuch through
neuralsuch neural
networks networks
in real time, suchin real tim
that the accurate positions of the user can be tested and validated
such that the accurate positions of the user can be tested and validated in further expe in further experimental
procedures (see Figure 4).
mental procedures (see Figure 4).

FigureFigure
4. The tracking process during AR alignment between patient and device.
4. The tracking process during AR alignment between patient and device.

ThereThere exist
exist a multitude of
a multitude of sensors
sensorsthat areare
that integrated into robotic
integrated platforms
into robotic for the
platforms for t
detection of precise locations in a surgical procedure, ranging from ultrasonic
detection of precise locations in a surgical procedure, ranging from ultrasonic sensors [3sensors [35],
mechanical sensors [36], and electromagnetic (EM) sensors [36] to optical tracking sen-
mechanical sensors [36], and electromagnetic (EM) sensors [36] to optical tracking senso
sors [37]. Today, the most acclaimed sensors for image-guided surgical navigation systems
[37]. Today, the most
include optical acclaimed
and EM sensorsAR
tracking sensors. fordisplay
image-guided surgical
systems require navigation
an integrated camerasystems
clude optical and EM tracking sensors. AR display systems require an integrated came
tracking mechanism, which involves the registration of the head location and directio
This process can be performed using different individual or a combination of tracki
sensors, with a wide range of applications in the clinical sector, e.g., devices such as Pola
Sensors 2023, 23, 6202 9 of 37

tracking mechanism, which involves the registration of the head location and direction.
This process can be performed using different individual or a combination of tracking
sensors, with a wide range of applications in the clinical sector, e.g., devices such as Po-
laris and MiniBird (Ascension Technology Corp., Milton, VT, USA), which attach to the
surgeon’s head for accurate simultaneous localization and mapping (SLAM). This is the
process by which a surgical robotic tool can construct and generate a collision-free map
and simultaneously identify its exact location using the map. It uses different filtering
techniques, such as Kalman filters (KF), particle filters, and graph-based SLAM. A range of
ML algorithms [38] are used in the development of a navigation structure in a discrete-time
state-space framework, such as the unscented KF, which approximates the state distribution
with a Gaussian Random Variable, where the posterior mean and covariance are captured
for propagation through the nonlinear system; the extended KF, to overcome the linearity
assumption for the next probability state; and the Monte Carlo sequential algorithms for
filtering through the estimation of trajectory samples. Other graphical SLAM techniques
adopt a node-to-node graph formulation technique, where the back end enables robot
pose correction in order to produce an independent topology of the robot, as explained
in [39]. The most common SLAM algorithm used in surgery includes the visual SLAM,
based on monocular and trinocular RGB-D camera images that use information from the
surrounding surgical environment to track the 3D landmarks through Bayesian methods,
as cited in the literature [40]. In this section, we focus on the surgery SLAM applications,
where several examples of surgical tracking systems are given, typical of robotic platforms
with AR integration: specialized robot manipulators for surgery, control algorithms for AR
visualization, and ML decision-making algorithms, amongst others.
(i) Electromagnetic Tracking Systems (EMTs)
Contrary to mechanical tracking, which depends on the positions of end-effectors
to deliver fast update rates via rotary encoders or potentiometers, electromagnetic (EM)
tracking uses a stationary source with orthogonal magnetic coils within an operating range
of 1–3 m. Nowadays, the only AR technique that limits conventional occlusion limits
is EM tracking, which operates based on field generation placed near the patient and
connected to the latter by coil wires [41]. The orientation and position of the tracking
sensors are based on the signal attenuation of the generated EM field, allowing a 360◦ range
of motion. A recent patent by Bucknor et al. [42] describes the development of an HMD
fitted with EM emitters to track the movements of the user’s head when in a 3D virtual
scenario. Such technologies can be exploited in the surgical scene to detect 3D virtual
objects projected onto a patient’s body during robotic surgery, as in Pagador et al. [43],
Liu et al. [44], and Diaz et al. [45], such that the handheld surgical tool emits EM fields
when in communication with the HMD, for the augmented visualization of organs. In
addition, the AR haptic system in [46] is calibrated to obtain precise tool coordinates within
the global positioning system (GPS). Satellite technology improvements such as real-time
kinematic (RTK) and high battery performance are required to increase the accuracy level
per centimeter, as well as ML algorithms such as the Second Thales Theorem. As in
most common tracking methods, EMTs are affected by visibility issues, occlusion, and
the complexity of the algorithms used to register the workspace coordinates with the
robot coordinate system. This form of tracking involves the inverse proportionality in
the sensor–generator distance and its ferro-magnetic sensitivity, which tends to lower the
output accuracy. This can be resolved by using EM tracking systems, such as in [47], where
the RoboTracker performs the automated positioning and orientation of the patient without
depending entirely on X-rays and conventional optics for accuracy.
(ii) Optical tracking systems (OTSs)
Optical tracking systems (OTSs) are extensively adopted in surgical navigation, the
first proof of concept used during the Second World War, when optical sighting systems
and gun detectors were a requirement for strategy planning. Zhou et al. [48] developed
an infrared-based system with fiducial markers, integrated with a Bumblebee2 stereo
Sensors 2023, 23, 6202 10 of 37

camera lens for reduced optical interference during augmented viewing. According to
Sorriento et al. [49], an OTS comprises a signal generator, a signal detector, and a control
system, which processes the signal for accurate pose estimation. The operating principle
of the optical tracking device includes determining the position of a 3D virtual object in
space by connecting at least three visible and scattered points to form a known geometric
pattern. Three non-collinear fiducial markers are required for the tracking of multiple
end-effectors in six DoFs, for facilitated pose estimation. Double markers are used to
detect positioning angles and the direction of the surgical tool tip when the values are
independent of the orientation. During the clinical procedure, fiducial cues are rigidly
registered to the surgical instruments and areas of interest to obtain location data in the
range of 40–60 Hz, which is the most common frequency of human kinesthesia [50]. The
information collated is then reconstructed by triangulation or back projection methods
using mathematical algorithms such as geometric configuration, LED activation order, and
displacement between sensors [50]. In video metric devices, pose estimation is determined
by processing sequences of images from calibrated video cameras, albeit encountering
background interference due to mechanical or optical stimuli [51]. However, IR-based
optical tracking systems can perform multifunctional tracking using up to six active wireless
tools, hence requiring lengthy computations and registration periods. This results in an
increased cost for OTSs, which also plays a pivotal role in the overall system’s cost of
manufacture and its market value.

3.2. Object Detection and AR Alignment for Robotic Surgery


Alongside the multitude of advances in other areas, such as dexterity and accurate im-
age acquisition, commercial surgical robots are currently equipped with AR technology for
the manipulation of resection tools. Their ability to visualize the patient-specific anatomy
during affected tissue extraction allows them to work within safe workspace boundaries.
While the precise mapping of medical images is unlikely due to the constant deformation of
tissue pre- and post-surgery, many research papers [52–55] are dedicated to exploring the
possibility of decoupling virtual objects and their sensory stimuli from the real world using
algorithmic approaches adapted from the DL repository. Amongst the most acclaimed
methods, projection-based AR, marker-based AR, markerless AR, and superimposition
AR are widely used in robotic platforms employed in the operation theater and remotely.
The section below provides examples of the types of AR tracking and the ways that they
facilitate robot-assisted surgery.

3.2.1. Intraoperative Planning for Surgical Robots


(i) Marker-based AR
In marker-based AR technology, the main objective remains to drive a robotic system
while performing coordinate estimation from the cameras relative to the markers. A
plethora of marker sizes and types, with Vuforia and ArUco being the most popular, are
utilized in a back-end working environment, enabling fluctuating marker information from
the robot to be registered by the AR interface. This type of AR is useful in surgery that
requires the triangulation of end-effectors to calculate their positions based on an added or
moving fiducial marker with respect to the reference point (see Figure 5 for an example of
a tumor biopsy using preoperative marker tracking using CBCT).
Sensors 2023,23,
Sensors2023, 23,6202
x FOR PEER REVIEW 11
11 of
of3738

Figure5.5.Intraoperative
Figure IntraoperativeCBCT
CBCTscan
scanshowing
showingthe
thepre-planned
pre-plannedtumor
tumorlocation
locationand
andposition.
position.

InInthethe existing
existing literature, Yavas Yavas etetal.al.[56]
[56]used
usedAR-based
AR-based neuronavigation
neuronavigation using
usingop-
tical tracking
optical trackingcameras
camerassuch suchas as LIDAR
LIDAR and and AR light detection.
detection. Using
Using markers
markerswithin withinthe the
operatingscene,
operating scene,3D 3Dpreoperative
preoperativeregistration
registrationand andsuperposition
superpositionisisperformed
performedsuccessfully
successfully
withtargeting
with targetingerrors errorsbetween
between0.5 0.5and
and3.53.5mm,
mm,with withfacilitated
facilitatedplacement
placementofofthe thefiducial
fiducial
marker.
marker.This Thisapplication
applicationhas hasbeen
beenexpanded
expandedby byauthors
authorssuch suchasasVanVanDuren
Durenetetal. al.[57]
[57]and
and
Luciano
Lucianoetet al.al.
[58], who
[58], whocreated simulations
created simulationsof wire guidance
of wire through
guidance hip and
through hip thoracic screw
and thoracic
placements
screw placements using fluoroscopic imaging simulators
using fluoroscopic embedded embedded
imaging simulators with orthogonal with cameras
orthogonal to
track virtual fiducial markers. Another widespread use of marker-based
cameras to track virtual fiducial markers. Another widespread use of marker-based AR in AR in preoperative
planning
preoperative and training
planningincludes haptic-feedback-enforced
and training robotic simulators
includes haptic-feedback-enforced roboticfor midwives
simulators
and novice obstetricians
for midwives and novicetoobstetricians
perceive thetorequired
perceiveforce during birth
the required forceto exert birth
during during to tool
exert
triangulation [59]. In cardiac
during tool triangulation [59].support
In cardiacrobotic systems,
support robotic novices
systems, andnovices
surgeons andalike can
surgeons
perform preoperative cardiac pulmonary resuscitation (CPR)
alike can perform preoperative cardiac pulmonary resuscitation (CPR) through a series of through a series of training
exercises
training on an augmented
exercises robot simulator,
on an augmented thus walking
robot simulator, thusthem
walkingthrough
themthe core steps
through to
the core
perform when a patient’s heart stops
steps to perform when a patient’s heart stops [60]. [60].
(ii) Markerless AR
(ii) Markerless AR
In its evolution from the detection of tangible markers in a surrounding environment,
In its evolution from the detection of tangible markers in a surrounding environment,
markerless AR registration requires no trigger for the detection of objects of interest in
markerless AR registration requires no trigger for the detection of objects of interest in a
a workspace. Users can extract specific areas during visualization, estimate the corre-
workspace. Users can extract specific areas during visualization, estimate the correspond-
sponding transformation from the reference to image coordinate frame, and overlay the
ing transformation
generated image above from thelandmark.
a real reference to Forimage coordinate
instance, Liu et al. frame, and overlay
[61] recounted thethe gener-
utility of
anated image above
AR-based guidance a real landmark.
system For instance,
for tongue Liu et al.
tumor removal [61] recounted
during the utility
transoral robotic of an
surgery,
AR-based
where guidance
the daVinci robotsystem for tongue
gripper tumor avoidance
used collision removal during transoral
algorithms robotic areas
to identify surgery,of
interest and adjust the area of extraction in its holographic view. In [62], severalareas
where the daVinci robot gripper used collision avoidance algorithms to identify proof-of
interest and adjust the area of extraction in its holographic
of-concept devices have been presented with low-cost digital platforms for vein location. view. In [62], several proof-of-
concept
They consistdevices have been presented
of high-intensity IR LEDs for with low-cost
virtual vesseldigital platforms
enhancement for vein
diffusion on location.
an aug-
They consist
mented HMD or of ahigh-intensity
smartphone such IR LEDs
as thefor virtual
Google vesselFinally,
Nexus. enhancement
Khuzagliyezdiffusion et al.on an
[63]
augmented HMD or a smartphone such as the Google Nexus.
described an AR-based visualization platform for the location of veins through ultrasound, Finally, Khuzagliyez et al.
[63] described an AR-based visualization platform for the location
using holographic-assisted, marker-free needle guidance for increased precision of cannu- of veins through ultra-
sound,Inusing
lation. holographic-assisted,
a commercial setting, devices marker-free needle guidance
such as AccuVein for increased precision
(https://www.accuvein.com/
of cannulation. In a commercial setting, devices
why-accuvein/ar/ (accessed on 12 February 2023)) [64] and the NextVein such as AccuVein (https://www.accu-
Vein Finder
vein.com/why-accuvein/ar/
(https://nextvein.com (accessed
(accessed on 12 February
on 12 February 2023))
2023)) [65] are [64] and the NextVein
merchandized as wearable Vein
Finder (https://nextvein.com
high-definition glasses that provide (accessedthe on 12 February
smart, real-time2023)) [65] are merchandized
3D visualization of veins andas
wearable
arteries high-definition
in separate layered glasses
views. Thisthat property
provide reduces
the smart, the real-time 3D visualization
risks of internal bleeding, pa- of
veins and arteries in separate layered views. This property
tient discomfort, and patient–doctor codependence by 45% due to the constant monitoring reduces the risks of internal
bleeding,
through patient
similar discomfort, These
IR techniques. and patient–doctor
devices all use codependence
markerless AR by using45% thedue to the
principle
Sensors 2023, 23, 6202 12 of 37

of digital laser scanning, hence improving the prospects of successful vein targeting by
3.5 times [66]. Kastner et al. [67] applied a markerless calibration technique in a HoloLens-
based robotic system, which operated using point cloud data acquired from a depth sensor.
Despite the slow processing time of the modified neural network controlling the movement
of the robot, precise localization and augmented visualization was successfully achieved,
albeit lowering the user experience. Another paper by Von Atzigen et al. [68] recounts
the possibility of navigating a bending rod through a patient’s spinal cord after pedi-
cle screw placement through CNN-assisted object detection techniques and AR-based
axial orientation.
(iii) HMD-Based AR for Surgery and Rehabilitation
In the NAVIO Surgical System (Smith & Nephew, London, UK), which is built onto the
HoloLens HMD, the surgeon performs intraoperative customized bone preparation and
confirmation of the correct cut guide sizes through the overlaying of augmented drawings
produced by surgical resection tools [69]. Moreover, a paper by Thøgersen et al. [70]
introduced the concept of relieving patients from phantom pain, mostly experienced in
amputees, after the loss of a limb or after a spinal cord injury. The real-time rendering of
two healthy limbs in an HMD enables the user to perform specific actions in a game-like
scenario while angular measurements of rotations are sent to a robotic platform from
inertial sensors. Studies showed that the two participants who used this low-cost AR
rehabilitation system for pain reduction immediately experienced lighter pain pre- and
post-session. Further studies by Rothgangel et al. [71] depict the repetitive use of AR for
analgesic outcomes on affected areas through a procedure called mirror box therapy. In
these exercises, amputees can observe a reflection of the affected limb as intact through
an avatar of similar appearance. They wear an HMD during the procedure and can
mimic movement in line with the functioning one when using wheelchairs and robotic
exoskeletons. These experiments are a turning point in the way that physiotherapists
perform rehabilitation, such that patients can perform movement actions with AR-based
software whilst supported by an exoskeletal structure.

3.2.2. Preoperative Planning for Surgical Robots


(i) Superimposition-based AR
There are a wide range of applications for AR devices in the field of surgical robotics,
which operate based on superimposition. The benefit of augmenting the surgeon’s FoV
with reconstructed medical images and computer vision interfaces is primarily the ability
to superpose useful information over real-world scenarios. This will enable an increase in
efficiency in surgical setup and clinical arrangements. Widely appreciated by the surgical
community, AR-based technologies include video-based displays that augment the sur-
geon’s FoV through video streaming [72], see-through displays that superimpose additional
virtual objects onto the surface of the target user’s direct view [73], and projection-based
displays that enable patient-specific models to be overlaid on patient skin, albeit reduc-
ing the geometrical accuracy and depth estimation. Liu et al. [74] described the use of
a superimposition-based tracking system used to set up and register a digital reference
frame of the craniofacial skeleton. The platform also consisted of an optical tracking system
and a workstation to upload real-time data, which are transferred to the HMD for visual-
ization. Pfefferle et al. [75] developed a renal biopsy system for needle tracking through
tissue of interest by superposing holographic lesion representations from relevant CT scans
(Figure 6). Moreover, Nicolau et al. [76] successfully introduced a miniature AR-based
optical tube through a patient’s abdomen to visualize the endoscopic structures, which are
not visible in direct camera view but are visible in the preoperative images. This type of 3D
visualization is the first step in the development of a fully functional AR system, whereby
the patient’s anatomy is transparent to the surgeon’s eye and important structures such
as polyps, tumors, and blood channels can be identified in preoperative planning. This
means that the control system operated by the surgeon can capture force feedback as well
as perceive the approximate depth reached from the navigating channels, the results of
Sensors 2023, 23, x FOR PEER REVIEW 13 of 38

Sensors 2023, 23, 6202 13 of 37

planning. This means that the control system operated by the surgeon can capture force
feedback as well as perceive the approximate depth reached from the navigating channels,
which cantheberesults
analyzed to create
of which can an
be interactive
analyzed toAR system
create for surgeons
an interactive ARand novices
system alike.
for surgeons and
Salah et al. [77] described
novices a similar
alike. Salah approach
et al. [77] foranavigation
described along the
similar approach forspinal cord and
navigation alongthe
the spinal
adjacent cord
vertebrae, discs,
and the and nerves
adjacent using discs,
vertebrae, the in and
situ superimposition
nerves using theofinreconstructed 3D
situ superimposition of
models over the patient’s body.
reconstructed 3D models over the patient’s body.

Figure 6. Figure 6. Superimposition-based


Superimposition-based AR tool navigation
AR tool navigation during
during right VATSright VATS segmentectomy
segmentectomy and biopsyand biopsy
using HD monitor screens. The red marker indicates the correct positioning of the stapler
using HD monitor screens. The red marker indicates the correct positioning of the stapler to proceed to proceed
with the dissection.
with the dissection.

Pessaux Pessaux et al.


et al. [78] [78] investigated
investigated the usethe ofuse
an of an AR-assisted
AR-assisted roboticrobotic
system system to perform
to per-
form accurate incisions and detect areas of interest during robotic liver segmentectomy. Liu et
accurate incisions and detect areas of interest during robotic liver segmentectomy.
Liu et al.al. [61]
[61] andand Navab
Navab et al.
et al. [61][61]
bothboth praised
praised thethe tool’s
tool’s guidance
guidance during
during robotic
robotic surgery due
surgery
due to theto ability
the ability to visualize
to visualize real-time real-time deformations
deformations with geometrical
with geometrical aids such aidsas such as fiducial
fiducial
lines in their
lines stereoscopic view. The
in their stereoscopic authors
view. in [79] used
The authors ARused
in [79] to reconstruct the external
AR to reconstruct the external
auditoryauditory
canal andcanal
the tympanic membranemembrane
and the tympanic of the middle earmiddle
of the cavity aseara cavity
3D representation
as a 3D representa-
from preoperative CT scans. ThisCT
tion from preoperative procedure
scans. This requires delicate
procedure navigation
requires delicateto prevent
navigation bleed-
to prevent
ing in thebleeding
middle ear or perforation to the ossicles, therefore requiring accurate
in the middle ear or perforation to the ossicles, therefore requiring accurate lo- localization
and an overall augmented
calization and an view.
overall Surgeons
augmented are then able
view. to detectare
Surgeons and target
then ablespecific
to detecttissue,
and target
which may be difficult
specific to ascertain
tissue, which may be ordifficult
shift constantly.
to ascertainA or
novel
shiftapproach
constantly. to Aperforming
novel approach to
cancer biopsies,
performing developed
cancer in partnership
biopsies, with KUKA
developed Robotics with
in partnership and SIEMENS,
KUKA Robotics has sur-and SIE-
faced through
MENS, thehas
MURAB
surfaced project
through[80], setting
the MURAB up a new workflow
project for magnetic
[80], setting up a new resonance
workflow for
imaging magnetic
(MRI) andresonance
ultrasoundimaging(US). Users can register the deformation of
(MRI) and ultrasound (US). Users can register the target areas
the defor-
using relative force feedback and volumetric data. Such procedures ensure
mation of the target areas using relative force feedback and volumetric data. Such proce- the precise
targetingdures
and extraction
ensure theofprecise
miniature lesionsand
targeting with precise control
extraction underlesions
of miniature the guidance of an control
with precise
AR-based navigation system (https://www.murabproject.eu/about-murab/
under the guidance of an AR-based navigation system (https://www.murabpro- (accessed on
22 Marchject.eu/about-murab/
2023)). (accessed on 22 March 2023)).
AR has also been used for
AR has also been used the location of subcutaneous
for the location veins in veins
of subcutaneous preoperative surgery,surgery,
in preoperative
as depicted in [70], where the proposed prototype has a USB
as depicted in [70], where the proposed prototype has a USB camera connected camera connected to anto an An-
Android droid
smartphone to capture live frames of the vein using infrared sensors.
smartphone to capture live frames of the vein using infrared sensors. The inherent The inherent
ability ofability
hemoglobin in the blood
of hemoglobin in theto blood
absorbtolarge
absorbvolumes of infrared
large volumes (IR) waves
of infrared (IR)triggers
waves triggers
the given the given phenomenon. The output images are then enhanced with contour[81],
phenomenon. The output images are then enhanced with contour filling filling [81],
segmented through a thresholding technique, and are then displayed on a screen for
segmented through a thresholding technique, and are then displayed on a screen for su-
superposition over a real-world anatomical landmark. Furthermore, Chen et al. [82]
perposition over a real-world anatomical landmark. Furthermore, Chen et al. [82] devised
devised a mechanism to track the location of cues on the human pelvis by superposing a
a mechanism to track the location of cues on the human pelvis by superposing a hologram
hologram of the latter over itself, using an HMD called the nVisor ST60 (NVIS Inc., Reston,
of the latter over itself, using an HMD called the nVisor ST60 (NVIS Inc., Reston, VA, USA).
VA, USA).
Another paper by Ma et al. [83] presented the use of AR systems for further preoper-
Another paper by Ma et al. [83] presented the use of AR systems for further preopera-
ative planning before pedicle screw placement in spine surgery, a procedure that involved
tive planning before pedicle screw placement in spine surgery, a procedure that involved
the use of ultrasound to generate 3D images from CT scans and then superpose them onto
the use of ultrasound to generate 3D images from CT scans and then superpose them
the areas of interest. Hajek et al. [84] used HoloLens to HoloLens communication to locate
onto the areas of interest. Hajek et al. [84] used HoloLens to HoloLens communication
to locate a bone in a user’s body; the devices were mounted on a C-arm fluoroscope and
Sensors 2023, 23, 6202 14 of 37

the patient, respectively. In studies by Elmi-Terander et al. [85], similar preplanning tech-
niques were used to direct the drill trajectories for transpedicular screw placement using a
cross-modality-based system called AR surgical navigation (ARSN). After procuring 3D
CT scans of the surrounding spinal structures, the output DICOM data are warped into a
3D reconstruction of the spine, which allows for feedback-enhanced tracking to locate the
areas of screw insertion. This is achieved by equipping the system with quadruple cameras,
which are able to record a wider field of view from different angles and, in turn, display the
3D superposed images over the estimated drilling trajectory on a monitor. Moreover, the
system proved to be reliable even during minimally invasive surgeries with percutaneous
placement of Jamshidi needle tips at areas of interest to calculate the screw entry point and
appropriate angles of insertion.
(ii) Projection-based AR
In projection-based (or spatial) AR, the areas of interest in the human body are dis-
played in a virtual world, without the use of HMDs and high-definition display screens.
Using projection mapping, the augmented model of a landmark may be overlaid and
dragged out of a screen using a tracking pen for realistic cardiac anatomy examination, as
described in [86]. The available CT datasets were used to reconstruct the cardiac vessels
and the associated separation between, for instance, pulmonary vessels of the hilum and
vena cava, and visualize the behavior of a typical heart during grafting in a transplant. This
method can be incorporated with a cross-examination using the computational model of
the Total Artificial Heart (SynCardia Systems, Tucson, AZ, USA) in its virtual form, for
accurate decision making, especially in terms of the biocompatibility of the scaffold heart
in patients of younger age. Wu et al. [87] described the use of an AR navigation system to
investigate the live deformation of surrounding tissue, nerves, and vessels via projections
of the spine onto the patient’s back, which was reinforced with reference markers to overlay
the image precisely over the patient. In other works, described in [88], the authors introduce
the use of projector-based AR platforms to control a custom needle-guided robot using
hand gestures over a preoperative ablation model projected over the area of interest. The
Leonardo project by Simoes et al. [89] presented a similar interaction framework to plan the
positioning of surgical ports by projecting the triangulation points over the patient’s torso.
(iii) HMD-based AR
According to Burström et al. [90], augmented reality surgical navigation (ARSN)
techniques have been applied in the automatic position tracking of a given instrument to
establish a real-time feedback loop of its location, leading to the enhanced identification of
the virtual bone screw target point and angulation. After conducting several experiments,
97.4% accuracy was achieved during the extrapolation of the output data coordinates. Addi-
tionally, another robotic platform designed for spine surgery is MicroOptical (MicroOptical
Corp., Westwood, MA, USA), which consists of an HMD for augmented intraoperative
fluoroscopy performed in the vicinity of the internal fractures and the spinal pedicle screw
placement. Out of the fifty case studies carried out on different patients, the operation time
was significantly reduced due to the reduced view diversion from the patient. This also
diminishes the rate of radiation absorbed by the medical professionals in the operating
theater from the fluoroscopy generator. Furthermore, Lee et al. [91] describe an alternative
to projection-based AR using a monitor screen to allow video sequence visualization in
thyroidectomy to decouple the tactile feedback stimulus from the robot feedback system,
during the resection of different anatomical structures. In a study by Agten et al. [92],
the HoloLens was used to perform augmented visualizations of needle placements and
insertions through a sawbone spine phantom made from opaque agar, as a simulation of
lumbar joint punctures. After the reconstruction of the output CT scans of the phantom, the
data were collated, loaded onto a headset, and projected onto the surgeon’s FoV as a holo-
gram for precise needle guidance during the procedures, of which 97.5% were successful.
Pratt et al. [69] conducted experiments to display and see through a patient’s inner vascu-
lature in 3D during reconstructive surgery, through a HoloLens. The device is equipped
Sensors 2023, 23, 6202 15 of 37

with hand gesture recognition, enabling any hand movements to be captured, registered,
and eventually converted into a hologram overlaying the existing lower extremities of the
human body.

4. Software Integration
From the master–slave testbed to the operating theater, AR plays a pivotal role in the
visualization of anatomical landmarks, particularly the ear, nose, and throat, as well as
gastro-intestinal areas. AR-assisted robotic surgery has facilitated the surgeon’s task in
reducing hand tremors and loss of triangulation during complicated procedures. Studies
show that the transition from invasive open surgery to indirect, perception-driven surgery
has resulted in a lower number of cases of tissue perforation, blood loss, and post-operative
trauma [93]. In contrast to open surgery, which involves the direct manipulation of tissue,
image-guided operations enable the surgeon to map medical image information, virtual or
otherwise, spatially onto real-world body structures. Usually, the virtual fixture is defined
as the superposition of augmented sensory data upon the user’s FoV from a software-
generated platform that uses fiducial markers to register the location of an anatomical
section in real time and space with respect to the user’s real-time scene. The use of
publicly available datasets obtained from cutting-edge technology, such as CT and magnetic
resonance imaging (MRI), in such scenarios enables minimal human error in data processing
and hence improved success rates of surgeries.

4.1. Patient-To-Image Registration


The preliminary steps in diagnosing the area of concern in a patient include the use
of computer guiding software to visualize inner anatomical landmarks. The loss of direct
feel of the inner anatomy, reduced depth perception due to the monocularity of cameras,
and distorted images have been addressed in novel techniques such as the segmentation
of tissue in medical scans and 3D modeling for an augmented 360-degree field of view
(FoV) [94]. In several papers by Londono et al. [95] and Pfefferle et al. [75], case studies of
kidney biopsies examine the development of AR systems for the superposition of holograms
over experimental phantoms. Studies show that preoperative CT scans from the lateral
decubitus position result in deformed tissue internally, in addition to discrepancies between
preoperative and intraoperative scans. Accurate image-guided surgery greatly depends on
the registration of preoperative medical scans with their corresponding ones within the
intraoperative anatomy. During the procedure, aligned coordinate frames are mapped onto
the output registered image. The need to compensate for the time lag during registration
means that multiple time frames are required at different regions of interest to enhance the
quality of the registered image.
Usually, the preferred choice of registration method depends on the type of robotic
environment that the surgeon is navigating, where feature-based registration attracts the
most attention within the academic community. These methods are less computationally
heavy and can be used to effectively match fiducials between preoperative and intraopera-
tive images, with primarily deformable methods of surface registration. Due to the sole use
of 2D parameters, the possibility of obtaining highly accurate 3D information is low, hence
driving the research community to establish novel sensing technologies for 3D marker
tracking. Registration methods such as point-based registration, feature-based registration,
segmentation-based registration, and fluoroscopy-based registration are widely used in the
image processing of medical scans. The geometric transformations of deformable objects
are computed using fiducial markers, which act as positioning cues and can be analyzed
for fiducial localization errors (FLEs). In cases where images have varying gray levels, DL
algorithms are able to segregate different features using parameters such as the sum of
squared or absolute differences (SSD), correlation coefficients, and mean squared difference
(MSD). For real-time X-ray image processing, a contrasting material, such as barium or
iodine, is used to create more subtle contrast differences for clinicians to analyze. The
process of 2D to 3D image registration involves the alignment of matching preoperative
Sensors 2023, 23, x FOR PEER REVIEW 16 of 38

mean squared difference (MSD). For real-time X-ray image processing, a contrasting ma-
Sensors 2023, 23, 6202 terial, such as barium or iodine, is used to create more subtle contrast differences for16 ofcli-
37

nicians to analyze. The process of 2D to 3D image registration involves the alignment of


matching preoperative and intraoperative features, which can be reconstructed in AR and
superposed over a features,
and intraoperative live fluoroscopic
which canimage with respectintoAR
be reconstructed reference points in over
and superposed the image
a live
fluoroscopic
sequence image
(Figure 7).with respect to reference points in the image sequence (Figure 7).

Figure 7. CT scans of the lung with its corresponding 3D reconstruction and marker localization,
Figure 7. CT scans of the lung with its corresponding 3D reconstruction and marker localization,
used by surgeons to locate tumors as indicated by the red marker.
used by surgeons to locate tumors as indicated by the red marker.
4.2. Camera Calibration for Optimal Alignment
4.2. Camera Calibration for Optimal Alignment
Automatic camera calibration and corresponding image alignment in intraoperative
Automatic
ultrasound camera
is used calibration
to determine and corresponding
internal image alignment
structural characteristics such asinthe intraoperative
focal length
ultrasound
and surfaceisanatomy
used to determine
of differentinternal
organs.structural
Analysis, characteristics
visualization, andsuchpre-planning
as the focal length
using
and surface
registered anatomy
medical of different
images enable the organs. Analysis,ofvisualization,
development patient-specific and pre-planning
models using
of the relevant
registered
anatomy. The medical imagesinenable
researchers the development
[96] created of patient-specific
a cross-modality AR model to models of the
correct the rele-
shifts in
vant anatomy.
positioning Thelesion
using researchers
holograms,in [96] created during
generated a cross-modality
a CT image AR model to correct
reconstruction the
process.
shifts
A US in positioning
transducer usingtwo-dimensional
obtains lesion holograms,scans generated
from during
the siteaofCT image and
interest reconstruction
is merged
process.
with A US tracking
magnetic transducer obtains
data two-dimensional
to produce a 3D resultant scans
scanfrom thewith
in line site aofCNN
interest and is
algorithm.
merged
This with magnetic
alleviates tracking
the probability of data
false to produceappearing
negatives a 3D resultant
in the scan in line
dataset, with a when
especially CNN
mapping magnetically
algorithm. This alleviates tracked ultrasoundof
the probability scans
falseonto non-rigidly
negatives registered
appearing in the3Ddataset,
scans for the
espe-
detection
cially when of mismatches in deformation.
mapping magnetically Furthermore,
tracked ultrasoundthis method
scans is also used registered
onto non-rigidly for needle
guidance,
3D scans forasthe
mentioned
detectionin of[75], to predict
mismatches trans-operative
in deformation. pathways during
Furthermore, navigation,
this method is also
as well as detecting areas of extraction for lesions on Unity3D via the
used for needle guidance, as mentioned in [75], to predict trans-operative pathways dur- collision avoidance
system.
ing The object-to-image
navigation, registration
as well as detecting areasisofoptimized
extractionby forplacing
lesions markers,
on Unity3D sufficiently far
via the col-
apart avoidance
lision in a non-linear
system.configuration, such that
The object-to-image their combined
registration centerby
is optimized coincides with the
placing markers,
projection
sufficientlyoffar
theapart
targetinin the workspace.
a non-linear configuration, such that their combined center coin-
cides with the projection of the target in the workspace.
4.3. 3D Visualization using Direct Volume Rendering
The next steps in creating an AR model include image processing techniques such
as direct volume rendering, which are used to remove outliers and delineators from raw
DICOM data. A method proposed by Calhoun et al. [97] involves voxel contrast adjustment
Sensors 2023, 23, 6202 17 of 37

and multimodal volume registration of the voxels in the CT images by replacing their
existing density with a specific color and enhancing their contrast through thresholding,
performed by a transfer function. Manual intensity thresholding removes all low-intensity
artefacts and background noise from the image, ready for rigid attachment to an organ in
virtuality. A transparency function is applied to filter out extreme contrasts in anatomical
or pathological 3D landmarks and any blob-like contours detected can be used in the initial
registration of CT scans under techniques such as topological structural analysis. The
deformation properties of the organs are modeled using software such as Osirix 12.0, 3D
Slicer 5.2.2, or VR-Render IRCAD2010, and the high contrast applied to output images
makes structures such as tumors, bones, and aneurism-prone vessels more visible to the
naked eye.

4.4. Surface Rendering after Segmentation of Pre-Processed Data


Surface rendering techniques in [98] depict the conversion of anatomical structures
into a mesh for delineation and segmentation. Tamadazte et al. [99] used the epipolar
geometry principle to acquire images from the left and right stereovision cameras. The
authors then used a point correspondence approach to resample and build a 3D triangular
mesh from local data points in its neighborhood. The current techniques utilized in
AR are developed using a software program called Unity3D and require patient-specific
polygons such as triangles for rapid processing. Furthermore, the anatomical scenes
detected using US transducers may be reconstructed using multi-view stereo (MVS), which
analyzes pieces of tissue extracted from an area, remeshes them by warping the raw
DICOM data, and displays them with appropriate textures using speeded up robust
feature (SURF) methods [100]. In most cases, segmentation may cause the loss of essential
information in the original volume data. Therefore, in the quest to improve the quality
of segmented images, Pandey et al. [101] introduced a faster and more robust system for
US to CT registration using shadow peak (SP) bone registration. In another study by
Hacihaliloglu et al. [102], similar bone surface segmentation techniques have been used to
determine the phase symmetry of bone fractures.

4.5. Path Computational Framework for Navigation and Planning


In studies by El-Hariri et al. [103] and Hussain et al. [79], the use of tracking mech-
anisms for marker-based biopsy guidance has been widely commended and applied in
surgery, such as that of the middle ear and the kidneys. Fiducial cues are registered to dif-
ferent locations on the patient’s body, using the robust surface matching of sphere markers
with the standard model, alongside laparoscopic video streams. Image-to-patient registra-
tion is performed by comparing the acquired live images to the available patient-to-image
datasets, which is a crucial operation to eliminate errors during automatic correction, as
explained by Wittman et al. [104]. Leeming et al. [105] used proximity queries to detect
internal changes in anatomy during the manipulation of a continuum robot for surgery
around a bone cavity. A covariance tree is used in this case, as a live modeling algorithm,
to maintain an explicit safety margin between the walls of an anatomical landmark during
the maneuvering of surgical tools. For cases of minimally invasive surgery, precautionary
measures such as CO2 inflation of the patient’s body and highlighting target locations with
contrasting colors (for example, with ICG) facilitate the surgeon’s task, especially when
performing cross-modality interventions with AR systems such as headsets. A study by
Zhang et al. [106] explained the tracking mechanisms used in US procedures for intraop-
erative use. The probe was equipped with a HoloLens-tracked reference frame, which
contained multiple reflective spheres on an organ. In terms of biopsy needle tracking,
Pratt et al. [81] introduced the concept of registered stylus guidance in line with a simu-
lated 3D replica reconstructed from CT images of the torso. During preoperative surgical
navigation, a calibrated probe is used to collect data from internal organs to send to the 3D
Slicer software over OpenIGTLink, whilst combining tracked data from the input instru-
ments. The stylus tip is calibrated about a pivot and can be moved to various positions in
Sensors 2023, 23, 6202 18 of 37

the anatomical plane while tracking it over the probe reference frame using an iterative
closest point (ICP)-based detection algorithm. Jiang et al. [106] proved that the projector
view for puncture surgery also improves the efficiency of perception-based navigation,
using superimposed markers to align the needle tip to a magenta circle. The researchers in
the above study generated an accurate AR positioning method using DL techniques such
as the Newton–Gauss method and Lie algebra to produce an optimized projection matrix.
Any projection is performed towards the target location of the body, hence reducing the
probability of parallax errors, as shown by Wu et al. [107].

5. Applications of Computer Vision in Surgical Robot Operation (DL-Based)


With the groundbreaking development of artificial intelligence (AI) in assistive surgi-
cal robots, the healthcare sector today has seen a booming increase in the data collected and
stored in databases, such as in the NHS. During the COVID-19 pandemic, technology has
lessened the burden of healthcare workers behind the scenes, minimizing the need to sort,
collect, and store data manually, as well as cutting down costs in decision-making tasks.
Training datasets for early symptom recognition, estimating patient mortality rates, and
abnormality detection in specific tissue images have enabled researchers to obviate error-
prone concepts during robot training and prepare novices for unexpected fallacies [108].
The surgical community has recognized the pivotal role that AR integration in DL-based
robotics plays, including increasing the transparency of the patient, higher accuracy, less
bleeding, and shorter recovery times. The possibility of reducing the exposure to harm-
ful radiation and pathogens has also proved beneficial for overall surgical efficiency in
clinics, especially in the post-pandemic world. Despite the multitude of benefits that AR
presents, there are still a number of issues that have been identified, as in Table 1, such
as incorrect interposition and mapping between real and virtual worlds, the inaccurate
visualization of organs of interest due to difficulties in estimating their positions, and a
lack of correspondence between the real tissue and the virtual tissue. The projections of
AR-based reconstructions may be inaccurate at times due to various real-time factors, such
as the indefinite structures of internal organs and boundaries, fluctuations in vital signs,
and subtle human body movements such as aspiration and blood pressure [109].

Table 1. Technical bottlenecks in the field of AR according to Roger’s theory.

Technical Bottlenecks Description


Compatibility with social practices Wearable devices such as Google Glass may create privacy issues.
AR is easy to learn by novice surgeons and can increase the
Complexity (user-friendliness or learning)
learning curve.
Modern DL algorithms such as deep transfer learning and supervised
Lack of accuracy in alignment and unsupervised learning are used to tackle the issues in real-to-virtual
world mapping. Lighting conditions can be adjusted for better alignment.
Easily deployed but may be expensive to test in several regions
Trialability to general public
simultaneously.

The introduction of the DL-based optimization of surgical robot performance, accord-


ing to Govers et al. [110], enables intelligent task planning and operation, in contrast to
manual robotics, which only applies pre-defined output reflexes. These intelligent robots
are environment-aware and can perform perception-based obstacle-aided navigation, for
the shortest displacement decision making within restricted passageways. According
to Conti et al. [111], physical robots that have embedded lasers, IR cameras, and ultra-
wideband radios can be trained using DL algorithms to track human–robot interactions in
augmented environments. Zhang et al. [112] described the use of AR to control physical
interactions, achieve sensor-based navigation, and perform complex trajectory planning
using DL methods under changing external stimuli. The following section provides an
Sensors 2023, 23, 6202 19 of 37

overview of the DL algorithms used to increase the efficiency of robot performance and
end-effector positioning accuracy in proof-of-concept robotic platforms.

5.1. Medical Image Registration


Recently, papers by Garon et al. [113] and Alhaija et al. [114] have described the
implementation of DL algorithms such as CNNs to allow marker-based image registration
within given parameters (See Appendix B). An AR-modified neural network is proposed
for efficient object detection and point cloud extraction in line with the ComplexYOLO
architecture. Another paper by Qi et al. [115] proposes a different neural network, known
as the PointNet network, for semantic segmentation as well as 3D object localization within
raw point cloud data. Estrada et al. [33] depict an array of deep neural network architectures
to train large datasets without the need for feature engineering. DL methods such as region-
based CNNs (RCNN) [116], you only look once (YOLO) [117], and single-shot detectors
(SSD) [118] have been applied in several works pertaining to surgical image registration.
Extending from this concept, the popular SLAM algorithm, classified as feature-based
operations and direct operations, can be used in the localization of anatomical defects.
Feature-based methods focus on the principal image locators or features, whilst the direct
method utilizes the data from each pixel in the image to determine the parameters of the
target image posture. The studies by Klein et al. [119] and Mur-Artal et al. [120] describe
the use of monocular feature-based tracking using a real-time pose estimation system called
Parallel Tracking and Mapping (PTAM), as well as an alternative called ORB-SLAM. These
algorithms reduce the batch operation period and create a large coordinate system within
keyframe constraints for more accurate pose estimation.

5.2. Increased Optimization of Robot Orientation Using Motion Planning and Camera Projection
The position and orientation of surgical robots are determined by the linkage arrange-
ments and their relative degrees of freedom. In each workspace, the configuration of the
manipulator is specified for each joint to allow inference of the position of any variable.
According to Adhami et al. [121], the concept of AR can be applied to determine the DoF of
a manipulator according to its configuration space using a systematic method of positioning
surgical robots with high n values to optimize their performance. Recent experiments con-
ducted by Gonzalez-Barbosa et al. [122] and Yildiz et al. [123] depicted the use of optimal
camera placement for the wider angular coverage of a specific workspace using a camera
projection model, two-step strategies for robotically assisted minimally invasive surgery
(RAMIS), and deep learning algorithms such as Wireless Video Sensor Networks. Similarly,
Gadre et al. [124] utilized the Microsoft HoloLens as an interface for the visualization of a
target curve for a real Baxter robot. Furthermore, studies by Fotouhi et al. [125] made use
of Kinect sensors in their experimental setup to register images of their robot from multiple
angles, which were used to determine the accuracy in AR alignment. The use of a DT of the
KUKA robot in motion enabled the surgeon to estimate the correct position and orientation
during an operation, via a reflective AR mirror. The accuracy parameters depended highly
on the precision taken in reconstructing medical images, with a 33.7% success rate.

5.3. Collision Detection during Surgical End-Effector Motion


Once the topology of the robot is achieved, the orientation profile is checked for the
collision-free volume (CFV) using a swept volume visualization process, as described
in [126]. A sequence of control coordinates is selected on the contour of the output profile to
specify the pose of the robot arm at each control point. The use of an AR interface enables
efficient CFV mapping and collision detection among the registered virtual models. In
recent papers, such as [127], self-collision detection checks are performed using V-COLLIDE,
where the robot links are converted to STL format. The end-effector is only considered
collision-free when the swept control points are within the CFV range, which is visualized
by the user via projection-based AR devices. Determination of the CFV range is particularly
essential for precise port placement in robotic-assisted laparoscopic surgery [128], where
Sensors 2023, 23, 6202 20 of 37

collision avoidance allows for maximum port access and the visualization of areas of
interest. To reduce the number of cuboids in each reconstructed mesh before a collision
detection procedure starts, several algorithms, such as the tight-fitting oriented bounding
boxes (OBB) and axis-aligned bounding boxes (AABB) algorithms, are applied. They are
used to calculate the shortest colliding distance in convex polyhedral collision models [129].
Zhang et al. [130] described the qualitative results of tissue reconstruction from the surface
meshes of point clouds to the anatomical margin of interest. The experiment proved that
the fast collision method used on the OBBs after automatic cube tessellation achieved a
feedback rate of approximately 1 kHz, hence able to provide unparalleled control during
robotic surgery. Coste-Maniere et al. [131] describe the possibility of AR-based collision
detection and the increased accuracy of virtual tool placement within flesh, ribs, and
target locations. The use of a heatmap superposed over a patient’s body has also been
explored in works by Weede et al. [132], to calculate the goodness value of the plane. The
authors in [133] discuss the evaluation and calibration of such robotic systems, which
showed a relatively high degree of accuracy, albeit with a few hindrances in terms of virtual
EE alignment with the lectern interaction tools due to limited DoF. These methods are,
however, highly successful in contour tracing and profiling to produce a virtual smooth and
collision-free workspace along the output curve. A list of existing techniques in collision
detection is tabulated below for readers to compare the most efficient learning methods
used to attain accurate trajectory planning (Table 2).

Table 2. Types of collision detection techniques and corresponding learning methods used during
path planning.

Author(s) Collision Avoidance Technique Learning Method Accuracy


Zero robot programming for
vision-based human–robot
Wise-ShopFloor framework is used to
Wang et al. [134] interactions, linking two Kinect N/A
determine initial and final pose.
sensors for retrieval of robot pose in
3D from a robot mesh model.
Fast path planning using virtual
Human tracking using unscented
potential fields, representing Lower avoidance time
Du et al. [135] Kalman filter, for mean and variance
obstacles and targets, as well as (>689.41 Hz).
determination of a set of sigma points.
Kinect sensors.
Preliminary filtering of mesh models
to reduce the number of cuboids in Use of oriented bounding boxes (OBBs)
Frame rates of 17.5 k OBBs
experiment. Virtual fixtures known as and filtering algorithms: Separating
Hongzhong using a bit width of 20,
active constraints used in generating Axis Test and Sweep and Prune. Use of
et al. [136] update rate of 25 Hz
resistive force. Automatic cube field-programmable gate arrays to
compared to 1 kHz.
tessellation used for 3D point design a faster GPU system.
detection and collision avoidance.
OPML motion planning using Learning-based Fastron algorithm used 100-times faster collision
Das et al. [137] standard geometric collision checkers to generate robot motion in complex detection than C-space
such as proxy collision detectors. obstacle-prone surroundings. modeling.
Rapidly exploring random graph (RRG)
Concentric tube robot teleoperation
algorithm aids roadmap construction in Tip error between 0.18 mm
Torres et al. [138] using automatic, collision avoidance
maximum reachable insertion and 0.21 mm of tip width.
roadmaps.
workspace.
Multicopter collision avoidance by
redirecting a drone onto a planned Use of the probabilistic RRT algorithm
Killian et al. [139] Speed of up to 6 m/s.
path; connects random nodes within for collision detection.
a search space on a virtual line.
Sensors 2023, 23, 6202 21 of 37

5.4. Reconfiguration and Workspace Visualization of Surgical Robots


Most malleable robots in surgical settings require accurate port placement and end-
effector positioning, adapting to the desired user requirements within a specific workspace
reconfiguration. Each revolute joint of the robot can be aligned and positioned using
augmented visual cues, hence guiding the user towards the required robot topology. This
method is gaining popularity in surgery due to the ease of motion tracking and calibration,
with demonstrated accuracy of up to ±2 mm. In works by Ranne et al. [140], an AR-
assisted robotic system with OptiTrack sensors is implemented for the smooth generation
of a virtual end-effector, which is placed in its maximum reachable space. The computation
of workspace configurations can be performed individually, which generates a virtual cue
in the user’s FoV. Previously, a VR platform developed by Lipton et al. [141], called Baxtor’s
Homonculus, introduced an intermediate virtual scenario for the mapping of the robot
reference frame to the user’s, decoupling the sensory stimulus from the translations of
end-effectors. However, it has been observed that the mapping of the robot’s reference
frame onto that of the user may be problematic, due to inaccuracies in the alignment of
the end-effector in virtual-to-real scenarios. With the aim of increasing the precision of
such mapping, Bolano et al. [142] used point cloud extraction to predict robot–end-effector
collision during the swept volume visualization and orientation profiling of the robot. The
algorithms used for the mapping of the P3-5 end-effectors of a robot arm are explained
in detail in [143], emphasizing the accuracy of inference and virtual feedback from the
HoloLens with respect to the origin. The user is given real-time feedback on the alignment
error between the current and desired position based on the mesh model generation of
a link and translation around the reference end-effector. Other applications of AR in the
orientation planning of end-effectors are described by Gao et al. [144], who investigate the
optimum inclination angles of a robot linkage whilst following a visual cue with respect
to a particular path. Human–robot interaction is smoother at optimum angles as display–
control misalignment can be reduced and precise port placement can be selected during an
operation without risking patient safety.

5.5. Increased Haptic Feedback for Virtual Scene Guidance


The applications of the daVinci research kit have been reported to be broad in the
academic community, ranging from collaborative research in RAMIS to independent sur-
geon manipulation using a stereoscopic system [145]. The development of an ROS interface
has been the stepping stone in initiating a novel motion planning framework in line with
haptic feedback. This relies mainly on MoveIt and the Fast Collision Library (FCL), which
are currently used to upload a specific mesh object in simulation and check for collisions
in the panned PSM environment. According to Zhang et al. [145], the simulated PSMs
produce deflections, which are fed through as input and produce sensing feedback at the
main manipulator system. The direction of velocity for each PSM end-effector, normal to
the surface, v, is instead utilized and described mathematically in the equations below:
when the surfaces U and V approach a point coordinate such that ∆ (U,V) converges to
a zero value, the surface normal becomes collinear and this feature may be expanded in
order to create a spherical proxy region (SPR) at the end-effector, which is the target area
for interactions.
In the existing literature, a multitude of haptic sensors have been proposed to enable
high-speed performance in data pre-processing and rendering. The stereotaxy phenomena
in surgical scenarios such as gamma-knife surgeries [146] create an illusion of registered
and reproduced 3D haptic feedback data in the form of a sensory stimulus. In a paper
by Srinivasan et al. [147], the effect of visually locating markers was investigated, which
allowed a correlation to be made between the perception feedback of the cues and the
actual haptic feedback obtained from an object. The textures of specific objects rendered in
a virtual world were studied by Basdogan et al. [148], to allow for the creation of tangential
frictional force-generated textured surfaces in line with a technique called bump mapping.
Furthermore, a review published by Latimer et al. [149] described the behavior of several
Sensors 2023, 23, 6202 22 of 37

polygonal rigid bodies during haptic interaction, as well as the challenges of their forward
collisions on surrounding forces. More recently, research by Costa et al. [150] generated a
simulated environment for anatomical tissue using the long elements method to estimate
object deformation in a gravitational field. Such theories can be widely exploited in the
world of AR-assisted surgery to better understand the behavior of end-effectors within the
bodily vessels both in real time and virtually.
In a surgical setting, works by Okamura et al. [151] and Westebring-Van der Putten
et al. [152] have been commended for their research into various haptic devices, the types of
interaction control, as well as the intelligent proxy-based algorithms used to assess deflec-
tions or collisions in a proxy workspace. Additionally, Wurdemann et al. [153] presented
a novel wearable device that could provide accurate haptic feedback, and Li et al. [154]
adapted this design to apply the pseudo-haptic feedback (PHF) technique for hardware-
free experimentation using visual cues. A field of surgery that requires considerable
haptic implementation is plastic surgery, which requires the overlaying of virtual images
for surgical guidance, such as in Tsai et al. [155], who used a haptic device to deflate a
protruding zygoma and for implant positioning during facial contouring. In works by
Schendel et al. [156], 3D visualization was used for the surgical planning and manipulation
of patient skin models for a cleft lip repair surgery, in accordance with the output from
haptic devices. The application of AR for cranio-maxillofacial complex fracture reduction
has been explored by Olsson et al. [157], whereby the patient’s bone mesh models are
generated and an immersive experience is created using software such as Unity3D 2023.1.2
for accurate end-effector guidance in educational training.

5.6. Improved Communication and Patient Safety


Robotic surgery has improved the way in which surgeons gain access to the difficult
internal anatomy, bringing significant advancements in transmission latency from the
first transatlantic robotic-assisted laparoscopic cholecystectomy performed in the late 20th
century [158]. With the advent of 5G mobile communication technology, the field of AR
in surgery has been revolutionized, with an increased ability to perform MIS via fiber
optic technology, for example, at a cheaper cost and more widespread throughout clinics
worldwide. The use of smartphone applications for teleconferencing has been widely
recommended by surgeons operating on robotic platforms, for instant access to web-based
resources as well as near-ubiquitous peer–doctor communication. This is especially useful
in the post-pandemic world, where the immediate advice of off-site staff is required, such as
in cases where a visual review would be beneficial for injury assessment. The development
of smart wearable devices such as HMD-based systems allows the user to obtain reliable
audio-visual data with minimal latency during an intervention due to the fast transfer of
data through 5G networks [159]. These devices also help surgeons to perform simulations
and adopt extended reality scenarios when mentoring novices and remote colleagues
during various surgical procedures, by using models of surgical specimens and case studies.
Recent studies have shown that any latency experienced during robotic surgery, remote
or otherwise, increases the risks of bleeding and mistargeted tool placement, which may
lead to complications. Moreover, 5G technology enables the relaying of haptic feedback to
the surgeon in real time, through gyroscopic all-motion cognition as well as tactile sensors.
This provides a sense of real touch for the determination of the depth, precision, texture,
and contours of tissue and organs. AR technology can facilitate the learning and simulation
of tissue resistance techniques, such as determining the weight and force required to insert
and remove a needle. There are also possibilities to incorporate DL algorithms within the
learning database to adapt the system for performance analysis with supervised memory
during simulation-based training and assessment.
The link between communication and patient safety has been highlighted in works
such as [8,12], where the different types of surgical hazards, such as trajectory misjudgment
and diversion, electrical faults, and time lags, can pose serious risks for the patient. Hence,
AR displays play a pivotal role in overcoming the effect of the lag between the remote oper-
Sensors 2023, 23, 6202 23 of 37

ator and the robot platform, providing instant visual feedback to the user. As established
by a clinical trial [160] performed at IRCCS San Raffaele Hospital in Milan, the efficiency of
remote proctoring in guiding the implantation of medical devices is significantly increased.
A combination of AR visors, 5G telecommunication, and multi-access edge computing
(MEC) enables the surgeon to access live medical imaging and a holographic model of
the human heart directly from the operating theater, through a low-latency 5G network.
The paper by Richter et al. [161] explains the use of a stereoscopic AR predictive display
(SARPD) to display the preoperative displacements within anatomical margins, eliminating
the risk of overshoot and oscillations in navigation. The use of an extended Kalman filtering
(EKF) framework enabled visual alignment between predicted and actual movements using
kinematic calibrations. Ye et al. [162] described an experimental setup that facilitated error-
free hand–eye coordination during end-effector placement, whilst successfully rendering
augmented objects such as slave–tool systems and geometric fiducial cues. According to
Takahashi et al. [163], the pre-transmission compression of surgical images requires an
acceptable level of delay before the irreparable loss of anatomical data. Generally, a delay
time of up to 140 milliseconds can establish sufficient connectivity for minimum data loss
and image compression. Despite the advancements in runtime for robotic platforms such
as the DaVinci, achieving up to 36 fps on AR displays, we emphasize the need for further
study on the accuracy of measurements as well as the rendering pipeline to reduce the
cognitive load during tool manipulation.

5.7. Digital Twins (DT) to Guide End-Effectors


In the post-pandemic world, medical DT provide the integrated and virtual visual-
ization of patient data and hence create a user-friendly software platform for surgeons to
access complex information such as physical, physiological, and cognitive characteristics.
DT can play a pivotal role in remote patient monitoring through advanced diagnostic
tools, whilst incorporating AR for precision medicine, for a more patient-centric method of
treatment. The core purpose of DT in AR-based robotic surgery is personalized medicine
using patient-specific modeling from deep learning databases to accurately determine the
cause and treatment of a disease. The table in [164] summarizes the various applications of
DT in the surgical field from 2011 to the present, with some notable systems cited, such
as the Philips Heart Navigator Tool, which combines CT scans from different angles of an
organ and generates a real-time 3D model for the accurate positioning of surgical tools and
faster preoperative planning. The need for the data-driven control of dosage effects as well
as device responses before treatment has risen, being important to predict the behavior of
patients after heart disease management, such as in works by Niederer et al. [165], where
mechanical models were used to investigate the effects of cardiac resynchronization therapy
(CRT). The use of DT to treat cardiovascular diseases through semi-active modeling of the
heart, with real-time blood flow and head vibration, facilitated the localization of stenosis
in a modeled human face. In other surgical uses, the company Sim & Cure employed
3D rotational angiography to generate an interactive model of an aneurysm, to direct the
tools towards the ideal implant coordinates [166]. The implementation of a cloud GPU,
computer vision, and ML technology enables the augmented visualization of anatomical
landmarks and blood circulation, through a DT model. The efficiency of surgical perfor-
mance is increased through combined simulation and AR platforms, hence significantly
improving the training graph via an increasing optimization gradient. Other DT surgical
models include post-operative bone structure modeling in 3D from CT scans, which enables
accurate rotation and imaging system orientation through the compensation of the subtalar
joint axis, as explained in Hernigou et al. [167].
Owing to its high success rates in industrial robot integration, the authors in [168]
have established a proof-of-concept AR-integrated system for surgical interventions based
on ROS and Unity3D. The work was based on the lightweight KUKA robot, which can
be manipulated using the TMFlow software in parallel with pre-defined programs. The
robot was controlled using Python, where scripts such as MqttOmronMovePublisher package
Sensors 2023, 23, 6202 24 of 37

the joint angles and publish them on the /omron/command MQTT topic, which is then
received by the /omron/move/command at the ROS side after conversion and JSON
deserialization. This information is then received by MqttOmronMoveSubscriber, which
attaches these joint angles to the corresponding robot arm and enables the DT controller to
achieve the desired configuration. In this way, the OMRON robot can perform multiple
movements whilst visualizing them on the ghost robot in the background. In another study,
a similar virtual-to-real mapping technique was used to simulate an abdominal surgery,
estimating the virtual-to-real 6-DoF robot’s alignment within an AR environment [124].
The integration of reflective AR mirrors enables a simultaneous view of the scene from
various angles, whilst images are captured by a camera sensor on the HMD to be analyzed
for alignment accuracy. The Euclidean distance between the reference frame of the camera
center is mapped to that of each joint by colliding the cursor with the AR workspace.
The errors achieved between these joints were compared using the fixed, reflective, and
single-view displays, with a misalignment error of 16.5 +/− 11.0 mm, which was lower
than when no reflective mirrors were used.
Nowadays, most DT technologies enable AR integration in order to support model ad-
justment based on user feedback, immersion, and intuitiveness. For example, the ARAILIS
prototype in [169] provides calibrated digital software for AR mapping and image seg-
mentation, via a SLAM algorithm for object detection. The ROS2 middleware enables
communication through a modular architecture, allowing for safety and privacy encryp-
tion. Furthermore, the output from the calibration process increases the precision of the
ORBSLAM algorithm for the supervised learning of the real-to-virtual world coordinate
mapping. The human-in-the-loop collaboration with robotic DTs is a crucial requirement
for dynamic modeling and data annotation—for example, to locate tumors, detect misalign-
ments, and transpose 3D models. This enables a multi-user system to be set up for other
medical bodies to refer to as a constant knowledge database, through a human–machine
collaborative intelligent platform.

6. Discussion
Despite the prevalent breakthroughs in AR for robotic surgery, there are several pitfalls
that have been reported by users of such technology, which need to be addressed. Several
papers have been reviewed in the existing literature, addressing the core gaps in the field
and potential improvements in the efficiency of robotic platforms. These papers were
classified in terms of hardware, software, and DL applications, with a total of 170 papers.
We realized that several papers lacked a focus on accuracy in feedback systems, alignment
during interactions, registration techniques, and patient safety for robots of varying LoA.
We also identified difficulties in sight diversion during hand–eye coordination for a surgeon,
which means that rather than switching between real and virtual scenes, an integrated
lectern or platform can be used for the experiments instead [169]. Employing virtual
monitors through the HoloLens reduces the discomfort of constant view diversion and
allows greater situational awareness, as in the case of Yang et al. [170], who utilized color-
coded margins to optimize the resection margins. In the case of occlusion, it is suggested
that the field of view for the operating surgeon and the spatial relationship of individual
landmarks be optimized so that only one plane is visible at a time. This would significantly
reduce the clinician’s cognitive load during a surgery, as well as improving the depth
perception of internal organs and reducing visual clutter and the latency of the entire
surgical system. Any visual clutter produced by excessive data in a surgeon’s FoV may
risk the safe placement of the robot end-effectors around the sensitive anatomical walls. A
transparent AR model with varying window levels enables the successful navigation of the
tools from the skin layers to the bone.
The integration of AR in surgery requires the precise calibration of the end-effectors to
localize the coordinates of the objects of interest within the workspace. Fiducial markers
are predominantly used in object localization on medical scans, especially in the pre-
planning of convoluted navigation procedures. Techniques such as the rigid and non-rigid
Sensors 2023, 23, 6202 25 of 37

registration of such markers, using image deformation methods to move tissue, enable the
accurate reconstruction of unstructured robotic environments, allowing for the generation
of a historical trajectory map. In the image registration stage, the accurate transferring of
matching features from preoperative to intraoperative images requires virtual potential
fields to reference the locations of areas of interest, such as tumors or blood vessels. The use
of manual registration techniques before a fully autonomous navigation procedure allows
the elimination, to an extent, of any errors that may be caused by misleading situational
awareness, as in the two-way DaVinci robot registration. The integration of DL methods
at this stage of validation and testing helps the system to learn the correct position of the
defect within an exhaustive database, hence ensuring higher repeatability.
Despite the plethora of advantages that the development of CNN databases for AR
navigation provides, our studies showed that the feasibility of many procedures is hindered
by human errors, such as the movement of the patient, the rapid deformation of the
tissue, such as the lung, and the instability of clinical equipment, as well as high levels of
background noise in captured medical images. It is therefore of the utmost importance
that researchers bridge the gap in alleviating these challenges by employing more efficient
deep learning algorithms, such as image deformation methods, to parametrize the no-
risk zones within an anatomical space. Registration algorithms such as the head and
hat algorithm and the iterative closest point (ICP) algorithm facilitate the extraction and
geometric transformation of specific 2D or 3D surfaces on deformable multimodal images,
based on neural network architectures (CNNs, RNNs, and LTSMs).
The ability to detect collision-free zones for accurate end-effector placement requires
complex calculations of workspace reconfigurations, whilst plotting the coordinates of
the output curve within the safe margin of interest. The authors in [115,129] have largely
inspired the extraction of point clouds to create mesh models of the volume swept, hence
enabling the user to visualize the exact maximum reachable space within which the robot
end-effector can reach. There is a need for the further development of such algorithms to
enable a correlation between real-time deflated images from dynamic organs such as the
heart and the lung, with respect to their marker counterparts on inflated medical images. In
these cases, the integration of a digital twin (DT) can become useful, to reflect and visualize
the robot motion whilst performing the operation. The ghost robot allows the user to view
the final configuration of the robot in RoboDK in a pre-planning stage, which enables
corrections to be made, if necessary, both in the simulated environment and via the MQTT
LAN bridge, which connects Unity3D to the collaborative robot.
The legal and ethical aspects of AR in surgery have been debated in several courts,
due to the skepticism that comes with performing operations from a distance or even cross-
territory. The process of the clinical evaluation and validation of AR-based surgical robots
remains in its early stages of development, and experiments performed on phantoms such
as 3D-printed organs followed by cadavers and patients are claimed to have the highest
rates of accuracy [171]. However, due to the lack of evidence of the increase in surgeon
comfort or clinical performance in such validation techniques, further research focused
on larger patient datasets, higher precision in 3D reconstruction, and depth perception
may enhance the outcomes of AR-ready clinical evaluation. Despite the ongoing criticism,
the current success of AR in surgical education may also encourage further research into
faster and more accurate robot performance [8]. Although in its early stages due to a
lack of objective metrics to assess its impact, this application can further improve the
performance of various Level 1 and Level 2 robots (see LoA in Section 1) under the patient
safety guidance and ethical approval [12]. We also noticed an increase in FDA clearance
under the CE marking of devices utilizing AR in surgery, which is a promising aspect
towards AR-assisted robot deployment in hospitals.

7. Conclusions
This paper provided a general overview of various surgical robotic platforms available
on the market and in the existing literature, with an emphasis on their system architectures,
Sensors 2023, 23, 6202 26 of 37

software platforms, and learning methods. A total of 170 papers were selected and analyzed
in a systematic manner for the purpose of identifying the relevant literature that described
the types of AR technologies used in surgery. AR remains a promising tool in facilitating
the surgeon’s task, from docking the station to port placement and end-effector guidance.
To counteract the difficulties experienced by manual operation, AR visualization helps
surgeons to perform interventions efficiently through HMDs, spatial projectors, and marker-
based or markerless interfaces. This review focused mainly on the plethora of AR interfaces
used in surgery, focusing on three main aspects: “hardware”, “software”, and “application”.
The roadblocks towards achieving optimum AR integration were addressed and a wide
range of solutions was presented to increase the efficiency of existing robots. The ability to
eliminate visual clutter and occlusion within the surgeon’s FoV opens the door to novel
augmented models with different layers and windows, which can be chosen according to
the degree of importance. In areas such as thoracic surgery, gynecology, and plastic surgery,
where the haptic feedback system provides an indication of the type of force required for
an intervention, we found that the use of AR integration with force feedback sensors, as
in the DaVinci master–slave console, increases the sensory stimulus of the surgeon, with
a direct correlation between the fiducial cues and the real-time feedback from the object
of interest.
Owing to its popularity amongst surgeons, AR is widely commended by the research
community for surgeries ranging from tumor detection to vein location through fiducial
markers, despite its restrictions in terms of spatial awareness and occlusion. To reduce
the risks of bleeding in conditions where the surgeon is required to coordinate hand–eye
movements, we introduce the novel concept of reflective AR and DT technologies, which
are in their pilot stages. The level of accuracy in areas such as suturing, knot tying, and
pick-and-place operations has significantly increased as compared to manual operations,
which inspires further research in this sector. At the time of this literature review, to the best
of our knowledge, there exists a limited pool of specialized papers in the field of AR for
surgical robots containing a detailed rundown of novel AR platforms with DL algorithmic
approaches. Our paper aims to identify the research gaps in areas such as hardware
components, software specifications, and the application of DL in various surgical stages.
We believe that we have laid the foundation for the future of AR for surgery, which will
not only be useful for researchers but also surgeons and novices who wish to learn about
AR-assisted surgical interventions with accurate tool placement, without limiting the reader
to previous conventional trends in the sector.

Author Contributions: Conceptualization, J.S. and M.S.; methodology, J.S. and M.S.; software,
J.S.; formal analysis, J.S. and M.S.; investigation, J.S.; data curation, J.S.; writing—original draft
preparation, J.S.; writing—review and editing, J.S. and M.S.; supervision, M.S. and K.S.; project
administration, J.S.; funding acquisition, M.S. All authors have read and agreed to the published
version of the manuscript.
Funding: This research was funded by the UK Engineering and Physical Sciences Research Council
(EPSRC), grant number 2620688.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Acknowledgments: The authors would like to thank the team at St Bartholomew’s Hospital,
especially Steven Stamenkovic, for his expert advice and guidance during this project.
Conflicts of Interest: The authors declare no conflict of interest.
Sensors 2023, 23, 6202 27 of 37

Appendix A. State of the Art and Proof of Concept in AR-Based Surgical Robots from Existing Literature

Author/ Name of Type of Surgical


Parameters Studied AR Interface Operating Principle CE Marking
Company Device AR Display Specialization
CT-scan-based image Graphical user The system is fixed to
Transpedicular screw
reconstruction, path interface for the spine, attached to
placement
SpineAssist [172] planning of screw fluoroscopy guidance Marker-based a frame triangulated Yes (2011)
(orthopedic)
placement, and using fiducial by percutaneously
Brain surgery
Mazor Robotics Inc., needle tracking. markers. placed guidewires.
Caesarea, Israel
Ten-times faster
3D reconstruction of
Hologram generation software processing Thoracolumbar
spine with selection
Renaissance [173] for localization of Superposition-based for target localization screw placement Yes (2011)
of desired vertebral
screw placement. due to DL (orthopedic)
segments.
algorithms.
Image reconstruction, Robotic arm with Transpedicular screw
3D intraoperative
Zimmer Biomet, path planning of floor-flexible base, placement
ROSA Spine [174] planning software for Superposition-based Yes (2015)
Warsaw, Indiana screw placement, and which can readjust its (orthopedic)
robotic arm control.
needle tracking. orientation. Brain surgery
Matching
3D intraoperative preoperative and
Image reconstruction,
planning software for intraoperative General spine and
MazorX [175] 3D volumetric assay Superposition-based Yes (2017)
robotic arm control fluoroscopy to brain surgery
of the surgical field.
and execution. reconstruct inner
anatomy.
MedRobotics, Can navigate around
Raynham, MA, USA paths at 180 degrees Transoral robotic
Intraoperative Built-in AR software to reach deeper areas surgery (TORS),
Flex Robotic visualization to give with magnified HD of interest in the transoral laser
Superposition-based Yes (2014)
System [176] surgeons a clear view for viewing of body by a steering microsurgery (TLM),
of the area of interest. anatomy. instrument, i.e., and Flex®
joystick. Use of two procedures
working channels.
Sensors 2023, 23, 6202 28 of 37

Author/ Name of Type of Surgical


Parameters Studied AR Interface Operating Principle CE Marking
Company Device AR Display Specialization
Hands-free voice
recognition for
Instrument tracking Reconstructs patient
facilitated robot
and navigation imaging data into 3D
Novarad® , Pasig, control. Voice User
VisAR [5] guidance, holograms Superposition-based Neurosurgery Yes (May 2022)
Philippines Interface (VUI).
submillimeter superimposed onto
Automatic data
accuracy. patient.
uploading to the
system.
Instrument tracking Use of smart glasses Overlays 3D CT-based knee
and 3D navigation to deliver an reconstructed models ligament balance and
Medacta, Castel San Superposition/marker-
NextAR [177] guidance, immersive adapted to the other hip, shoulder, Yes (2021)
Pietro, Switzerland based
submillimeter experience to patient’s anatomy and joint arthroplasty
accuracy. surgeons. and biomechanics. interventions.
MRI-based
Image-guided robotic
image-guided AR-based immersive
interventions inside
IMRIS Inc., navigation, force environment for
an MRI, with sensory
Winnipeg, MB, NeuroARM [178] feedback from recreation of haptic, Marker-based Brain surgery Yes (2016)
stimulus from
Canada controllers for tumor olfactory, and touch
workstation to guide
localization and stimuli.
the end-effector.
resection.
Denavit–Hartenberg
derivations of Use of head tracking
Jacobian, servo HoloLens-based HoloLens for camera
Ma et al., Chinese 6-DoF robotic stereo
control, and head tracking using HMD calibration and
University of flexible endoscope Marker-based Cardiothoracic No
tracking for wider for image-guided visualization of tool
Hong Kong (RSFE) [179]
angle view, user endoscopic tracking. placement of flexible
evaluation, task load endoscope
comparison.
Sensors 2023, 23, 6202 29 of 37

Author/ Name of Type of Surgical


Parameters Studied AR Interface Operating Principle CE Marking
Company Device AR Display Specialization
User evaluation,
camera-to-joint
Digital twin with
reference frame
HMD-based robotic ghost robot for
Euclidean distance
Fotouhi et al., John KUKA robot-based arm guidance and mapping of
compared for no AR, Marker-based Cardiothoracic No
Hopkins University reflective AR [125] positioning using virtual-to-real robot
reflective mirror AR,
reflective mirrors. linkages from a
and single-view AR,
reference point.
joint error
calculation.
AR-based HMD used
to visualize the
Distance motion of surgical tip
Stereo-view capture
computation for in an image-guided Custom laparoscopic
Forte et al., Max of medical images
Robotic dry-lab lym- Euclidean arm procedure. Image box trainer
Planck Institute for acquired by robot Marker-based No
phadenectomy [180] measurements, user processing of CT containing a piece of
Intelligent Systems and HD
evaluation of AR scans to locate pixels simulated tissue
visualization.
alignment accuracy. of virtual marker
placed in virtual
scene.
Point cloud
generation for
localization of AR-based
Augmented reality
markers, system Overlay of point experimental setup
Qian et al., John assistance for Superposition/rigid
evaluation using clouds on test for guiding of a General surgery No
Hopkins University minimally invasive marker-based
accuracy parameters anatomy. surgical tool to a
surgery [181]
such as frame rate, defect in anatomy.
peg transfer
experiment.
Sensors 2023, 23, 6202 30 of 37

Appendix B. Types of Neural Networks Used in Image Registration for AR Reconstruction in Surgery

Authors Model Performance Metrics Purpose Accuracy Optimization Algorithm Equipment


Bending parameters such as Perspective-n-point
Markerless navigation and
Stereo neural networks axial displacement, algorithm and random Head-mounted AR device
Von Atzigen et al. [80] localization of pedicles of 67.26% to 76.51%
(adapted from YOLO) reorientation, bending time, sample consensus (HoloLens) with C++
screw heads.
frame rate. (RANSAC), SLAM.
SurgeonAssistNet Evaluating the online 7.4× fewer model
Parameters of the GRU cell
composed of performance of the parameters, achieved 10.2×
and dense layer, model size, 5.2× decrease in CPU Optical see-through
Doughty et al. [182] EfficientNet-Lite-B0 for HoloLens during virtual faster FLOPS, and used 3×
inference time, accuracy, inference time. head-mounted displays
feature extraction and gated augmentation of anatomical less time for inference with
precision, and recall.
recurrent unit RNN landmarks. respect to SV-RCNet.
Intersection over union
Semantic segmentation of CNN with encoder–decoder
CNN-based architectures (IoU), Euclidean distance
intraoperative proctectomy, IoU = 0.894 (σ = 0.076) structure for real-time In vivo robot-assisted
such as UNet, ResNet, between points of interest,
Tanzi et al. [118] for 3D reconstruction of compared to 0.339 image segmentation and radical prostatectomy using
MobileNet for semantic geodesic distance, number
virtual models to preserve (σ = 0.195). training of a dataset in DaVinci surgical console
segmentation of data of iterations per second
nerves of the prostate. Keras and TensorFlow.
(it/s).
Image registration Immersed boundary
Use of an artificial neural
frequency, latency between methods (FEM, MJED,
Adapted UNet architecture network to learn and
data acquisition, input Mean target registration Multiplicative Jacobian
Brunet et al. [183] for simulation of predict mesh deformation RGB-D cameras
displacements, stochastic error = 2.9 mm, 100× faster. Energy Decomposition) for
preoperative organs in human anatomical
gradients, target discretization of non-linear
boundaries.
registration error (TRE). material on mesh.
For autonomous robotic
DC-Net with images in the
ultrasound using Semantic segmentation of
propagation direction feed
Visual deep learning deep-learning-based vessel scans for organ Final model Dice score of Philips L15-i07 probe driven
through, binary
Marahrens et al. [184] algorithm such as UNet, control, for better kinematic deformation analysis using 0.887 as compared to 0.982 by US machine, dVRK
classification task,
DC-Net sensing and orientation of a dVRK and Philips L15-7io in [179]. software
IMU-fused kinematics for
the US probe with respect to probe.
trajectory comparison.
the organ surface.
Sensors 2023, 23, 6202 31 of 37

References
1. Chen, B.; Marvin, S.; While, A. Containing COVID-19 in China: AI and the robotic restructuring of future cities. Dialogues Hum.
Geogr. 2020, 10, 238–241. [CrossRef]
2. Raje, S.; Reddy, N.; Jerbi, H.; Randhawa, P.; Tsaramirsis, G.; Shrivas, N.V.; Pavlopoulou, A.; Stojmenović, M.; Piromalis, D.
Applications of Healthcare Robots in Combating the COVID-19 Pandemic. Appl. Bionics Biomech. 2021, 2021, 7099510. [CrossRef]
[PubMed]
3. Leal Ghezzi, T.; Campos Corleta, O. 30 years of robotic surgery. World J. Surg. 2016, 40, 2550–2557. [CrossRef] [PubMed]
4. Wörn, H.; Mühling, J. Computer- and robot-based operation theatre of the future in cranio-facial surgery. Int. Congr. Ser. 2001,
1230, 753–759. [CrossRef]
5. VisAR: Augmented Reality Surgical Navigation. Available online: https://www.novarad.net/visar (accessed on 6 March 2022).
6. Proximie: Saving Lives by Sharing the World’s Best Clinical Practice. Available online: https://www.proximie.com/ (accessed
on 6 March 2022).
7. Haidegger, T. Autonomy for Surgical Robots: Concepts and Paradigms. IEEE Trans. Med. Robot. Bionics 2019, 1, 65–76. [CrossRef]
8. Attanasio, A.; Scaglioni, B.; De Momi, E.; Fiorini, P.; Valdastri, P. Autonomy in surgical robotics. Annu. Rev. Control. Robot. Auton.
Syst. 2021, 4, 651–679. [CrossRef]
9. Ryu, J.; Joo, H.; Woo, J. The safety design concept for surgical robot employing degree of autonomy. In Proceedings of the 2017
17th International Conference on Control, Automation and Systems (ICCAS), Jeju, Republic of Korea, 18–21 October 2017; IEEE:
Piscataway, NJ, USA, 2017; pp. 1918–1921.
10. IEC 80601-2-77:2019. Medical Electrical Equipment—Part 2-77: Particular Requirements for the Basic Safety and Essential
Performance of Robotically Assisted Surgical Equipment. Available online: https://www.iso.org/standard/68473.html (accessed
on 20 March 2023).
11. IEC 60601-1-11:2015. Medical Electrical Equipment—Part 1-11: General Requirements for Basic Safety and Essential
Performance—Collateral Standard: Requirements for Medical Electrical Equipment and Medical Electrical Systems Used in the
Home Healthcare Environment. Available online: https://www.iso.org/standard/65529.html (accessed on 20 March 2023).
12. Simaan, N.; Yasin, R.M.; Wang, L. Medical technologies and challenges of robot-assisted minimally invasive intervention and
diagnostics. Annu. Rev. Control Robot. Auton. Syst. 2018, 1, 465–490. [CrossRef]
13. Hoeckelmann, M.; Rudas, I.J.; Fiorini, P.; Kirchner, F.; Haidegger, T. Current capabilities development potential in surgical robotics.
Int. J. Adv. Robot. Syst. 2015, 12, 61. [CrossRef]
14. Van Krevelen, D.; Poelman, R. Augmented Reality: Technologies, Applications, and Limitations; Departement of Computer Sciences,
Vrije University Amsterdam: Amsterdam, The Netherlands, 2007.
15. Microsoft. HoloLens 2; Microsoft: Redmond, WA, USA, 2019. Available online: https://www.micro-soft.com/en-us/hololens
(accessed on 10 March 2022).
16. Peugnet, F.; Dubois, P.; Rouland, J.F. Virtual reality versus conventional training in retinal photocoagulation: A first clinical
assessment. Comput. Aided Surg. 1998, 3, 20–26. [CrossRef]
17. Khor, W.S.; Baker, B.; Amin, K.; Chan, A.; Patel, K.; Wong, J. Augmented and virtual reality in surgery-the digital surgical
environment: Applications, limitations and legal pitfalls. Ann. Transl. Med. 2016, 4, 454. [CrossRef]
18. Oculus.com. Oculus Rift S: PC-Powered VR Gaming Headset|Oculus. 2022. Available online: https://www.oculus.com/rift-s/
?locale=en_GB (accessed on 6 April 2022).
19. MetaQuest. Available online: https://www.oculus.com/experiences/quest/?locale=en_GB (accessed on 10 March 2022).
20. Limmer, M.; Forster, J.; Baudach, D.; Schüle, F.; Schweiger, R.; Lensch, H.P.A. Robust Deep-Learning-Based Road-Prediction
for Augmented Reality Navigation Systems at Night. In Proceedings of the IEEE 19th International Conference on Intelligent
Transportation Systems, Rio de Janeiro, Brazil, 1–4 November 2016; pp. 1888–1895.
21. Chen, C.; Zhu, H.; Li, M.; You, S. A Review of Visual-Inertial Simultaneous Localization and Mapping from Filtering-Based and
Optimization-Based Perspectives. Robotics 2018, 7, 45. [CrossRef]
22. Venkatesan, M.; Mohan, H.; Ryan, J.R.; Schürch, C.M.; Nolan, G.P.; Frakes, D.H.; Coskun, A.F. Virtual and augmented reality for
biomedical applications. Cell Rep. Med. 2021, 2, 100348. [CrossRef] [PubMed]
23. Nilsson, N.J. The Quest for Artificial Intelligence: A History of Ideas and Achievements; Cambridge University Press: Cambridge, UK,
2009. [CrossRef]
24. Kerr, B.; O’Leary, J. The training of the surgeon: Dr. Halsted’s greatest legacy. Am. Surg. 1999, 65, 1101–1102. [CrossRef] [PubMed]
25. Milgram, P.; Takemura, H.; Utsumi, A.; Kishino, F. Augmented Reality: A class of displays on the reality-virtuality continuum. In
Proceedings of the Photonics for Industrial Applications, Boston, MA, USA, 31 October–4 November 1994.
26. Azuma, R.T. A survey of augmented reality. Presence Teleoperators Virtual Environ. 1997, 6, 355–385. [CrossRef]
27. Kress, B.C.; Cummings, W.J. Optical architecture of HoloLens mixed reality headset. In Digital Optical Technologies; SPIE:
Bellingham, WA, USA, 2017; Volume 10335, pp. 124–133.
28. Kendoul, F. Towards a Unified Framework for UAS Autonomy and Technology Readiness Assessment (ATRA). In Autonomous
Control Systems and Vehicles; Nonami, K., Kartidjo, M., Yoon, K.J., Budiyono, A., Eds.; Springer: Tokyo, Japan, 2013; pp. 55–71.
29. Pott, P.P.; Scharf, H.P.; Schwarz, M.L. Today’s state of the art in surgical robotics. Comput. Aided Surg. 2005, 10, 101–132. [PubMed]
30. Tsuda, S.; Kudsi, O.Y. Robotic Assisted Minimally Invasive Surgery; Springer: Cham, Switzerland, 2018.
Sensors 2023, 23, 6202 32 of 37

31. Barcali, E.; Iadanza, E.; Manetti, L.; Francia, P.; Nardi, C.; Bocchi, L. Augmented Reality in Surgery: A Scoping Review. Appl. Sci.
2022, 12, 6890. [CrossRef]
32. Brito, P.Q.; Stoyanova, J. Marker versus markerless augmented reality. Which has more impact on users? Int. J. Hum. Comput.
Interact. 2018, 34, 819–833. [CrossRef]
33. Estrada, J.; Paheding, S.; Yang, X.; Niyaz, Q. Deep-Learning- Incorporated Augmented Reality Application for Engineering Lab
Training. Appl. Sci. 2022, 12, 5159. [CrossRef]
34. Rothberg, J.M.; Ralston, T.S.; Rothberg, A.G.; Martin, J.; Zahorian, J.S.; Alie, S.A.; Sanchez, N.J.; Chen, K.; Chen, C.; Thiele, K.; et al.
Ultrasound-on-chip platform for medical imaging, analysis, and collective intelligence. Proc. Natl. Acad. Sci. USA 2021,
118, e2019339118. [CrossRef]
35. Alam, M.S.; Gunawan, T.; Morshidi, M.; Olanrewaju, R. Pose estimation algorithm for mobile augmented reality based on inertial
sensor fusion. Int. J. Electr. Comput. Eng. 2022, 12, 3620–3631. [CrossRef]
36. Attivissimo, F.; Lanzolla, A.M.L.; Carlone, S.; Larizza, P.; Brunetti, G. A novel electromagnetic tracking system for surgery
navigation. Comput. Assist. Surg. 2018, 23, 42–52. [CrossRef]
37. Lee, D.; Yu, H.W.; Kim, S.; Yoon, J.; Lee, K.; Chai, Y.J.; Choi, Y.J.; Koong, H.-J.; Lee, K.E.; Cho, H.S.; et al. Vision-based tracking
system for augmented reality to localize recurrent laryngeal nerve during robotic thyroid surgery. Sci. Rep. 2020, 10, 8437.
[CrossRef]
38. Scaradozzi, D.; Zingaretti, S.; Ferrari, A.J.S.C. Simultaneous localization and mapping (SLAM) robotics techniques: A possible
application in surgery. Shanghai Chest 2018, 2, 5. [CrossRef]
39. Konolige, K.; Bowman, J.; Chen, J.D.; Mihelich, P.; Calonder, M.; Lepetit, V.; Fua, P. View-based maps. Int. J. Robot. Res. 2010, 29,
941–957. [CrossRef]
40. Cheein, F.A.; Lopez, N.; Soria, C.M.; di Sciascio, F.A.; Lobo Pereira, F.; Carelli, R. SLAM algorithm applied to robotics assistance
for navigation in unknown environments. J. Neuroeng. Rehabil. 2010, 7, 10. [CrossRef]
41. Geist, E.; Shimada, K. Position error reduction in a mechanical tracking linkage for arthroscopic hip surgery. Int. J. Comput. Assist.
Radiol. Surg. 2011, 6, 693–698. [CrossRef]
42. Bucknor, B.; Lopez, C.; Woods, M.J.; Aly, H.; Palmer, J.W.; Rynk, E.F. Electromagnetic Tracking with Augmented Reality Systems.
U.S. Patent and Trademark Office, U.S. Patent No. US10948721B2, 16 March 2021. Available online: https://patents.google.com/
patent/US10948721B2/en (accessed on 23 March 2023).
43. Pagador, J.B.; Sánchez, L.F.; Sánchez, J.A.; Bustos, P.; Moreno, J.; Sánchez-Margallo, F.M. Augmented reality haptic (ARH):
An approach of electromagnetic tracking in minimally invasive surgery. Int. J. Comput. Assist. Radiol. Surg. 2011, 6, 257–263.
[CrossRef]
44. Liu, S.; Feng, Y. Real-time fast-moving object tracking in severely degraded videos captured by unmanned aerial vehicle. Int. J.
Adv. Robot. Syst. 2018, 15, 1–10. [CrossRef]
45. Diaz, R.; Yoon, J.; Chen, R.; Quinones-Hinojosa, A.; Wharen, R.; Komotar, R. Real-time video-streaming to surgical loupe mounted
head-up display for navigated meningioma resection. Turk. Neurosurg. 2017, 28, 682–688. [CrossRef]
46. Zhu, S.; Morin, L.; Pressigout, M.; Moreau, G.; Servières, M. Video/GIS registration system based on skyline matching method.
In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013;
pp. 3632–3636. [CrossRef]
47. Amarillo, A.; Oñativia, J.; Sanchez, E. RoboTracker: Collaborative robotic assistant device with electromechanical patient tracking
for spinal surgery. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems, Madrid,
Spain, 1–5 October 2018; pp. 1312–1317. [CrossRef]
48. Zhou, Z.; Wu, B.; Duan, J.; Zhang, X.; Zhang, N.; Liang, Z. Optical surgical instrument tracking system based on the principle of
stereo vision. J. Biomed. Opt. 2017, 22, 65005. [CrossRef] [PubMed]
49. Sorriento, A.; Porfido, M.B.; Mazzoleni, S.; Calvosa, G.; Tenucci, M.; Ciuti, G.; Dario, P. Optical and Electromagnetic Tracking
Systems for Biomedical Applications: A Critical Review on Potentialities and Limitations. IEEE Rev. Biomed. Eng. 2020, 13,
212–232. [CrossRef]
50. Sirokai, B.; Kiss, M.; Kovács, L.; Benyó, B.I.; Benyó, Z.; Haidegger, T. Best Practices in Electromagnetic Tracking System Assessment.
2012. Available online: https://repozitorium.omikk.bme.hu/bitstream/handle/10890/4783/137019.pdf?sequence=1 (accessed
on 22 March 2023).
51. Pfister, S.T. Algorithms for Mobile Robot Localization and Mapping, Incorporating Detailed Noise Modeling and Multi-Scale
Feature Extraction. Ph.D. Thesis, California Institute of Technology, Pasadena, CA, USA, 2006. [CrossRef]
52. Komorowski, J.; Rokita, P. Camera Pose Estimation from Sequence of Calibrated Images. arXiv 2018, arXiv:1809.11066.
53. Ghasemi, Y.; Jeong, H.; Choi, S.H.; Park, K.B.; Lee, J.Y. Deep learning-based object detection in augmented reality: A systematic
review. Comput. Ind. 2022, 139, 103661. [CrossRef]
54. Lee, T.; Jung, C.; Lee, K.; Seo, S. A study on recognizing multi-real world object and estimating 3D position in augmented reality.
J. Supercomput. 2022, 78, 7509–7528. [CrossRef]
55. Portalés, C.; Gimeno, J.; Salvador, A.; García-Fadrique, A.; Casas-Yrurzum, S. Mixed Reality Annotation of Robotic-Assisted
Surgery videos with real-time tracking and stereo matching. Comput. Graph. 2023, 110, 125–140. [CrossRef]
56. Yavas, G.; Caliskan, K.E.; Cagli, M.S. Three-dimensional-printed marker-based augmented reality neuronavigation: A new
neuronavigation technique. Neurosurg. Focus 2021, 51, E20. [CrossRef]
Sensors 2023, 23, 6202 33 of 37

57. Van Duren, B.H.; Sugand, K.; Wescott, R.; Carrington, R.; Hart, A. Augmented reality fluoroscopy simulation of the guide-wire
insertion in DHS surgery: A proof of concept study. Med. Eng. Phys. 2018, 55, 52–59. [CrossRef]
58. Luciano, C.J.; Banerjee, P.P.; Bellotte, B.; Oh, G.M.; Lemole, M., Jr.; Charbel, F.T.; Roitberg, B. Learning retention of thoracic
pedicle screw placement using a high-resolution augmented reality simulator with haptic feedback. Neurosurgery 2011, 69
(Suppl. Operative), ons14–ons19; discussion ons19. [CrossRef]
59. Virtual Reality Simulations in Healthcare. Available online: https://www.forbes.com/sites/forbestechcouncil/2022/01/24
/virtual-reality-simulations-in-healthcare/?sh=46cb0870382a (accessed on 2 May 2022).
60. Hou, L.; Dong, X.; Li, K.; Yang, C.; Yu, Y.; Jin, X.; Shang, S. Comparison of Augmented Reality-assisted and Instructor-assisted
Cardiopulmonary Resuscitation: A Simulated Randomized Controlled Pilot Trial. Clin. Simul. Nurs. 2022, 68, 9–18. [CrossRef]
61. Liu, W.P.; Richmon, J.D.; Sorger, J.M.; Azizian, M.; Taylor, R.H. Augmented reality and cone beam CT guidance for transoral
robotic surgery. J. Robotic Surg. 2015, 9, 223–233. [CrossRef]
62. Taha, M.; Sayed, M.; Zayed, H. Digital Vein Mapping Using Augmented Reality. Int. J. Intell. Eng. Syst. 2020, 13, 512–521.
[CrossRef]
63. Kuzhagaliyev, T.; Clancy, N.T.; Janatka, M.; Tchaka, K.; Vasconcelos, F.; Clarkson, M.J.; Gurusamy, K.; Hawkes, D.J.; Davidson,
B.; Stoyanov, D. Augmented Reality Needle Ablation Guidance Tool for Irreversible Electroporation in the Pancreas. In Medical
Imaging 2018: Image-Guided Procedures, Robotic Interventions, and Modeling; Proc. SPIE: Houston, TX, USA, 2018; Volume 10576.
64. AccuVein® Vein Vein Visualization: The Future of Healthcare Is Here. Available online: https://www.accuvein.com/why-
accuvein/ar/ (accessed on 1 May 2022).
65. NextVein. Available online: https://nextvein.com (accessed on 1 May 2022).
66. Ai, D.; Yang, J.; Fan, J.; Zhao, Y.; Song, X.; Shen, J.; Shao, L.; Wang, Y. Augmented reality based real-time subcutaneous vein
imaging system. Biomed. Opt. Express 2016, 7, 2565–2585. [CrossRef] [PubMed]
67. Kästner, L.; Frasineanu, V.; Lambrecht, J. A 3D-Deep-Learning-based Augmented Reality Calibration Method for Robotic
Environments using Depth Sensor Data. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation
(ICRA), Paris, France, 31 May–31 August 2020; pp. 1135–1141. [CrossRef]
68. Von Atzigen, M.; Liebmann, F.; Hoch, A.; Miguel Spirig, J.; Farshad, M.; Snedeker, J.; Fürnstahl, P. Marker-free surgical navigation
of rod bending using a stereo neural network and augmented reality in spinal fusion. Med. Image Anal. 2022, 77, 102365.
[CrossRef]
69. Pratt, P.; Ives, M.; Lawton, G.; Simmons, J.; Radev, N.; Spyropoulou, L.; Amiras, D. Through the HoloLens™ looking glass:
Augmented reality for extremity reconstruction surgery using 3D vascular models with perforating vessels. Eur. Radiol. Exp.
2018, 2, 2. [CrossRef] [PubMed]
70. Thøgersen, M.; Andoh, J.; Milde, C.; Graven-Nielsen, T.; Flor, H.; Petrini, L. Individualized augmented reality training reduces
phantom pain and cortical reorganization in amputees: A proof of concept study. J. Pain 2020, 21, 1257–1269. [CrossRef] [PubMed]
71. Rothgangel, A.; Bekrater-Bodmann, R. Mirror therapy versus augmented/virtual reality applications: Towards a tailored
mechanism-based treatment for phantom limb pain. Pain Manag. 2019, 9, 151–159. [CrossRef]
72. Mischkowski, R.A.; Zinser, M.J.; Kubler, A.C.; Krug, B.; Seifert, U.; Zoller, J.E. Application of an augmented reality tool for
maxillary positioning in orthognathic surgery: A feasibility study. J. Craniomaxillofac. Surg. 2006, 34, 478–483. [CrossRef]
73. Wang, J.; Suenaga, H.; Hoshi, K.; Yang, L.; Kobayashi, E.; Sakuma, I.; Liao, H. Augmented reality navigation with automatic
marker-free image registration using 3-D image overlay for dental surgery. IEEE Trans. Bio Med. Eng. 2014, 61, 1295–1304.
[CrossRef]
74. Liu, K.; Gao, Y.; Abdelrehem, A.; Zhang, L.; Chen, X.; Xie, L.; Wang, X. Augmented reality navigation method for recontouring
surgery of craniofacial fibrous dysplasia. Sci. Rep. 2021, 11, 10043. [CrossRef]
75. Pfefferle, M.; Shahub, S.; Shahedi, M.; Gahan, J.; Johnson, B.; Le, P.; Vargas, J.; Judson, B.O.; Alshara, Y.; Li, O.; et al. Renal biopsy
under augmented reality guidance. In Proceedings of the SPIE Medical Imaging, Houston, TX, USA, 16 March 2020. [CrossRef]
76. Nicolau, S.; Soler, L.; Mutter, D.; Marescaux, J. Augmented reality in laparoscopic surgical oncology. Surg. Oncol. 2011, 20,
189–201. [CrossRef]
77. Salah, Z.; Preim, B.; Elolf, E.; Franke, J.; Rose, G. Improved navigated spine surgery utilizing augmented reality visualization. In
Bildverarbeitung für die Medizin; Springer: Berlin/Heidelberg, Germany, 2011; pp. 319–323.
78. Pessaux, P.; Diana, M.; Soler, L.; Piardi, T.; Mutter, D.; Marescaux, J. Towards cybernetic surgery: Robotic and augmented
reality-assisted liver segmentectomy. Langenbecks Arch. Surg. 2015, 400, 381–385. [CrossRef] [PubMed]
79. Hussain, R.; Lalande, A.; Marroquin, R.; Guigou, C.; Grayeli, A.B. Video-based augmented reality combining CT-scan and
instrument position data to microscope view in middle ear surgery. Sci. Rep. 2020, 10, 6767. [CrossRef] [PubMed]
80. MURAB Project. Available online: https://www.murabproject.eu (accessed on 30 April 2022).
81. Zeng, F.; Wei, F. Hole filling algorithm based on contours information. In Proceedings of the 2nd International Conference on
Information Science and Engineering, Hangzhou, China, 4–6 December 2010. [CrossRef]
82. Chen, X.; Xu, L.; Wang, Y.; Wang, H.; Wang, F.; Zeng, X.; Wang, Q.; Egger, J. Development of a surgical navigation system based
on augmented reality using an optical see-through head-mounted display. J. Biomed. Inf. 2015, 55, 124–131. [CrossRef]
83. Ma, L.; Zhao, Z.; Chen, F.; Zhang, B.; Fu, L.; Liao, H. Augmented reality surgical navigation with ultrasound-assisted registration
for pedicle screw placement: A pilot study. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 2205–2215. [CrossRef] [PubMed]
Sensors 2023, 23, 6202 34 of 37

84. Hajek, J.; Unberath, M.; Fotouhi, J.; Bier, B.; Lee, S.C.; Osgood, G.; Maier, A.; Navab, N. Closing the Calibration Loop: An
Inside-out-tracking Paradigm for Augmented Reality in Orthopedic Surgery. In Medical Image Computing and Computer Assisted
Intervention–MICCAI 2018; Frangi, A., Schnabel, J., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Lecture Notes in
Computer Science; Springer: Cham, Switzerland, 2018; Volume 11073.
85. Elmi-Terander, A.; Nachabe, R.; Skulason, H.; Pedersen, K.; Soderman, M.; Racadio, J.; Babic, D.; Gerdhem, P.; Edstrom, E.
Feasibility and accuracy of thoracolumbar minimally invasive pedicle screw placement with augmented reality navigation
technology. Spine 2018, 43, 1018–1023. [CrossRef] [PubMed]
86. Dickey, R.M.; Srikishen, N.; Lipshultz, L.I.; Spiess, P.E.; Carrion, R.E.; Hakky, T.S. Augmented reality assisted surgery: A urologic
training tool. Asian J. Androl. 2016, 18, 732–734. [PubMed]
87. Wu, J.R.; Wang, M.L.; Liu, K.C.; Hu, M.H.; Lee, P.Y. Real-time advanced spinal surgery via visible patient model and augmented
reality system. Comput. Methods Programs Biomed. 2014, 113, 869–881. [CrossRef]
88. Wen, R.; Yang, L.; Chui, C.K.; Lim, K.B.; Chang, S. Intraoperative Visual Guidance and Control Interface for Augmented Reality Robotic
Surgery; IEEE: Piscataway, NJ, USA, 2010; pp. 947–952.
89. Simoes, M.; Cao, C.G. Leonardo: A first step towards an interactive decision aid for port-placement in robotic surgery. In
Proceedings of the 2013 IEEE International Conference on Systems, Man, and Cybernetics, Manchester, UK, 13–16 October 2013;
IEEE: Piscataway, NJ, USA, 2013; pp. 491–496.
90. Burström, G.; Nachabe, R.; Persson, O.; Edström, E.; Elmi Terander, A. Augmented and Virtual Reality Instrument Tracking for
Minimally Invasive Spine Surgery: A Feasibility and Accuracy Study. Spine 2019, 44, 1097–1104. [CrossRef]
91. Lee, D.; Kong, H.J.; Kim, D.; Yi, J.W.; Chai, Y.J.; Lee, K.E.; Kim, H.C. Preliminary study on application of augmented reality
visualization in robotic thyroid surgery. Ann. Surg. Treat Res. 2018, 95, 297–302. [CrossRef]
92. Agten, C.A.; Dennler, C.; Rosskopf, A.B.; Jaberg, L.; Pfirrmann, C.W.A.; Farshad, M. Augmented Reality-Guided Lumbar Facet
Joint Injections. Investig. Radiol. 2018, 53, 495–498. [CrossRef]
93. Ghaednia, H.; Fourman, M.S.; Lans, A.; Detels, K.; Dijkstra, H.; Lloyd, S.; Sweeney, A.; Oosterhoff, J.H.; Schwab, J.H. Augmented
and virtual reality in spine surgery, current applications and future potentials. Spine J. 2021, 21, 1617–1625. [CrossRef]
94. Nachabe, R.; Strauss, K.; Schueler, B.; Bydon, M. Radiation dose and image quality comparison during spine surgery with two
different, intraoperative 3D imaging navigation systems. J. Appl. Clin. Med. Phys. 2019, 20, 136–145. [CrossRef]
95. Londoño, M.C.; Danger, R.; Giral, M.; Soulillou, J.P.; Sánchez-Fueyo, A.; Brouard, S. A need for biomarkers of operational
tolerance in liver and kidney transplantation. Am. J. Transplant. 2012, 12, 1370–1377. [CrossRef]
96. Georgi, M.; Patel, S.; Tandon, D.; Gupta, A.; Light, A.; Nathan, A. How is the Digital Surgical Environment Evolving? The Role of
Augmented Reality in Surgery and Surgical Training. Preprints.org 2021, 2021100048. [CrossRef]
97. Calhoun, V.D.; Adali, T.; Giuliani, N.R.; Pekar, J.J.; Kiehl, K.A.; Pearlson, G.D. Method for multimodal analysis of independent
source differences in schizophrenia: Combining gray matter structural and auditory oddball functional data. Hum. Brain Mapp.
2006, 27, 47–62. [CrossRef] [PubMed]
98. Kronman, A.; Joskowicz, L. Image segmentation errors correction by mesh segmentation and deformation. In International
Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2013;
pp. 206–213.
99. Tamadazte, B.; Voros, S.; Boschet, C.; Cinquin, P.; Fouard, C. Augmented 3-d view for laparoscopy surgery. In Workshop on
Augmented Environments for Computer-Assisted Interventions; Springer: Berlin/Heidelberg, Germany, 2012; pp. 117–131.
100. Wang, A.; Wang, Z.; Lv, D.; Fang, Z. Research on a novel non-rigid registration for medical image based on SURF and APSO. In
Proceedings of the 2010 3rd International Congress on Image and Signal Processing, Yantai, China, 16–18 October 2010; IEEE:
Piscataway, NJ, USA, 2010; Volume 6, pp. 2628–2633.
101. Pandey, P.; Guy, P.; Hodgson, A.J.; Abugharbieh, R. Fast and automatic bone segmentation and registration of 3D ultrasound to
CT for the full pelvic anatomy: A comparative study. Int. J. Comput. Assist. Radiol. Surg. 2018, 13, 1515–1524. [CrossRef] [PubMed]
102. Hacihaliloglu, I. Ultrasound imaging and segmentation of bone surfaces: A review. Technology 2017, 5, 74–80. [CrossRef]
103. El-Hariri, H.; Pandey, P.; Hodgson, A.J.; Garbi, R. Augmented reality visualisation for orthopaedic surgical guidance with pre-and
intra-operative multimodal image data fusion. Healthc. Technol. Lett. 2018, 5, 189–193. [CrossRef]
104. Wittmann, W.; Wenger, T.; Zaminer, B.; Lueth, T.C. Automatic correction of registration errors in surgical navigation systems.
IEEE Trans. Biomed. Eng. 2011, 58, 2922–2930. [CrossRef]
105. Zhang, Y.; Wang, K.; Jiang, J.; Tan, Q. Research on intraoperative organ motion tracking method based on fusion of inertial and
electromagnetic navigation. IEEE Access 2021, 9, 49069–49081. [CrossRef]
106. Jiang, Z.; Gao, Z.; Chen, X.; Sun, W. Remote Haptic Collaboration for Virtual Training of Lumbar Puncture. J. Comput. 2013, 8,
3103–3110. [CrossRef]
107. Wu, C.; Wan, J.W. Multigrid methods with newton-gauss-seidel smoothing and constraint preserving interpolation for obstacle
problems. Numer. Math. Theory Methods Appl. 2015, 8, 199–219. [CrossRef]
108. Livyatan, H.; Yaniv, Z.; Joskowicz, L. Gradient-based 2-D/3-D rigid registration of fluoroscopic X-ray to CT. IEEE Trans. Med.
Imaging 2003, 22, 1395–1406. [CrossRef] [PubMed]
109. Martínez, H.; Skournetou, D.; Hyppölä, J.; Laukkanen, S.; Heikkilä, A. Drivers and Bottlenecks in the Adoption of Augmente-
dReality Applications. J. Multimed. Theory Appl. 2014, 2.
Sensors 2023, 23, 6202 35 of 37

110. Govers, F.X. Artificial Intelligence for Robotics: Build Intelligent Robots that Perform Human Tasks Using AI Techniques; Packt Publishing
Limited: Birmingham, UK, 2018.
111. Conti, A.; Guerra, M.; Dardari, D.; Decarli, N.; Win, M.Z. Network experimentation for cooperative localization. IEEE J. Sel. Areas
Commun. 2012, 30, 467–475. [CrossRef]
112. Zhang, X.; Wang, J.; Wang, T.; Ji, X.; Shen, Y.; Sun, Z.; Zhang, X. A markerless automatic deformable registration framework
for augmented reality navigation of laparoscopy partial nephrectomy. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1285–1294.
[CrossRef] [PubMed]
113. Garon, M.; Lalonde, J.F. Deep 6-DOF tracking. IEEE Trans. Vis. Comput. Graph. 2017, 23, 2410–2418. [CrossRef] [PubMed]
114. Abu Alhaija, H.; Mustikovela, S.K.; Mescheder, L.; Geiger, A.; Rother, C. Augmented reality meets deep learning for car instance
segmentation in urban scenes. In Proceedings of the BMVC 2017 and Workshops, London, UK, 4–7 September 2017.
115. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660.
116. Canalini, L.; Klein, J.; Miller, D.; Kikinis, R. Segmentation-based registration of ultrasound volumes for glioma resection in
image-guided neurosurgery. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 1697–1713. [CrossRef]
117. Doughty, M.; Singh, K.; Ghugre, N.R. SurgeonAssist-Net: Towards Context-Aware Head-Mounted Display-Based Augmented
Reality for Surgical Guidance. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer:
Cham, Switzerland, 2021; pp. 667–677.
118. Tanzi, L.; Piazzolla, P.; Porpiglia, F.; Vezzetti, E. Real-time deep learning semantic segmentation during intra-operative surgery
for 3D augmented reality assistance. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 1435–1445. [CrossRef]
119. Klein, G.; Murray, D. Parallel Tracking and Mapping for Small AR Workspaces. In Proceedings of the 2007 6th IEEE and ACM
International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; pp. 225–234. [CrossRef]
120. Mur-Artal, R.; Montiel, J.M.; Tardós, J.D. ORB-SLAM: A Versatile and Accurate Monocular SLAM System. IEEE Trans. Robot.
2015, 31, 1147–1163. [CrossRef]
121. Adhami, L.; Maniere, E.C. Optimal Planning for Minimally Invasive Surgical Robots. IEEE Trans. Robot. Autom. 2003, 19, 854–863.
[CrossRef]
122. Gonzalez-Barbosa, J.J.; Garcia-Ramirez, T.; Salas, J.; Hurtado-Ramos, J.B. Optimal camera placement for total coverage. In
Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 844–848.
123. Yildiz, E.; Akkaya, K.; Sisikoglu, E.; Sir, M.Y. Optimal camera placement for providing angular coverage in wireless video sensor
networks. IEEE Trans. Comput. 2013, 63, 1812–1825. [CrossRef]
124. Gadre, S.Y.; Rosen, E.; Chien, G.; Phillips, E.; Tellex, S.; Konidaris, G. End-User Robot Programming Using Mixed Reality. In
Proceedings of the International Conference on Robotics and Automation, Montreal, QC, Canada, 20–24 May 2019; pp. 2707–2713.
125. Fotouhi, J.; Song, T.; Mehrfard, A.; Taylor, G.; Wang, Q.; Xian, F.; Martin-Gomez, A.; Fuerst, B.; Armand, M.; Unberath, M.; et al.
Reflective-ar display: An interaction methodology for virtual-to-real alignment in medical robotics. IEEE Robot. Autom. Lett. 2020,
5, 2722–2729. [CrossRef]
126. Fang, H.C.; Ong, S.K.; Nee, A.Y.C. Orientation planning of robot end-effector using augmented reality. Int. J. Adv. Manuf. Technol.
2013, 67, 2033–2049. [CrossRef]
127. Bade, A.; Devadas, S.; Daman, D.; Suaib, N.M. Modeling and Simulation of Collision Response between Deformable Objects.
Available online: https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&doi=9d62cae770ce2e6e5d76013782cee973a3
be87e7 (accessed on 22 March 2023).
128. Sun, L.W.; Yeung, C.K. Port placement and pose selection of the da Vinci surgical system for collision-free intervention based
on performance optimization. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems,
San Diego, CA, USA, 29 October–2 November 2007; pp. 1951–1956.
129. Lee, I.; Lee, K.K.; Sim, O.; Woo, K.S.; Buyoun, C.; Oh, J.H. Collision detection system for the practical use of the humanoid robot.
In Proceedings of the IEEE-RAS 15th International Conference on Humanoid Robots, Seoul, Republic of Korea, 3–5 November
2015; pp. 972–976.
130. Zhang, Z.; Xin, Y.; Liu, B.; Li, W.X.; Lee, K.H.; Ng, C.F.; Stoyanov, D.; Cheung, R.C.C.; Kwok, K.-W. FPGA-Based High-Performance
Collision Detection: An Enabling Technique for Image-Guided Robotic Surgery. Front. Robot. AI 2016, 3, 51. [CrossRef]
131. Coste-Manière, È.; Olender, D.; Kilby, W.; Schulz, R.A. Robotic whole body stereotactic radiosurgery: Clinical advantages of the
CyberKnife®integrated system. Int. J. Med. Robot. Comput. Assist. Surg. 2005, 1, 28–39. [CrossRef]
132. Weede, O.; Mehrwald, M.; Wörn, H. Knowledge-based system for port placement and robot setup optimization in minimally
invasive surgery. IFAC Proc. Vol. 2012, 45, 722–728. [CrossRef]
133. Gao, S.; Lv, Z.; Fang, H. Robot-assisted and conventional freehand pedicle screw placement: A systematic review and meta-
analysis of randomized controlled trials. Eur. Spine J. 2018, 27, 921–930. [CrossRef]
134. Wang, L. Collaborative robot monitoring and control for enhanced sustainability. Int. J. Adv. Manuf. Technol. 2015, 81, 1433–1445.
[CrossRef]
135. Du, G.; Long, S.; Li, F.; Huang, X. Active Collision Avoidance for Human-Robot Interaction with UKF, Expert System, and
Artificial Potential Field Method. Front. Robot. AI 2018, 5, 125. [CrossRef]
136. Hongzhong, Z.; Fujimoto, H. Suppression of current quantization effects for precise current control of SPMSM using dithering
techniques and Kalman filter. IEEE Transac. Ind. Inform. 2014, 10, 1361–1371. [CrossRef]
Sensors 2023, 23, 6202 36 of 37

137. Das, N.; Yip, M. Learning-based proxy collision detection for robot motion planning applications. IEEE Trans. Robot. 2020, 36,
1096–1114. [CrossRef]
138. Torres, L.G.; Kuntz, A.; Gilbert, H.B.; Swaney, P.J.; Hendrick, R.J.; Webster, R.J.; Alterovitz, R. A motion planning approach to
automatic obstacle avoidance during concentric tube robot teleoperation. In Proceedings of the IEEE International Conference on
Robotics and Automation, Seattle, WA, USA, 26–30 May 2015; pp. 2361–2367.
139. Killian, L.; Backhaus, J. Utilizing the RRT*-Algorithm for Collision Avoidance in UAV Photogrammetry Missions. 2021. Available
online: https://arxiv.org/abs/2108.03863 (accessed on 22 March 2023).
140. Ranne, A.; Clark, A.B.; Rojas, N. Augmented Reality-Assisted Reconfiguration and Workspace Visualization of Malleable Robots:
Workspace Modification Through Holographic Guidance. IEEE Robot. Autom. Mag. 2022, 29, 10–21. [CrossRef]
141. Lipton, J.I.; Fay, A.J.; Rus, D. Baxter’s homunculus: Virtual reality spaces for teleoperation in manufacturing. IEEE Robot. Autom.
Lett. 2017, 3, 179–186. [CrossRef]
142. Bolano, G.; Fu, Y.; Roennau, A.; Dillmann, R. Deploying Multi-Modal Communication Using Augmented Reality in a Shared
Workspace. In Proceedings of the 2021 18th International Conference on Ubiquitous Robots (UR), Gangneung, Republic of Korea,
12–14 July 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 302–307.
143. Da Col, T.; Caccianiga, G.; Catellani, M.; Mariani, A.; Ferro, M.; Cordima, G.; De Momi, E.; Ferrigno, G.; de Cobelli, O. Automating
Endoscope Motion in Robotic Surgery: A Usability Study on da Vinci-Assisted Ex Vivo Neobladder Reconstruction. Front. Robot.
AI 2021, 8, 707704. [CrossRef]
144. Gao, W.; Tang, Q.; Yao, J.; Yang, Y. Automatic motion planning for complex welding problems by considering angular redundancy.
Robot. Comput. Integr. Manuf. 2020, 62, 101862. [CrossRef]
145. Zhang, Z.; Munawar, A.; Fischer, G.S. Implementation of a motion planning framework for the davinci surgical system research
kit. In Proceedings of the Hamlyn Symposium on Medical Robotics, London, UK, 12–15 July 2014.
146. Moon, H.C.; Park, S.J.; Kim, Y.D.; Kim, K.M.; Kang, H.; Lee, E.J.; Kim, M.-S.; Kim, J.W.; Kim, Y.H.; Park, C.-K.; et al. Navigation of
frameless fixation for gamma knife radiosurgery using fixed augmented reality. Sci. Rep. 2022, 12, 4486. [CrossRef]
147. Srinivasan, M.A.; Beauregard, G.L.; Brock, D.L. The impact of visual information on the haptic perception of stiffness in virtual
environments. Proc. ASME Dyn. Syst. Control Div. 1996, 58, 555–559.
148. Basdogan, C.; Ho, C.; Srinivasan, M.A.; Small, S.; Dawson, S. Force interactions in laparoscopic simulations: Haptic rendering of
soft tissues. In Medicine Meets Virtual Reality; IOS Press: Amsterdam, The Netherlands, 1998; pp. 385–391.
149. Latimer, C.W. Haptic Interaction with Rigid Objects Using Real-Time Dynamic Simulation. Ph.D. Thesis, Massachusetts Institute
of Technology, Boston, MA, USA, 1997.
150. Balanuik, R.; Costa, I.; Salisbury, J. Long Elements Method for Simulation of Deformable Objects. US Patent 2003/0088389 A1,
8 May 2003.
151. Okamura, A.M. Methods for haptic feedback in teleoperated robot-assisted surgery. Ind. Rob. 2004, 31, 499–508. [CrossRef]
152. Westebring-Van Der Putten, E.; Goossens, R.; Jakimowicz, J.; Dankelman, J. Haptics in minimally invasive surgery—A review.
Minim. Invasive Ther. Allied. Technol. 2008, 17, 3–16. [CrossRef]
153. Wurdemann, H.A.; Secco, E.L.; Nanayakkara, T.; Althoefer, K.; Mucha, L.; Rohr, K. Mapping tactile information of a soft
manipulator to a haptic sleeve in RMIS. In Proceedings of the 3rd Joint Workshop on New Technologies for Computer and Robot
Assisted Surgery, Verona, Italy, 11–13 September 2013; pp. 140–141.
154. Li, M.; Konstantinova, J.; Secco, E.L.; Jiang, A.; Liu, H.; Nanayakkara, T.; Seneviratne, L.D.; Dasgupta, P.; Althoefer, K.;
Wurdemann, H.A. Using visual cues to enhance haptic feedback for palpation on virtual model of soft tissue. Med. Biol. Eng.
Comput. 2015, 53, 1177–1186. [CrossRef] [PubMed]
155. Tsai, M.D.; Hsieh, M.S. Computer-based system for simulating spine surgery. In Proceedings of the 22nd IEEE International
Symposium on Computer-Based Medical Systems, Albuquerque, NM, USA, 2–5 August 2009; pp. 1–8.
156. Schendel, S.; Montgomery, K.; Sorokin, A.; Lionetti, G. A surgical simulator for planning and performing repair of cleft lips.
J. Cranio-Maxillofac. Surg. 2005, 33, 223–228. [CrossRef] [PubMed]
157. Olsson, P.; Nysjö, F.; Hirsch, J.M.; Carlbom, I.B. A haptics-assisted cranio-maxillofacial surgery planning system for restoring
skeletal anatomy in complex trauma cases. Int. J. Comput. Assist. Radiol. Surg. 2013, 8, 887–894. [CrossRef] [PubMed]
158. Richter, F.; Zhang, Y.; Zhi, Y.; Orosco, R.K.; Yip, M.C. Augmented reality predictive displays to help mitigate the effects of delayed
telesurgery. In Proceedings of the International Conference on Robotics and Automation, Montreal, QC, Canada, 20–24 May 2019;
pp. 444–450.
159. Ye, M.; Zhang, L.; Giannarou, S.; Yang, G.Z. Real-time 3D tracking of articulated tools for robotic surgery. In Interna-
tional Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2016;
pp. 386–394.
160. Marohn CM, R.; Hanly CE, J. Twenty-first century surgery using twenty-first century technology: Surgical robotics. Curr. Surg.
2004, 61, 466–473. [CrossRef]
161. Dake, D.K.; Ofosu, B.A. 5G enabled technologies for smart education. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 201–206. [CrossRef]
162. Grieco, M.; Elmore, U.; Vignali, A.; Caristo, M.E.; Persiani, R. Surgical Training for Transanal Total Mesorectal Excision in a Live
Animal Model: A Preliminary Experience. J. Laparoendosc. Adv. Surg. Tech. 2022, 32, 866–870. [CrossRef]
163. Takahashi, Y.; Hakamada, K.; Morohashi, H.; Akasaka, H.; Ebihara, Y.; Oki, E.; Hirano, S.; Mori, M. Verification of delay time and
image compression thresholds for telesurgery. Asian J. Endosc. Surg. 2022, 16, 255–261. [CrossRef]
Sensors 2023, 23, 6202 37 of 37

164. Sun, T.; He, X.; Li, Z. Digital twin in healthcare: Recent updates and challenges. Digit. Health 2023, 9, 20552076221149651.
[CrossRef]
165. Niederer, S.A.; Plank, G.; Chinchapatnam, P.; Ginks, M.; Lamata, P.; Rhode, K.S.; Rinaldi, C.A.; Razavi, R.; Smith, N.P. Length-
dependent tension in the failing heart and the efficacy of cardiac resynchronization therapy. Cardiovasc. Res. 2011, 89, 336–343.
[CrossRef]
166. Lebras, A.; Boustia, F.; Janot, K.; Lepabic, E.; Ouvrard, M.; Fougerou-Leurent, C.; Ferre, J.-C.; Gauvrit, J.-Y.; Eugene, F. Re-
hearsals using patient-specific 3D-printed aneurysm models for simulation of endovascular embolization of complex intracranial
aneurysms: 3D SIM study. J. Neuroradiol. 2021, 50, 86–92. [CrossRef]
167. Hernigou, P.; Safar, A.; Hernigou, J.; Ferre, B. Subtalar axis determined by combining digital twins and artificial intelligence:
Influence of the orientation of this axis for hindfoot compensation of varus and valgus knees. Int. Orthop. 2022, 46, 999–1007.
[CrossRef]
168. Diachenko, D.; Partyshev, A.; Pizzagalli, S.L.; Bondarenko, Y.; Otto, T.; Kuts, V. Industrial collaborative robot Digital Twin
integration and control using Robot Operating System. J. Mach. Eng. 2022, 22, 57–67. [CrossRef]
169. Riedel, P.; Riesner, M.; Wendt, K.; Aßmann, U. Data-Driven Digital Twins in Surgery utilizing Augmented Reality and Machine
Learning. In Proceedings of the 2022 IEEE International Conference on Communications Workshops (ICC Workshops), Seoul,
Republic of Korea, 16–20 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 580–585.
170. Qian, L.; Wu, J.Y.; DiMaio, S.P.; Navab, N.; Kazanzides, P. A review of augmented reality in robotic-assisted surgery. IEEE Trans.
Med. Robot. Bionics 2019, 2, 1–16. [CrossRef]
171. Yang, Z.; Shang, J.; Liu, C.; Zhang, J.; Liang, Y. Identification of oral cancer in OCT images based on an optical attenuation model.
Lasers Med. Sci. 2020, 35, 1999–2007. [CrossRef] [PubMed]
172. SpineAssist: Robotic Guidance System for Assisting in Spine Surgery. Available online: https://www.summitspine.com/
spineassist-robotic-guidance-system-for-assisting-in-spine-surgery-2/ (accessed on 12 April 2022).
173. Renaissance. Available online: https://neurosurgicalassociatespc.com/mazor-robotics-renaissance-guidance-system/patient-
information-about-renaissance/ (accessed on 12 April 2022).
174. ROSA Spine. Available online: https://www.zimmerbiomet.lat/en/medical-professionals/robotic-solutions/rosa-spine.html
(accessed on 12 April 2022).
175. MAZOR X STEALTH EDITION: Robotic Guidance System for Spinal Surgery. Available online: https://www.medtronic.com/us-
en/healthcare-professionals/products/spinal-orthopaedic/spine-robotics/mazor-x-stealth-edition.html (accessed on 12 April
2022).
176. Flex Robotic System. Available online: https://novusarge.com/en/medical-products/flex-robotic-system/ (accessed on 12 April
2022).
177. Medacta Announces First Surgeries in Japan with NextAR Augmented Reality Surgical Platform. Available online:
https://www.surgicalroboticstechnology.com/news/medacta-announces-first-surgeries-in-japan-with-nextar-augmented-
reality-surgical-platform/ (accessed on 10 June 2022).
178. Sutherland, G.R.; McBeth, P.B.; Louw, D.F. NeuroArm: An MR compatible robot for microsurgery. In International Congress Series;
Elsevier: Amsterdam, The Netherlands, 2003; Volume 1256, pp. 504–508.
179. Ma, X.; Song, C.; Qian, L.; Liu, W.; Chiu, P.W.; Li, Z. Augmented reality-assisted autonomous view adjustment of a 6-DOF robotic
stereo flexible endoscope. IEEE Trans. Med. Robot. Bionics 2022, 4, 356–367. [CrossRef]
180. Forte, M.P.; Gourishetti, R.; Javot, B.; Engler, T.; Gomez, E.D.; Kuchenbecker, K.J. Design of interactive augmented reality functions
for robotic surgery and evaluation in dry-lab lymphadenectomy. Int. J. Med. Robot. Comput. Assist. Surg. 2022, 18, e2351.
[CrossRef]
181. Qian, L.; Zhang, X.; Deguet, A.; Kazanzides, P. Aramis: Augmented reality assistance for minimally invasive surgery using a
head-mounted display. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference,
Shenzhen, China, 13–17 October 2019, Proceedings, Part V 22; Springer International Publishing: New York, NY, USA, 2019; pp. 74–82.
182. Doughty, M.; Ghugre, N.R.; Wright, G.A. Augmenting performance: A systematic review of optical see-through head-mounted
displays in surgery. J. Imaging 2022, 8, 203. [CrossRef]
183. Brunet, J.N.; Mendizabal, A.; Petit, A.; Golse, N.; Vibert, E.; Cotin, S. Physics-based deep neural network for augmented reality
during liver surgery. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019; Springer: Cham, Switzerland,
2019; pp. 137–145.
184. Marahrens, N.; Scaglioni, B.; Jones, D.; Prasad, R.; Biyani, C.S.; Valdastri, P. Towards Autonomous Robotic Minimally Invasive
Ultrasound Scanning and V essel Reconstruction on Non-Planar Surfaces. Front. Robot. AI 2022, 9, 178. [CrossRef] [PubMed]

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.

You might also like