Read Me Guide
Read Me Guide
File
The contents of this document are for information purposes only. INTERPOL and UNICRI assume
no liability or responsibility for any inaccurate or incomplete information, nor for any actions
taken in reliance thereon. The published material is distributed without warranty of any kind,
either express or implied, and the responsibility for the interpretation and use of the material lies
with the reader. In no event shall, INTERPOL or UNICRI be liable for damages arising from its use.
INTERPOL and UNICRI take no responsibility for the content of any external website referenced
in this publication or for any defamatory, offensive or misleading information which might be
contained on these third-party websites. Any links to external websites do not constitute an
endorsement by INTERPOL or UNICRI, and are only provided as a convenience. It is the responsibility
of the reader to evaluate the content and usefulness of information obtained from other sites.
The views, thoughts and opinions expressed in the content of this publication belong solely to the
authors and do not necessarily reflect the views or policies of, nor do they imply any endorsement
by, INTERPOL or the United Nations, their member countries or member states, their governing
bodies, or contributory organizations. Therefore, INTERPOL and UNICRI carry no responsibility for
the opinions expressed in this publication.
INTERPOL and UNICRI do not endorse or recommend any product, process, or service. Therefore,
mention of any products, processes, or services in this document cannot be construed as an
endorsement or recommendation by INTERPOL or UNICRI.
The designation employed and presentation of the material in this document do not imply the
expression of any opinion whatsoever on the part of the Secretariat of the United Nations, UNICRI
or INTERPOL, concerning the legal status of any country, territory, city or area of its authorities, or
concerning the delimitation of its frontiers or boundaries.
The contents of this document may be quoted or reproduced, provided that the source of
information is acknowledged. INTERPOL and UNICRI would like to receive a copy of the document
in which this publication is used or quoted.
             FOREWORD
                                                      JUERGEN STOCK
                                            Secretary General, INTERPOL
             FOREWORD
The Toolkit for Responsible AI Innovation in Law Enforcement is a tailor-made set of resources for
the global law enforcement community, developed by INTERPOL and UNICRI with the financial
support of the European Union. Its objective is to support law enforcement agencies to navigate
the complex task of institutionalizing responsible AI innovation and integrating AI systems into
their work. It has been designed in full recognition of both the opportunities and the challenges
that AI presents for the law enforcement community, and it is intended to offer support to agencies
interested in or already committed to responsible AI innovation. More specifically, this AI Toolkit
provides law enforcement agencies with a theoretical foundation on responsible AI innovation
based on human rights law, ethics and policing principles, as well as several practical tools to
support them with putting responsible AI innovation theory into practice at each and every stage
of their AI journey.
                                                                                                 6
                                                                                         README File
This AI Toolkit is comprised of seven individual standalone resources. Together, these resources
provide agencies with the necessary knowledge and understanding to be able to benefit from the
positive potential of these systems while navigating the possible challenges and risks, whatever
the agency’s level of experience with AI. The resources have all been designed to be informative,
practical and actionable, providing guidance and best practices to move forward along the path
of responsible AI innovation.
The seven resources are grouped into three categories (see the table):
The AI Toolkit is designed to provide the law enforcement community with a balanced perspective
on responsible AI innovation that is grounded in human rights, ethics, and established good
policing principles, all contextualized with practical examples of law enforcement-specific use
cases which are provided throughout the AI Toolkit.
It is important to note that the AI Toolkit is not intended to create binding obligations or to limit
or undermine any of an agency’s binding obligations under international or domestic law. Rather,
the AI Toolkit provides guidance, considerations and recommendations to be followed voluntarily.
Equally, the AI Toolkit is not intended to replace or subsume any of the management strategies,
policies or procedures already in place in law enforcement agencies, such as project management,
risk and needs analysis, data protection and assessment, or risk-benefit analysis. Instead, it is
designed to complement existing institutional and organizational processes and procedures: it is
most useful when used in conjunction with other relevant management tools.
                                                                                                   7
                                                                                       README File
However, it is also important to highlight the fact that ‘law enforcement’ can be a broad term.
For the purposes of the AI Toolkit, ‘law enforcement’ refers to the police and other similar state
authorities that exercise police functions such as investigating crimes, protecting people and
property, and maintaining public order and safety. In some cases, the term ‘law enforcement’
may include border control authorities, corrections officers, or counter-terrorism entities, for
example. In some cases, the judiciary, in particular prosecutors, is also considered to be part of
‘law enforcement’. While the AI Toolkit was not designed primarily with these latter kinds of ‘law
enforcement’ users in mind, it may nevertheless be of value for these authorities when exercising
police functions.
It is crucial to note that, apart from law enforcement personnel, many other non-law enforcement
stakeholders play a key role in implementing responsible AI innovation in law enforcement. These
include technology developers in the private sector or academia, civil society, the general public,
and other criminal justice actors such as the judiciary, prosecutors and lawyers – if they are not
already considered part of ‘law enforcement’. The AI Toolkit has been developed with the role of
these ‘secondary’ stakeholders in mind, and seeks to simultaneously increase general knowledge
and awareness of the underlying themes of responsible AI innovation in law enforcement,
demystify the issues involved, and facilitate more informed discussions among and between all
groups of stakeholders.
This approach takes into account the fact that each agency’s experience with AI, and their journey
towards responsible AI innovation, is unique. While some agencies may already be using AI
                                                                                                 8
                                                                                         README File
systems to support their day-to-day work and may have the corresponding in-house expertise,
other agencies may still be getting to grips with general knowledge and understanding of AI and
exploring the available technologies. This approach also allows for the fact that the AI Toolkit will
be relevant for many different types of law enforcement personnel at different stages of the life
cycle of an AI system.
                                                                                                   9
                                                                                       README File
          This pathway begins with all AI Toolkit users – regardless of how familiar the agency
          or individual is with AI – getting acquainted with the concepts upon which the AI
 1
          Toolkit is founded and the key terms that arise repeatedly throughout. With this in
          mind, it is recommended that AI Toolkit users start by familiarizing themselves with
          the Introduction to Responsible AI Innovation and the Principles for Responsible AI
          Innovation, referring as needed to the Technical Reference Book for more technical
          detail on AI.
          Having established this foundation, it is recommended that each agency carries out a
          self-assessment of its organizational readiness using the Organizational Readiness
 2
          Assessment Questionnaire, supported by the Organizational Roadmap. Again, it is
          recommended that this assessment be conducted regardless of the level of experience
          that an agency already has regarding the use of AI systems. This exercise will help
          the agency to understand at what stage it is in terms of the necessary capabilities for
          beginning or continuing the journey toward responsible AI innovation.
          AI Toolkit users – particular those moving from a more general exploration of the
          opportunities AI systems present toward the actual deployment of an AI system – are
          then advised to move on to the Risk Assessment Questionnaire and the Responsible
          AI Innovation in Action Workbook.
 3
          communities associated with a specific AI use case, thus allowing them to better align
          their decisions with the Principles for Responsible AI Innovation. Given that assessing
          risks is an iterative process, it is recommended that the Risk Assessment Questionnaire
          be conducted periodically.
The following table provides an overview of the main aims of each resource and a provisional
list of the users that may find the resources valuable. This is intended to provide further support
for agencies in determining when to consult each resource and who should be assigned the
responsibility of familiarizing themselves with its content.
                                                                                                10
                                                                            README File
                                                                                        11
                                                                            README File
                                                                                       12
                                                                                          README File
Given the rapid advancements in this field, the specific impact and risk of any single AI system
cannot be determined solely by looking at the nature of the use case itself or whether it was
especially developed for the law enforcement context. On the contrary, understanding and
managing the risks that may derive from implementing AI systems in law enforcement involves
a consideration of a diverse range of circumstances. In practical terms, this means that there
can be no definitive ‘threshold’ to establish when the AI Toolkit should be consulted or when its
guidance, recommendations and considerations should be acted upon.
Despite this, given their nature, the core principles for responsible AI innovation can be considered
applicable to all uses of AI systems and at all stages of the life cycle. They are general principles
designed to be adaptable to different contexts, and respecting them will contribute to promoting
good practices as agencies leverage the positive potential of using AI systems. With regard to the
                                                                                                   13
                                                                                          README File
applicability of the guidance, recommendations and considerations in the rest of the AI Toolkit,
agencies should consider the following three questions:
1. Does the use of the AI system pertain to the execution of the core functions of the law
   enforcement agency, namely, to protect the community, prevent and investigate crime, and
   ensure justice?
2. Having completed the Risk Assessment Questionnaire, have medium, high, or extremely
   high risks pertaining to the AI system been identified?
3. In the event that the AI system has been identified as presenting a low risk, are there
   any other potential effects on human rights that would necessitate considering the full
   application of the AI Toolkit?
If the answer to any of these three questions is yes, then the recommendation is to take into
account the guidance, recommendations and considerations contained within the AI Toolkit to
the fullest extent possible.
The AI Toolkit has been developed through a highly inclusive, consultative, and iterative process.
This process has been specifically designed with a view to ensuring that the AI Toolkit is technically
and operationally sound, promoting close interdisciplinary coordination with industry, academia,
criminal justice practitioners, civil society organizations and the general public, and fostering a
sense of transparency and a broader acceptance of the use of AI in law enforcement
                                                                                                   14
                                                                                       README File
INTERPOL and UNICRI established this group as an informal external review body to support
and guide the development of the AI Toolkit from a substantive perspective. It was comprised
of 56 experts from law enforcement, academia, the private sector, and civil society from over 30
different countries. The Core Group of Experts was convened in a virtual format on six occasions
during the development process.
INTERPOL and UNICRI carried out regular consultations with representatives of the law
enforcement community throughout the development of the AI Toolkit. In addition, specific group
consultation sessions were organized with criminal justice practitioners, representatives from the
human rights community, and technology providers. When necessary, individual subject matter
experts/boards/organizations were also contacted on an ad-hoc basis for additional consultation
on specific topics. Finally, as public engagement, trust, and law enforcement transparency
play a vital role in the concept of responsible AI innovation, a global public opinion survey was
conducted to map public perceptions of the use of AI systems by law enforcement agencies,
and consultations were organized on specific public concerns around groups in vulnerable
circumstances. Feedback received during this process informed the drafting of the sections of
the AI Toolkit which address the importance of public engagement.
Peer Review:
Ahead of the AI Toolkit’s public release (V1.0), INTERPOL and UNICRI facilitated a technical review
of the AI Toolkit involving more than 30 experts with backgrounds in law enforcement, academia,
human rights, industry, criminal justice, the public sector, and international organizations from
across the globe. During this process, detailed feedback on each of the resources was received
and integrated into the AI Toolkit.
To complement the technical peer review, the AI Toolkit has undergone testing by national and
local law enforcement agencies from 15 different countries, identified in response to a call for
volunteers issued by INTERPOL to all its member countries. Feedback received from testing
countries has been compiled and used to improve the AI Toolkit, resulting in the current revised
version.
                                                                                                15
                                                                               README File
ACKNOWLEDGMENTS
 INTERPOL and UNICRI would like to express their deep appreciation to
 the many experts from within the law enforcement community and other
 national authorities, as well as industry, academia, civil society, and
 international and regional organizations, that have devoted their time,
 insights and perspectives to the process of developing the AI Toolkit.
 Their contribution is invaluable and has been pivotal in shaping the AI
 Toolkit and ensuring its content is technically accurate, comprehensive
 from all perspectives, understandable, and implementable by law
 enforcement agencies. INTERPOL and UNICRI would also like to
 specifically thank the law enforcement agencies who participated in
 testing the AI Toolkit, helping to ensure that it is a practical and useful
 set of resources that will benefit the law enforcement community.
 Without the collective contributions and collaboration of all those
 involved, the AI Toolkit would not have been possible. This collaborative
 effort is a testament to the strength of international cooperation, and it
 demonstrates the shared commitment of all involved to ensuring that
 responsible AI innovation is the future of policing.
How to cite this publication: UNICRI and INTERPOL. (Revised February 2024). Toolkit for
Responsible AI Innovation in Law Enforcement: README file.
© United Nations Interregional Crime and Justice Research Institute (UNICRI), 2024
www.ai-lawenforcement.org
XVIII